Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
5,675
29,513,460,559
IssuesEvent
2023-06-04 07:50:17
debanjum/khoj
https://api.github.com/repos/debanjum/khoj
opened
Create benchmarks for search
maintain
We need a consistent mechanism for evaluating search quality across different data sources over time for quality assessment.
True
Create benchmarks for search - We need a consistent mechanism for evaluating search quality across different data sources over time for quality assessment.
main
create benchmarks for search we need a consistent mechanism for evaluating search quality across different data sources over time for quality assessment
1
3,105
11,868,460,111
IssuesEvent
2020-03-26 09:14:04
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFM - freac.portable
Status: Available For Maintainer(s)
## Current Maintainer - [x] I am the maintainer of the package and wish to pass it to someone else; ## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/freac.portable Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/freac.portable
True
RFM - freac.portable - ## Current Maintainer - [x] I am the maintainer of the package and wish to pass it to someone else; ## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/freac.portable Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/freac.portable
main
rfm freac portable current maintainer i am the maintainer of the package and wish to pass it to someone else i don t want to become the maintainer i have followed the package triage process and i do not want to become maintainer of the package there is no existing open maintainer request for this package checklist issue title starts with rfm existing package details package url package source url
1
33,824
2,772,815,007
IssuesEvent
2015-05-03 01:37:56
synergy/synergy
https://api.github.com/repos/synergy/synergy
reopened
Auto-start GUI setting on Linux
bug priority-soon
**Imported issue:** * Author: Nick Bolton * Date: 2011-12-06 15:48:52 * Legacy ID: 3025
1.0
Auto-start GUI setting on Linux - **Imported issue:** * Author: Nick Bolton * Date: 2011-12-06 15:48:52 * Legacy ID: 3025
non_main
auto start gui setting on linux imported issue author nick bolton date legacy id
0
49,254
6,019,188,633
IssuesEvent
2017-06-07 14:02:51
sbsdev/daisyproducer
https://api.github.com/repos/sbsdev/daisyproducer
closed
Import of GD productions from ABACUS fails
deployed on testing
The xml from ABACUS now contains an new element `<drucker/>`. That change makes the validation fail and hence the import.
1.0
Import of GD productions from ABACUS fails - The xml from ABACUS now contains an new element `<drucker/>`. That change makes the validation fail and hence the import.
non_main
import of gd productions from abacus fails the xml from abacus now contains an new element that change makes the validation fail and hence the import
0
73,844
19,842,831,676
IssuesEvent
2022-01-21 00:29:32
tsunamayo/Starship-EVO
https://api.github.com/repos/tsunamayo/Starship-EVO
opened
[New build - DEFAULT] 22w03c: Space Battle Editor
Build Release Note
The "spawn NPC" F8 menu evolved in a full blown Space Battle editor. This is mainly to help me improve, balance and optimize space combat, but also to design larger battles to be added with faction encounters. - Spawn several ships at the same time - Choose the NPC behaviour: foe / friend. - Set a spawning distance. - Export a given space battle blueprint. You will need to be flying a ship for it to work correctly. Hotfixes: #4488 #3416 Decals are not mirrored correctly on some orientations. #4491 Rotor preview incorrect at various grid-size. #4495 #4496 Children entity can appears at the wrong place after a blueprint deletion.
1.0
[New build - DEFAULT] 22w03c: Space Battle Editor - The "spawn NPC" F8 menu evolved in a full blown Space Battle editor. This is mainly to help me improve, balance and optimize space combat, but also to design larger battles to be added with faction encounters. - Spawn several ships at the same time - Choose the NPC behaviour: foe / friend. - Set a spawning distance. - Export a given space battle blueprint. You will need to be flying a ship for it to work correctly. Hotfixes: #4488 #3416 Decals are not mirrored correctly on some orientations. #4491 Rotor preview incorrect at various grid-size. #4495 #4496 Children entity can appears at the wrong place after a blueprint deletion.
non_main
space battle editor the spawn npc menu evolved in a full blown space battle editor this is mainly to help me improve balance and optimize space combat but also to design larger battles to be added with faction encounters spawn several ships at the same time choose the npc behaviour foe friend set a spawning distance export a given space battle blueprint you will need to be flying a ship for it to work correctly hotfixes decals are not mirrored correctly on some orientations rotor preview incorrect at various grid size children entity can appears at the wrong place after a blueprint deletion
0
479,681
13,804,541,105
IssuesEvent
2020-10-11 09:29:09
AY2021S1-TIC4001-2/tp
https://api.github.com/repos/AY2021S1-TIC4001-2/tp
closed
Add saveIncomeCategories method to Storage class
priority.High type.Task
... for storing the income categories in the income category list to the harddisk.
1.0
Add saveIncomeCategories method to Storage class - ... for storing the income categories in the income category list to the harddisk.
non_main
add saveincomecategories method to storage class for storing the income categories in the income category list to the harddisk
0
2,869
10,275,929,668
IssuesEvent
2019-08-24 12:47:57
arcticicestudio/arctic
https://api.github.com/repos/arcticicestudio/arctic
closed
ESLint
context-workflow scope-dx scope-maintainability scope-quality scope-stability type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/63634799-a8555900-c65b-11e9-8988-dd0bc5d9c03b.png" /></p> Integrate [ESLint][], the _pluggable_ and de-facto standard linting utility for JavaScript. ### Configuration Preset The configuration presets that will be used are [@arcticicestudio/eslint-config][pr] that implements the [Arctic Ice Studio JavaScript Style][stg-js]. It comes with the following peer dependencies: - [eslint][esl-gh] - [babel-eslint][esl-pr-b] It it built on top of [@arcticicestudio/eslint-config-base][pr-b] that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [eslint-config-prettier][esl-c-pr] - [eslint-plugin-babel][esl-p-b] - [eslint-plugin-import][esl-p-i] - [eslint-plugin-jsx-a11y][esl-p-a11y] - [eslint-plugin-prettier][esl-p-pr] - [eslint-plugin-react-hooks][esl-p-r-h] - [eslint-plugin-react][esl-p-r] Since _arctic_ will be built with [TypeScript][ts], the [@arcticicestudio/eslint-config-typescript][pr-ts] preset will be extended to add support for _TypeScript_ source file linting and compatibility with [Prettier][] through the [`@arcticicestudio/eslint-config-typescript/prettier` extension entry point][pr-ts-d#ep]. This preset requires the following peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [typescript][gh-ts] Since the custom presets are still in major version `0` note that the version range should be `>=0.x.x <1.0.0` to avoid the “SemVer Major Zero Caveat”. When defining package versions with the the carat `^` or tilde `~` range selector it won't affect packages with a major version of `0`. _yarn_ will resolve these packages to their exact version until the major version is greater or equal to `1`. To avoid this caveat the more detailed version range `>=0.x.x <1.0.0` should be used to resolve all versions greater or equal to `0.x.x` but less than `1.0.0`. This will always use the latest `0.x.x` version and removes the need to increment the version manually on each new release. To allow to lint TypeScript code the `@typescript-eslint/parser` parser will be used and [specified as parser][esl-d-parser] next to the main [Babel parser][esl-pr-b]. Also to make use of the latest experimental Babel features and proposals, [eslint-plugin-babel][esl-p-b] will be added with the following rule configurations: - `babel/camelcase` with level `error` - doesn't complain about optional chaining (`let foo = bar?.a_b;`). Note that the [core rule `camelcase`][esl-r-cc] must be disabled! - `babel/no-unused-expressions` with level `error` - doesn't fail when using `do` expressions or optional chaining (`a?.b()`). Note that the [core rule `no-unused-expressions`][esl-r-nue] must be disabled! See the [documentation of provided rules][esl-p-b#r]and required configurations to use them. The `.eslintrc.js` configuration file will be placed in the project root next to the `.eslintignore` file to define ignore pattern. #### Webpack Import Resolving Strategy To prepare for a better developer experience with Webpack (that will be used later on through Gatsby) the [resolvers of the eslint-plugin-import][esl-p-i#res] will be configured for the `src` and `src/components` paths. ### Package Script To allow to run the JavaScript linting separately a `lint:js` npm script/task will be added to be included in the main `lint` script flow. To use the great [auto-fixing][esl-d-cli#af] feature another `format:js` script/task will be added. ## Tasks - [x] Install required packages to as development dependencies: - [@arcticicestudio/eslint-config-typescript][npm-esl-c-ais-ts] - [@arcticicestudio/eslint-config][npm-esl-c-ais] - [eslint-config-prettier][npm-esl-c-pr] - [eslint-plugin-babel][npm-esl-p-b] - [eslint-plugin-import][npm-esl-p-i] - [eslint-plugin-jsx-a11y][npm-esl-p-a11y] - [eslint-plugin-prettier][npm-esl-p-pr] - [eslint-plugin-react-hooks][npm-esl-p-r-h] - [eslint-plugin-react][npm-esl-p-r] - [eslint][npm-esl] - [typescript][npm-ts] - [x] Implement `.eslintrc.js` configuration file. - [x] Extend installed presets. - [x] [Prepare Webpack](#webpack-preparations) compatibility with resolvers. - [x] Integrate [eslint-plugin-babel][npm-esl-p-b] - [x] Enable `babel/no-unused-expressions` and `babel/camelcase` rules including the deactivation of their associated core rules. - [x] Add `babel` to the array of enabled plugins. - [x] Implement `.eslintignore` ignore pattern file. - [x] Implement npm `format:fix-js`, `format:fix-ts`, `lint:js` and `lint:ts` scripts. - [x] Lint current code base for the first time and fix possible JavaScript style guide violations. [esl-c-pr]: https://github.com/prettier/eslint-config-prettier [esl-d-cli#af]: https://eslint.org/docs/user-guide/command-line-interface#fixing-problems [esl-d-parser]: https://eslint.org/docs/user-guide/configuring#specifying-parser [esl-p-a11y]: https://github.com/evcohen/eslint-plugin-jsx-a11y [esl-p-b]: https://github.com/babel/eslint-plugin-babel [esl-p-b#r]: https://github.com/babel/eslint-plugin-babel#rules [esl-p-i]: https://github.com/benmosher/eslint-plugin-import [esl-p-i#res]: https://github.com/benmosher/eslint-plugin-import#resolvers [esl-p-pr]: https://github.com/prettier/eslint-plugin-prettier [esl-p-r-h]: https://github.com/facebook/react/tree/master/packages/eslint-plugin-react-hooks [esl-p-r]: https://github.com/yannickcr/eslint-plugin-react [esl-pr-b]: https://github.com/babel/babel-eslint [esl-r-cc]: https://eslint.org/docs/rules/camelcase [esl-r-nue]: https://eslint.org/docs/rules/no-unused-expressions [esl-ts-p]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/eslint-plugin [esl-ts-pa]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/parser [eslint]: https://eslint.org [esl-gh]: https://github.com/eslint/eslint [gh-ts]: https://github.com/Microsoft/TypeScript [npm-esl-c-ais-ts]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config-typescript [npm-esl-c-ais]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config [npm-esl-c-pr]: https://www.npmjs.com/package/eslint-config-prettier [npm-esl-p-a11y]: https://www.npmjs.com/package/eslint-plugin-jsx-a11y [npm-esl-p-b]: https://www.npmjs.com/package/eslint-plugin-babel [npm-esl-p-i]: https://www.npmjs.com/package/eslint-plugin-import [npm-esl-p-pr]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-esl-p-r-h]: https://www.npmjs.com/package/eslint-plugin-react-hooks [npm-esl-p-r]: https://www.npmjs.com/package/eslint-plugin-react [npm-esl]: https://www.npmjs.com/package/eslint [npm-ts]: https://www.npmjs.com/package/typescript [pr-b]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-base [pr-ts-d#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points [pr-ts]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-typescript [pr]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config [prettier]: https://prettier.io [stg-js]: https://arcticicestudio.github.io/styleguide-javascript [ts]: https://www.typescriptlang.org
True
ESLint - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63634799-a8555900-c65b-11e9-8988-dd0bc5d9c03b.png" /></p> Integrate [ESLint][], the _pluggable_ and de-facto standard linting utility for JavaScript. ### Configuration Preset The configuration presets that will be used are [@arcticicestudio/eslint-config][pr] that implements the [Arctic Ice Studio JavaScript Style][stg-js]. It comes with the following peer dependencies: - [eslint][esl-gh] - [babel-eslint][esl-pr-b] It it built on top of [@arcticicestudio/eslint-config-base][pr-b] that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [eslint-config-prettier][esl-c-pr] - [eslint-plugin-babel][esl-p-b] - [eslint-plugin-import][esl-p-i] - [eslint-plugin-jsx-a11y][esl-p-a11y] - [eslint-plugin-prettier][esl-p-pr] - [eslint-plugin-react-hooks][esl-p-r-h] - [eslint-plugin-react][esl-p-r] Since _arctic_ will be built with [TypeScript][ts], the [@arcticicestudio/eslint-config-typescript][pr-ts] preset will be extended to add support for _TypeScript_ source file linting and compatibility with [Prettier][] through the [`@arcticicestudio/eslint-config-typescript/prettier` extension entry point][pr-ts-d#ep]. This preset requires the following peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [typescript][gh-ts] Since the custom presets are still in major version `0` note that the version range should be `>=0.x.x <1.0.0` to avoid the “SemVer Major Zero Caveat”. When defining package versions with the the carat `^` or tilde `~` range selector it won't affect packages with a major version of `0`. _yarn_ will resolve these packages to their exact version until the major version is greater or equal to `1`. To avoid this caveat the more detailed version range `>=0.x.x <1.0.0` should be used to resolve all versions greater or equal to `0.x.x` but less than `1.0.0`. This will always use the latest `0.x.x` version and removes the need to increment the version manually on each new release. To allow to lint TypeScript code the `@typescript-eslint/parser` parser will be used and [specified as parser][esl-d-parser] next to the main [Babel parser][esl-pr-b]. Also to make use of the latest experimental Babel features and proposals, [eslint-plugin-babel][esl-p-b] will be added with the following rule configurations: - `babel/camelcase` with level `error` - doesn't complain about optional chaining (`let foo = bar?.a_b;`). Note that the [core rule `camelcase`][esl-r-cc] must be disabled! - `babel/no-unused-expressions` with level `error` - doesn't fail when using `do` expressions or optional chaining (`a?.b()`). Note that the [core rule `no-unused-expressions`][esl-r-nue] must be disabled! See the [documentation of provided rules][esl-p-b#r]and required configurations to use them. The `.eslintrc.js` configuration file will be placed in the project root next to the `.eslintignore` file to define ignore pattern. #### Webpack Import Resolving Strategy To prepare for a better developer experience with Webpack (that will be used later on through Gatsby) the [resolvers of the eslint-plugin-import][esl-p-i#res] will be configured for the `src` and `src/components` paths. ### Package Script To allow to run the JavaScript linting separately a `lint:js` npm script/task will be added to be included in the main `lint` script flow. To use the great [auto-fixing][esl-d-cli#af] feature another `format:js` script/task will be added. ## Tasks - [x] Install required packages to as development dependencies: - [@arcticicestudio/eslint-config-typescript][npm-esl-c-ais-ts] - [@arcticicestudio/eslint-config][npm-esl-c-ais] - [eslint-config-prettier][npm-esl-c-pr] - [eslint-plugin-babel][npm-esl-p-b] - [eslint-plugin-import][npm-esl-p-i] - [eslint-plugin-jsx-a11y][npm-esl-p-a11y] - [eslint-plugin-prettier][npm-esl-p-pr] - [eslint-plugin-react-hooks][npm-esl-p-r-h] - [eslint-plugin-react][npm-esl-p-r] - [eslint][npm-esl] - [typescript][npm-ts] - [x] Implement `.eslintrc.js` configuration file. - [x] Extend installed presets. - [x] [Prepare Webpack](#webpack-preparations) compatibility with resolvers. - [x] Integrate [eslint-plugin-babel][npm-esl-p-b] - [x] Enable `babel/no-unused-expressions` and `babel/camelcase` rules including the deactivation of their associated core rules. - [x] Add `babel` to the array of enabled plugins. - [x] Implement `.eslintignore` ignore pattern file. - [x] Implement npm `format:fix-js`, `format:fix-ts`, `lint:js` and `lint:ts` scripts. - [x] Lint current code base for the first time and fix possible JavaScript style guide violations. [esl-c-pr]: https://github.com/prettier/eslint-config-prettier [esl-d-cli#af]: https://eslint.org/docs/user-guide/command-line-interface#fixing-problems [esl-d-parser]: https://eslint.org/docs/user-guide/configuring#specifying-parser [esl-p-a11y]: https://github.com/evcohen/eslint-plugin-jsx-a11y [esl-p-b]: https://github.com/babel/eslint-plugin-babel [esl-p-b#r]: https://github.com/babel/eslint-plugin-babel#rules [esl-p-i]: https://github.com/benmosher/eslint-plugin-import [esl-p-i#res]: https://github.com/benmosher/eslint-plugin-import#resolvers [esl-p-pr]: https://github.com/prettier/eslint-plugin-prettier [esl-p-r-h]: https://github.com/facebook/react/tree/master/packages/eslint-plugin-react-hooks [esl-p-r]: https://github.com/yannickcr/eslint-plugin-react [esl-pr-b]: https://github.com/babel/babel-eslint [esl-r-cc]: https://eslint.org/docs/rules/camelcase [esl-r-nue]: https://eslint.org/docs/rules/no-unused-expressions [esl-ts-p]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/eslint-plugin [esl-ts-pa]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/parser [eslint]: https://eslint.org [esl-gh]: https://github.com/eslint/eslint [gh-ts]: https://github.com/Microsoft/TypeScript [npm-esl-c-ais-ts]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config-typescript [npm-esl-c-ais]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config [npm-esl-c-pr]: https://www.npmjs.com/package/eslint-config-prettier [npm-esl-p-a11y]: https://www.npmjs.com/package/eslint-plugin-jsx-a11y [npm-esl-p-b]: https://www.npmjs.com/package/eslint-plugin-babel [npm-esl-p-i]: https://www.npmjs.com/package/eslint-plugin-import [npm-esl-p-pr]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-esl-p-r-h]: https://www.npmjs.com/package/eslint-plugin-react-hooks [npm-esl-p-r]: https://www.npmjs.com/package/eslint-plugin-react [npm-esl]: https://www.npmjs.com/package/eslint [npm-ts]: https://www.npmjs.com/package/typescript [pr-b]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-base [pr-ts-d#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points [pr-ts]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-typescript [pr]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config [prettier]: https://prettier.io [stg-js]: https://arcticicestudio.github.io/styleguide-javascript [ts]: https://www.typescriptlang.org
main
eslint integrate the pluggable and de facto standard linting utility for javascript configuration preset the configuration presets that will be used are that implements the it comes with the following peer dependencies it it built on top of that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies since arctic will be built with the preset will be extended to add support for typescript source file linting and compatibility with through the this preset requires the following peer dependencies since the custom presets are still in major version note that the version range should be x x to avoid the “semver major zero caveat” when defining package versions with the the carat or tilde range selector it won t affect packages with a major version of yarn will resolve these packages to their exact version until the major version is greater or equal to to avoid this caveat the more detailed version range x x should be used to resolve all versions greater or equal to x x but less than this will always use the latest x x version and removes the need to increment the version manually on each new release to allow to lint typescript code the typescript eslint parser parser will be used and next to the main also to make use of the latest experimental babel features and proposals will be added with the following rule configurations babel camelcase with level error doesn t complain about optional chaining let foo bar a b note that the must be disabled babel no unused expressions with level error doesn t fail when using do expressions or optional chaining a b note that the must be disabled see the and required configurations to use them the eslintrc js configuration file will be placed in the project root next to the eslintignore file to define ignore pattern webpack import resolving strategy to prepare for a better developer experience with webpack that will be used later on through gatsby the will be configured for the src and src components paths package script to allow to run the javascript linting separately a lint js npm script task will be added to be included in the main lint script flow to use the great feature another format js script task will be added tasks install required packages to as development dependencies implement eslintrc js configuration file extend installed presets webpack preparations compatibility with resolvers integrate enable babel no unused expressions and babel camelcase rules including the deactivation of their associated core rules add babel to the array of enabled plugins implement eslintignore ignore pattern file implement npm format fix js format fix ts lint js and lint ts scripts lint current code base for the first time and fix possible javascript style guide violations
1
3,145
12,059,001,120
IssuesEvent
2020-04-15 18:28:37
arcticicestudio/igloo
https://api.github.com/repos/arcticicestudio/igloo
closed
“taskwarrior“ & “timewarrior“ snowblock decommission
scope-maintainability snowblock-taskwarrior snowblock-timewarrior type-task
Related to #248 --- Both _snowblocks_ for [Taskwarrior][] and [Timewarrior][] are not required anymore since they have been replaced with my own custom 💙 [Go][] application that is currently private/closed source, but planned to be open sourced later on. Both tools are great and provide a lot of features, but are also kind of overloaded with unused and unnecessary functions. I also missed the possibility to integrate the data and API into my other Go applications as well as web-based projects with a quite more modern _techstack_ ([Protocol Buffers][proto], [NATS][] Messaging, [React][] SPA etc.). Therefore the _snowblocks_ will be removed while the data is still available through the [_Git_ repository history/logs][git-docs-hist]. [git-docs-hist]: https://git-scm.com/book/en/v2/Git-Basics-Viewing-the-Commit-History [go]: https://go.dev [nats]: https://nats.io [proto]: https://developers.google.com/protocol-buffers [react]: https://reactjs.org [taskwarrior]: https://taskwarrior.org [timewarrior]: https://timewarrior.net
True
“taskwarrior“ & “timewarrior“ snowblock decommission - Related to #248 --- Both _snowblocks_ for [Taskwarrior][] and [Timewarrior][] are not required anymore since they have been replaced with my own custom 💙 [Go][] application that is currently private/closed source, but planned to be open sourced later on. Both tools are great and provide a lot of features, but are also kind of overloaded with unused and unnecessary functions. I also missed the possibility to integrate the data and API into my other Go applications as well as web-based projects with a quite more modern _techstack_ ([Protocol Buffers][proto], [NATS][] Messaging, [React][] SPA etc.). Therefore the _snowblocks_ will be removed while the data is still available through the [_Git_ repository history/logs][git-docs-hist]. [git-docs-hist]: https://git-scm.com/book/en/v2/Git-Basics-Viewing-the-Commit-History [go]: https://go.dev [nats]: https://nats.io [proto]: https://developers.google.com/protocol-buffers [react]: https://reactjs.org [taskwarrior]: https://taskwarrior.org [timewarrior]: https://timewarrior.net
main
“taskwarrior“ “timewarrior“ snowblock decommission related to both snowblocks for and are not required anymore since they have been replaced with my own custom 💙 application that is currently private closed source but planned to be open sourced later on both tools are great and provide a lot of features but are also kind of overloaded with unused and unnecessary functions i also missed the possibility to integrate the data and api into my other go applications as well as web based projects with a quite more modern techstack messaging spa etc therefore the snowblocks will be removed while the data is still available through the
1
266,547
28,379,520,106
IssuesEvent
2023-04-13 01:04:29
AlexRogalskiy/AlexRogalskiy
https://api.github.com/repos/AlexRogalskiy/AlexRogalskiy
opened
CVE-2023-26964 (Medium) detected in h2-0.1.26.crate, hyper-0.12.35.crate
Mend: dependency security vulnerability
## CVE-2023-26964 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>h2-0.1.26.crate</b>, <b>hyper-0.12.35.crate</b></p></summary> <p> <details><summary><b>h2-0.1.26.crate</b></p></summary> <p>An HTTP/2.0 client and server</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/h2/0.1.26/download">https://crates.io/api/v1/crates/h2/0.1.26/download</a></p> <p> Dependency Hierarchy: - rss-1.9.0.crate (Root Library) - reqwest-0.9.24.crate - hyper-0.12.35.crate - :x: **h2-0.1.26.crate** (Vulnerable Library) </details> <details><summary><b>hyper-0.12.35.crate</b></p></summary> <p>A fast and correct HTTP library.</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/hyper/0.12.35/download">https://crates.io/api/v1/crates/hyper/0.12.35/download</a></p> <p> Dependency Hierarchy: - rss-1.9.0.crate (Root Library) - reqwest-0.9.24.crate - :x: **hyper-0.12.35.crate** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in hyper v0.13.7. h2-0.2.4 Stream stacking occurs when the H2 component processes HTTP2 RST_STREAM frames. As a result, the memory and CPU usage are high which can lead to a Denial of Service (DoS). <p>Publish Date: 2023-04-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26964>CVE-2023-26964</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-26964 (Medium) detected in h2-0.1.26.crate, hyper-0.12.35.crate - ## CVE-2023-26964 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>h2-0.1.26.crate</b>, <b>hyper-0.12.35.crate</b></p></summary> <p> <details><summary><b>h2-0.1.26.crate</b></p></summary> <p>An HTTP/2.0 client and server</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/h2/0.1.26/download">https://crates.io/api/v1/crates/h2/0.1.26/download</a></p> <p> Dependency Hierarchy: - rss-1.9.0.crate (Root Library) - reqwest-0.9.24.crate - hyper-0.12.35.crate - :x: **h2-0.1.26.crate** (Vulnerable Library) </details> <details><summary><b>hyper-0.12.35.crate</b></p></summary> <p>A fast and correct HTTP library.</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/hyper/0.12.35/download">https://crates.io/api/v1/crates/hyper/0.12.35/download</a></p> <p> Dependency Hierarchy: - rss-1.9.0.crate (Root Library) - reqwest-0.9.24.crate - :x: **hyper-0.12.35.crate** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in hyper v0.13.7. h2-0.2.4 Stream stacking occurs when the H2 component processes HTTP2 RST_STREAM frames. As a result, the memory and CPU usage are high which can lead to a Denial of Service (DoS). <p>Publish Date: 2023-04-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26964>CVE-2023-26964</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in crate hyper crate cve medium severity vulnerability vulnerable libraries crate hyper crate crate an http client and server library home page a href dependency hierarchy rss crate root library reqwest crate hyper crate x crate vulnerable library hyper crate a fast and correct http library library home page a href dependency hierarchy rss crate root library reqwest crate x hyper crate vulnerable library found in base branch master vulnerability details an issue was discovered in hyper stream stacking occurs when the component processes rst stream frames as a result the memory and cpu usage are high which can lead to a denial of service dos publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
269,553
28,960,216,928
IssuesEvent
2023-05-10 01:24:13
shahul01/nlk-socket-simple-demo
https://api.github.com/repos/shahul01/nlk-socket-simple-demo
opened
socket.io-4.5.4.tgz: 1 vulnerabilities (highest severity is: 6.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-4.5.4.tgz</b></p></summary> <p></p> <p>Path to dependency file: /server/package.json</p> <p>Path to vulnerable library: /server/node_modules/engine.io/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (socket.io version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-31125](https://www.mend.io/vulnerability-database/CVE-2023-31125) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | engine.io-6.2.1.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2023-31125</summary> ### Vulnerable Library - <b>engine.io-6.2.1.tgz</b></p> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-6.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-6.2.1.tgz</a></p> <p>Path to dependency file: /server/package.json</p> <p>Path to vulnerable library: /server/node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - socket.io-4.5.4.tgz (Root Library) - :x: **engine.io-6.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Engine.IO is the implementation of transport-based cross-browser/cross-device bi-directional communication layer for Socket.IO. An uncaught exception vulnerability was introduced in version 5.1.0 and included in version 4.1.0 of the `socket.io` parent package. Older versions are not impacted. A specially crafted HTTP request can trigger an uncaught exception on the Engine.IO server, thus killing the Node.js process. This impacts all the users of the `engine.io` package, including those who use depending packages like `socket.io`. This issue was fixed in version 6.4.2 of Engine.IO. There is no known workaround except upgrading to a safe version. <p>Publish Date: 2023-05-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-31125>CVE-2023-31125</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-31125">https://www.cve.org/CVERecord?id=CVE-2023-31125</a></p> <p>Release Date: 2023-05-08</p> <p>Fix Resolution: engine.io - 6.4.2</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
socket.io-4.5.4.tgz: 1 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-4.5.4.tgz</b></p></summary> <p></p> <p>Path to dependency file: /server/package.json</p> <p>Path to vulnerable library: /server/node_modules/engine.io/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (socket.io version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-31125](https://www.mend.io/vulnerability-database/CVE-2023-31125) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | engine.io-6.2.1.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2023-31125</summary> ### Vulnerable Library - <b>engine.io-6.2.1.tgz</b></p> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-6.2.1.tgz">https://registry.npmjs.org/engine.io/-/engine.io-6.2.1.tgz</a></p> <p>Path to dependency file: /server/package.json</p> <p>Path to vulnerable library: /server/node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - socket.io-4.5.4.tgz (Root Library) - :x: **engine.io-6.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Engine.IO is the implementation of transport-based cross-browser/cross-device bi-directional communication layer for Socket.IO. An uncaught exception vulnerability was introduced in version 5.1.0 and included in version 4.1.0 of the `socket.io` parent package. Older versions are not impacted. A specially crafted HTTP request can trigger an uncaught exception on the Engine.IO server, thus killing the Node.js process. This impacts all the users of the `engine.io` package, including those who use depending packages like `socket.io`. This issue was fixed in version 6.4.2 of Engine.IO. There is no known workaround except upgrading to a safe version. <p>Publish Date: 2023-05-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-31125>CVE-2023-31125</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-31125">https://www.cve.org/CVERecord?id=CVE-2023-31125</a></p> <p>Release Date: 2023-05-08</p> <p>Fix Resolution: engine.io - 6.4.2</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_main
socket io tgz vulnerabilities highest severity is vulnerable library socket io tgz path to dependency file server package json path to vulnerable library server node modules engine io package json vulnerabilities cve severity cvss dependency type fixed in socket io version remediation available medium engine io tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file server package json path to vulnerable library server node modules engine io package json dependency hierarchy socket io tgz root library x engine io tgz vulnerable library found in base branch master vulnerability details engine io is the implementation of transport based cross browser cross device bi directional communication layer for socket io an uncaught exception vulnerability was introduced in version and included in version of the socket io parent package older versions are not impacted a specially crafted http request can trigger an uncaught exception on the engine io server thus killing the node js process this impacts all the users of the engine io package including those who use depending packages like socket io this issue was fixed in version of engine io there is no known workaround except upgrading to a safe version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io step up your open source security game with mend
0
5,000
25,723,359,976
IssuesEvent
2022-12-07 15:06:40
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Table not found error
type: bug work: frontend status: ready restricted: maintainers
## Steps to reproduce 1. Navigate to http://localhost:8000/mathesar_tables/ 1. Click on "Library Management" to open schema with id `217` (or similar) 1. On the "Authors" table card, hover the "Go to Table" hyperlink. Observe that it uses schema `217` (or equivalent) in the URL. Good. 1. In the navigation header, open the "Choose a Schema" dropdown, and click on "public", navigating to http://localhost:8000/mathesar_tables/1/ 1. Observe that table cards now use schema id `1` in their URLs. Good. 1. Use the same schema switcher to navigate back to "Library Management". 1. Hover the "Go to Table" hyperlinks for various tables. 1. Expect these URLs to use schema `217` (or equivalent). 1. Instead observe the table URLs to use schema `1`. 1. Clicking on one of these URLs rightly displays an error message like: > Table with id 1267 not found.
True
Table not found error - ## Steps to reproduce 1. Navigate to http://localhost:8000/mathesar_tables/ 1. Click on "Library Management" to open schema with id `217` (or similar) 1. On the "Authors" table card, hover the "Go to Table" hyperlink. Observe that it uses schema `217` (or equivalent) in the URL. Good. 1. In the navigation header, open the "Choose a Schema" dropdown, and click on "public", navigating to http://localhost:8000/mathesar_tables/1/ 1. Observe that table cards now use schema id `1` in their URLs. Good. 1. Use the same schema switcher to navigate back to "Library Management". 1. Hover the "Go to Table" hyperlinks for various tables. 1. Expect these URLs to use schema `217` (or equivalent). 1. Instead observe the table URLs to use schema `1`. 1. Clicking on one of these URLs rightly displays an error message like: > Table with id 1267 not found.
main
table not found error steps to reproduce navigate to click on library management to open schema with id or similar on the authors table card hover the go to table hyperlink observe that it uses schema or equivalent in the url good in the navigation header open the choose a schema dropdown and click on public navigating to observe that table cards now use schema id in their urls good use the same schema switcher to navigate back to library management hover the go to table hyperlinks for various tables expect these urls to use schema or equivalent instead observe the table urls to use schema clicking on one of these urls rightly displays an error message like table with id not found
1
1,304
5,495,033,343
IssuesEvent
2017-03-15 02:18:33
Code4SocialGood/C4SG
https://api.github.com/repos/Code4SocialGood/C4SG
closed
Replace CSS Framework
Architecture Front-End
- **Currently using** Materialize CSS (http://materializecss.com/) + Angular2-Material (https://github.com/InfomediaLtd/angular2-materialize) - **Replace with** Angular/Material 2 (https://material.angular.io/)
1.0
Replace CSS Framework - - **Currently using** Materialize CSS (http://materializecss.com/) + Angular2-Material (https://github.com/InfomediaLtd/angular2-materialize) - **Replace with** Angular/Material 2 (https://material.angular.io/)
non_main
replace css framework currently using materialize css material replace with angular material
0
2,458
8,639,897,895
IssuesEvent
2018-11-23 22:30:02
F5OEO/rpitx
https://api.github.com/repos/F5OEO/rpitx
closed
why rpitx developed for transceiver?
V1 related (not maintained)
Hi,it is great,it is possible for developing it for receiver and transmitter tool. can anyone offer any idea?? best regards stackprogramer
True
why rpitx developed for transceiver? - Hi,it is great,it is possible for developing it for receiver and transmitter tool. can anyone offer any idea?? best regards stackprogramer
main
why rpitx developed for transceiver hi it is great it is possible for developing it for receiver and transmitter tool can anyone offer any idea best regards stackprogramer
1
357,417
25,176,382,458
IssuesEvent
2022-11-11 09:37:55
PangKuangWei/pe
https://api.github.com/repos/PangKuangWei/pe
opened
UG Bug: Grammar Mistake (Comma Usage for 2 independent clauses)
severity.VeryLow type.DocumentationBug
There is a missing comma before the word `but` shown in the screenshot below. There should be a comma here because these 2 parts are independent clauses. ![image.png](https://raw.githubusercontent.com/PangKuangWei/pe/main/files/241f7930-a85c-4c9b-a2ba-9785951a5ca9.png) <!--session: 1668152656976-1bbc2a95-284a-4da4-b139-83a160afbe25--> <!--Version: Web v3.4.4-->
1.0
UG Bug: Grammar Mistake (Comma Usage for 2 independent clauses) - There is a missing comma before the word `but` shown in the screenshot below. There should be a comma here because these 2 parts are independent clauses. ![image.png](https://raw.githubusercontent.com/PangKuangWei/pe/main/files/241f7930-a85c-4c9b-a2ba-9785951a5ca9.png) <!--session: 1668152656976-1bbc2a95-284a-4da4-b139-83a160afbe25--> <!--Version: Web v3.4.4-->
non_main
ug bug grammar mistake comma usage for independent clauses there is a missing comma before the word but shown in the screenshot below there should be a comma here because these parts are independent clauses
0
827,355
31,767,119,845
IssuesEvent
2023-09-12 09:26:42
oceanbase/odc
https://api.github.com/repos/oceanbase/odc
opened
[Bug]: When importing the file, the id is undefined and does not exist.
type-bug priority-medium
### ODC version odc421 ### OB version Oceanbase410 ### What happened? When importing the file, the id is undefined and does not exist. ![image](https://github.com/oceanbase/odc/assets/59468957/f7eb7e1d-d257-4b4b-a964-4d8bfe21e330) ### What did you expect to happen? Execute normally ### How can we reproduce it (as minimally and precisely as possible)? 1. Import the file in the import panel, uncheck Ignore the first line 2. Select the corresponding library and table 3. Select next 4. Go back to the previous step and check Ignore first line 5. Click again ### Anything else we need to know? _No response_ ### Cloud _No response_
1.0
[Bug]: When importing the file, the id is undefined and does not exist. - ### ODC version odc421 ### OB version Oceanbase410 ### What happened? When importing the file, the id is undefined and does not exist. ![image](https://github.com/oceanbase/odc/assets/59468957/f7eb7e1d-d257-4b4b-a964-4d8bfe21e330) ### What did you expect to happen? Execute normally ### How can we reproduce it (as minimally and precisely as possible)? 1. Import the file in the import panel, uncheck Ignore the first line 2. Select the corresponding library and table 3. Select next 4. Go back to the previous step and check Ignore first line 5. Click again ### Anything else we need to know? _No response_ ### Cloud _No response_
non_main
when importing the file the id is undefined and does not exist odc version ob version what happened when importing the file the id is undefined and does not exist what did you expect to happen execute normally how can we reproduce it as minimally and precisely as possible import the file in the import panel uncheck ignore the first line select the corresponding library and table select next go back to the previous step and check ignore first line click again anything else we need to know no response cloud no response
0
4,654
24,096,902,233
IssuesEvent
2022-09-19 19:37:04
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Running sam sync with esbuild does not include dependencies
stage/bug-repro maintainer/need-followup
### Description: Running sam sync with esbuild does not include the dependencies. Running sam deploy works. ### Steps to reproduce: 1. Include a dependency: ```json { "name": "hello_world", "main": "app.js", "dependencies": { "uuidv4": "^6.2.12" }, ... ``` 2. Include it in your app.ts ```ts import type { APIGatewayProxyHandler } from 'aws-lambda'; import { uuid } from 'uuidv4'; export const lambdaHandler: APIGatewayProxyHandler = async (event) => { try { console.log('event', event); return { body: JSON.stringify({ message: 'hello ' + uuid() + process.env.FLAVOR, }), statusCode: 200, }; } catch (err) { console.log(err); return { body: JSON.stringify({ message: 'some error happened', }), statusCode: 500, }; } }; ``` 3. Use esbuild as the BuildMethod: ```yaml HelloWorldFunction: Type: AWS::Serverless::Function # More info about Function Resource: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#awsserverlessfunction Properties: CodeUri: hello/ Handler: app.lambdaHandler Runtime: nodejs14.x Architectures: - x86_64 Environment: Variables: FLAVOR: !Ref FLAVOR Events: HelloWorld: Type: Api # More info about API Event Source: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#api Properties: Path: /hello Method: get Metadata: # Manage esbuild properties BuildMethod: esbuild BuildProperties: Minify: true Target: 'es2020' Sourcemap: true EntryPoints: - app.ts ``` 4. Run: ```console sam sync --stack-name ... --watch ``` ### Observed result: There is no error in the terminal: ```console 2022-03-03 16:05:33 - Waiting for stack create/update to complete CloudFormation events from stack operations ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- ResourceStatus ResourceType LogicalResourceId ResourceStatusReason ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- UPDATE_IN_PROGRESS AWS::CloudFormation::Stack test-app Transformation succeeded CREATE_IN_PROGRESS AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack - CREATE_IN_PROGRESS AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack Resource creation Initiated CREATE_COMPLETE AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack - UPDATE_IN_PROGRESS AWS::Lambda::Function HelloWorldFunction - UPDATE_COMPLETE AWS::Lambda::Function HelloWorldFunction - UPDATE_COMPLETE AWS::CloudFormation::Stack test-app - UPDATE_COMPLETE_CLEANUP_IN_PROGRESS AWS::CloudFormation::Stack test-app - ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- CloudFormation outputs from deployed stack ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Outputs ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Key HelloWorldFunctionIamRole Description Implicit IAM Role created for Hello World function Value arn:aws:iam::102512246328:role/test-app-HelloWorldFunctionRole-IFB6LCDCZFB8 Key HelloWorldApi Description API Gateway endpoint URL for Prod stage for Hello World function Value https://xfi7je9i27.execute-api.eu-north-1.amazonaws.com/Prod/hello/ Key HelloWorldFunction Description Hello World Lambda Function ARN Value arn:aws:lambda:eu-north-1:102512246328:function:test-app-HelloWorldFunction-mJ9kipF6iwEi ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Stack update succeeded. Sync infra completed. {'StackId': 'arn:aws:cloudformation:eu-north-1:102512246328:stack/test-app/0c378cc0-9960-11ec-9d59-0e1e6bf4f0ce', 'ResponseMetadata': {'RequestId': '041daa8a-5770-4086-b8d2-3ddf51d2c7f9', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amzn-requestid': '041daa8a-5770-4086-b8d2-3ddf51d2c7f9', 'content-type': 'text/xml', 'content-length': '379', 'date': 'Thu, 03 Mar 2022 15:05:29 GMT'}, 'RetryAttempts': 0}} Infra sync completed. ``` but in cloud watch i get this: ```json { "errorType": "Runtime.ImportModuleError", "errorMessage": "Error: Cannot find module 'uuidv4'\nRequire stack:\n- /var/task/app.js\n- /var/runtime/UserFunction.js\n- /var/runtime/index.js", "stack": [ "Runtime.ImportModuleError: Error: Cannot find module 'uuidv4'", "Require stack:", "- /var/task/app.js", "- /var/runtime/UserFunction.js", "- /var/runtime/index.js", " at _loadUserApp (/var/runtime/UserFunction.js:202:13)", " at Object.module.exports.load (/var/runtime/UserFunction.js:242:17)", " at Object.<anonymous> (/var/runtime/index.js:43:30)", " at Module._compile (internal/modules/cjs/loader.js:1085:14)", " at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)", " at Module.load (internal/modules/cjs/loader.js:950:32)", " at Function.Module._load (internal/modules/cjs/loader.js:790:12)", " at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:76:12)", " at internal/main/run_main_module.js:17:47" ] } ``` ### Expected result: Have the lambda function not crashing when using sam sync. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows x64 3. `sam --version`: 1.40.1 6. AWS region: eu-north-1
True
Running sam sync with esbuild does not include dependencies - ### Description: Running sam sync with esbuild does not include the dependencies. Running sam deploy works. ### Steps to reproduce: 1. Include a dependency: ```json { "name": "hello_world", "main": "app.js", "dependencies": { "uuidv4": "^6.2.12" }, ... ``` 2. Include it in your app.ts ```ts import type { APIGatewayProxyHandler } from 'aws-lambda'; import { uuid } from 'uuidv4'; export const lambdaHandler: APIGatewayProxyHandler = async (event) => { try { console.log('event', event); return { body: JSON.stringify({ message: 'hello ' + uuid() + process.env.FLAVOR, }), statusCode: 200, }; } catch (err) { console.log(err); return { body: JSON.stringify({ message: 'some error happened', }), statusCode: 500, }; } }; ``` 3. Use esbuild as the BuildMethod: ```yaml HelloWorldFunction: Type: AWS::Serverless::Function # More info about Function Resource: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#awsserverlessfunction Properties: CodeUri: hello/ Handler: app.lambdaHandler Runtime: nodejs14.x Architectures: - x86_64 Environment: Variables: FLAVOR: !Ref FLAVOR Events: HelloWorld: Type: Api # More info about API Event Source: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#api Properties: Path: /hello Method: get Metadata: # Manage esbuild properties BuildMethod: esbuild BuildProperties: Minify: true Target: 'es2020' Sourcemap: true EntryPoints: - app.ts ``` 4. Run: ```console sam sync --stack-name ... --watch ``` ### Observed result: There is no error in the terminal: ```console 2022-03-03 16:05:33 - Waiting for stack create/update to complete CloudFormation events from stack operations ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- ResourceStatus ResourceType LogicalResourceId ResourceStatusReason ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- UPDATE_IN_PROGRESS AWS::CloudFormation::Stack test-app Transformation succeeded CREATE_IN_PROGRESS AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack - CREATE_IN_PROGRESS AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack Resource creation Initiated CREATE_COMPLETE AWS::CloudFormation::Stack AwsSamAutoDependencyLayerNestedStack - UPDATE_IN_PROGRESS AWS::Lambda::Function HelloWorldFunction - UPDATE_COMPLETE AWS::Lambda::Function HelloWorldFunction - UPDATE_COMPLETE AWS::CloudFormation::Stack test-app - UPDATE_COMPLETE_CLEANUP_IN_PROGRESS AWS::CloudFormation::Stack test-app - ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- CloudFormation outputs from deployed stack ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Outputs ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Key HelloWorldFunctionIamRole Description Implicit IAM Role created for Hello World function Value arn:aws:iam::102512246328:role/test-app-HelloWorldFunctionRole-IFB6LCDCZFB8 Key HelloWorldApi Description API Gateway endpoint URL for Prod stage for Hello World function Value https://xfi7je9i27.execute-api.eu-north-1.amazonaws.com/Prod/hello/ Key HelloWorldFunction Description Hello World Lambda Function ARN Value arn:aws:lambda:eu-north-1:102512246328:function:test-app-HelloWorldFunction-mJ9kipF6iwEi ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Stack update succeeded. Sync infra completed. {'StackId': 'arn:aws:cloudformation:eu-north-1:102512246328:stack/test-app/0c378cc0-9960-11ec-9d59-0e1e6bf4f0ce', 'ResponseMetadata': {'RequestId': '041daa8a-5770-4086-b8d2-3ddf51d2c7f9', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amzn-requestid': '041daa8a-5770-4086-b8d2-3ddf51d2c7f9', 'content-type': 'text/xml', 'content-length': '379', 'date': 'Thu, 03 Mar 2022 15:05:29 GMT'}, 'RetryAttempts': 0}} Infra sync completed. ``` but in cloud watch i get this: ```json { "errorType": "Runtime.ImportModuleError", "errorMessage": "Error: Cannot find module 'uuidv4'\nRequire stack:\n- /var/task/app.js\n- /var/runtime/UserFunction.js\n- /var/runtime/index.js", "stack": [ "Runtime.ImportModuleError: Error: Cannot find module 'uuidv4'", "Require stack:", "- /var/task/app.js", "- /var/runtime/UserFunction.js", "- /var/runtime/index.js", " at _loadUserApp (/var/runtime/UserFunction.js:202:13)", " at Object.module.exports.load (/var/runtime/UserFunction.js:242:17)", " at Object.<anonymous> (/var/runtime/index.js:43:30)", " at Module._compile (internal/modules/cjs/loader.js:1085:14)", " at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)", " at Module.load (internal/modules/cjs/loader.js:950:32)", " at Function.Module._load (internal/modules/cjs/loader.js:790:12)", " at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:76:12)", " at internal/main/run_main_module.js:17:47" ] } ``` ### Expected result: Have the lambda function not crashing when using sam sync. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows x64 3. `sam --version`: 1.40.1 6. AWS region: eu-north-1
main
running sam sync with esbuild does not include dependencies description running sam sync with esbuild does not include the dependencies running sam deploy works steps to reproduce include a dependency json name hello world main app js dependencies include it in your app ts ts import type apigatewayproxyhandler from aws lambda import uuid from export const lambdahandler apigatewayproxyhandler async event try console log event event return body json stringify message hello uuid process env flavor statuscode catch err console log err return body json stringify message some error happened statuscode use esbuild as the buildmethod yaml helloworldfunction type aws serverless function more info about function resource properties codeuri hello handler app lambdahandler runtime x architectures environment variables flavor ref flavor events helloworld type api more info about api event source properties path hello method get metadata manage esbuild properties buildmethod esbuild buildproperties minify true target sourcemap true entrypoints app ts run console sam sync stack name watch observed result there is no error in the terminal console waiting for stack create update to complete cloudformation events from stack operations resourcestatus resourcetype logicalresourceid resourcestatusreason update in progress aws cloudformation stack test app transformation succeeded create in progress aws cloudformation stack awssamautodependencylayernestedstack create in progress aws cloudformation stack awssamautodependencylayernestedstack resource creation initiated create complete aws cloudformation stack awssamautodependencylayernestedstack update in progress aws lambda function helloworldfunction update complete aws lambda function helloworldfunction update complete aws cloudformation stack test app update complete cleanup in progress aws cloudformation stack test app cloudformation outputs from deployed stack outputs key helloworldfunctioniamrole description implicit iam role created for hello world function value arn aws iam role test app helloworldfunctionrole key helloworldapi description api gateway endpoint url for prod stage for hello world function value key helloworldfunction description hello world lambda function arn value arn aws lambda eu north function test app helloworldfunction stack update succeeded sync infra completed stackid arn aws cloudformation eu north stack test app responsemetadata requestid httpstatuscode httpheaders x amzn requestid content type text xml content length date thu mar gmt retryattempts infra sync completed but in cloud watch i get this json errortype runtime importmoduleerror errormessage error cannot find module nrequire stack n var task app js n var runtime userfunction js n var runtime index js stack runtime importmoduleerror error cannot find module require stack var task app js var runtime userfunction js var runtime index js at loaduserapp var runtime userfunction js at object module exports load var runtime userfunction js at object var runtime index js at module compile internal modules cjs loader js at object module extensions js internal modules cjs loader js at module load internal modules cjs loader js at function module load internal modules cjs loader js at function executeuserentrypoint internal modules run main js at internal main run main module js expected result have the lambda function not crashing when using sam sync additional environment details ex windows mac amazon linux etc os windows sam version aws region eu north
1
4,083
19,285,831,764
IssuesEvent
2021-12-11 00:45:43
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Multiple queryStringParameters ignored in sam local start-api for HTTP APIs
area/local/start-api stage/needs-investigation stage/bug-repro maintainer/need-followup
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- Briefly describe the bug you are facing.--> I have a simple lambda function which takes multiple query string parameters, however only the last query string parameter is kept. ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> I have defined a function as follows: ```yaml TestFunction: Type: AWS::Serverless::Function Properties: CodeUri: rest_api/test/ Events: GetTest: Type: HttpApi Properties: Path: /api/test Method: get ``` Then run the application with `sam local start-api --debug` Then sent a query to it using postman as a GET request: http://localhost:3000/api/test?test_ids=1&test_ids=2 ### Observed result: <!-- Please provide command output with `--debug` flag set. --> The event output for the function is the following ```json Constructed String representation of Event Version 2.0 to invoke Lambda. Event: { "version": "2.0", "routeKey": "GET /api/test", "rawPath": "/api/test", "rawQueryString": "test_ids=1&test_ids=2", "cookies": [], "headers": { "Content-Type": "application/json", "User-Agent": "PostmanRuntime/7.28.3", "Accept": "*/*", "Postman-Token": "cb0df40e-920d-4f8d-bbf8-9381f9ee5e45", "Host": "localhost:3000", "Accept-Encoding": "gzip, deflate, br", "Connection": "keep-alive", "Content-Length": "21", "X-Forwarded-Proto": "http", "X-Forwarded-Port": "3000" }, "queryStringParameters": { "test_ids": "2" }, "requestContext": { "accountId": "123456789012", "apiId": "1234567890", "http": { "method": "GET", "path": "/api/test", "protocol": "HTTP/1.1", "sourceIp": "127.0.0.1", "userAgent": "Custom User Agent String" }, "requestId": "46b22e74-8e33-4f90-8f5c-3af887356c8e", "routeKey": "GET /api/test", "stage": null }, "body": "", "pathParameters": {}, "stageVariables": null, "isBase64Encoded": false } ``` Critically: ```json "queryStringParameters": { "test_ids": "2" }, ``` ### Expected result: <!-- Describe what you expected. --> I expected the following parsing for the queryStringParameters, as per the documentation ```json "queryStringParameters": { "test_ids": "1,2" }, ``` ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 1.29.0 3. AWS region: ap-southeast-2 `Add --debug flag to command you are running`
True
Multiple queryStringParameters ignored in sam local start-api for HTTP APIs - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- Briefly describe the bug you are facing.--> I have a simple lambda function which takes multiple query string parameters, however only the last query string parameter is kept. ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> I have defined a function as follows: ```yaml TestFunction: Type: AWS::Serverless::Function Properties: CodeUri: rest_api/test/ Events: GetTest: Type: HttpApi Properties: Path: /api/test Method: get ``` Then run the application with `sam local start-api --debug` Then sent a query to it using postman as a GET request: http://localhost:3000/api/test?test_ids=1&test_ids=2 ### Observed result: <!-- Please provide command output with `--debug` flag set. --> The event output for the function is the following ```json Constructed String representation of Event Version 2.0 to invoke Lambda. Event: { "version": "2.0", "routeKey": "GET /api/test", "rawPath": "/api/test", "rawQueryString": "test_ids=1&test_ids=2", "cookies": [], "headers": { "Content-Type": "application/json", "User-Agent": "PostmanRuntime/7.28.3", "Accept": "*/*", "Postman-Token": "cb0df40e-920d-4f8d-bbf8-9381f9ee5e45", "Host": "localhost:3000", "Accept-Encoding": "gzip, deflate, br", "Connection": "keep-alive", "Content-Length": "21", "X-Forwarded-Proto": "http", "X-Forwarded-Port": "3000" }, "queryStringParameters": { "test_ids": "2" }, "requestContext": { "accountId": "123456789012", "apiId": "1234567890", "http": { "method": "GET", "path": "/api/test", "protocol": "HTTP/1.1", "sourceIp": "127.0.0.1", "userAgent": "Custom User Agent String" }, "requestId": "46b22e74-8e33-4f90-8f5c-3af887356c8e", "routeKey": "GET /api/test", "stage": null }, "body": "", "pathParameters": {}, "stageVariables": null, "isBase64Encoded": false } ``` Critically: ```json "queryStringParameters": { "test_ids": "2" }, ``` ### Expected result: <!-- Describe what you expected. --> I expected the following parsing for the queryStringParameters, as per the documentation ```json "queryStringParameters": { "test_ids": "1,2" }, ``` ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 1.29.0 3. AWS region: ap-southeast-2 `Add --debug flag to command you are running`
main
multiple querystringparameters ignored in sam local start api for http apis make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description i have a simple lambda function which takes multiple query string parameters however only the last query string parameter is kept steps to reproduce i have defined a function as follows yaml testfunction type aws serverless function properties codeuri rest api test events gettest type httpapi properties path api test method get then run the application with sam local start api debug then sent a query to it using postman as a get request observed result the event output for the function is the following json constructed string representation of event version to invoke lambda event version routekey get api test rawpath api test rawquerystring test ids test ids cookies headers content type application json user agent postmanruntime accept postman token host localhost accept encoding gzip deflate br connection keep alive content length x forwarded proto http x forwarded port querystringparameters test ids requestcontext accountid apiid http method get path api test protocol http sourceip useragent custom user agent string requestid routekey get api test stage null body pathparameters stagevariables null false critically json querystringparameters test ids expected result i expected the following parsing for the querystringparameters as per the documentation json querystringparameters test ids additional environment details ex windows mac amazon linux etc os windows sam version aws region ap southeast add debug flag to command you are running
1
41,914
6,953,504,736
IssuesEvent
2017-12-06 21:16:54
prometheus/prometheus
https://api.github.com/repos/prometheus/prometheus
closed
Integrations link broken
component/documentation kind/bug low hanging fruit
On the Prometheus configuration page in the file discovery section: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#file_sd_config There's a link to integrations that resolves to: https://github.com/prometheus/prometheus/blob/release-2.0/docs/operating/configuration/#%3Cfile_sd_config It should resolve to: https://prometheus.io/docs/operating/integrations/#file-service-discovery In configuration.md it looks like it should be right. ```markdown There is a list of [integrations](/docs/operating/integrations/#file-service-discovery) with this discovery mechanism. ``` Perhaps needs to be rebuilt?
1.0
Integrations link broken - On the Prometheus configuration page in the file discovery section: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#file_sd_config There's a link to integrations that resolves to: https://github.com/prometheus/prometheus/blob/release-2.0/docs/operating/configuration/#%3Cfile_sd_config It should resolve to: https://prometheus.io/docs/operating/integrations/#file-service-discovery In configuration.md it looks like it should be right. ```markdown There is a list of [integrations](/docs/operating/integrations/#file-service-discovery) with this discovery mechanism. ``` Perhaps needs to be rebuilt?
non_main
integrations link broken on the prometheus configuration page in the file discovery section there s a link to integrations that resolves to it should resolve to in configuration md it looks like it should be right markdown there is a list of docs operating integrations file service discovery with this discovery mechanism perhaps needs to be rebuilt
0
378,103
11,196,184,260
IssuesEvent
2020-01-03 09:22:09
centreon/centreon
https://api.github.com/repos/centreon/centreon
closed
[Login Form] Only ok in private mode
area/authentication kind/bug priority/major status/more-info-needed
# BUG REPORT INFORMATION ### Prerequisites > The opened issue, must be code related. GitHub is not meant for support. Feel free to check the CONTRIBUTING section for more details. ***Versions*** For the RPM based systems -- Copy/Paste the result of the following command -- $ rpm -qa | grep centreon centreon-plugin-Network-Loadbalancers-F5-Bigip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hitachi-Hnas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Aerohive-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Blade-Chassis-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oracle-Infiniband-Snmp-20190704-123459.el7.centos.noarch centreon-widget-engine-status-19.04.0-5.el7.centos.noarch centreon-plugin-Operatingsystems-Windows-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Discovery-Nmap-20181019-1.el7.noarch centreon-plugin-Cloud-Ovh-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Alcatel-Omniswitch-6850-20170329-1.el7.noarch centreon-plugin-Applications-Databases-Elasticsearch-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Linux-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Elb-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Poller-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Tomcat-Webmanager-20190704-123459.el7.centos.noarch centreon-19.04.3-6.el7.centos.noarch centreon-plugin-Hardware-Storage-Hitachi-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Mag-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Waas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Quantum-Scalar-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Brocade-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Trapeze-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Paloalto-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-20190704-123459.el7.centos.noarch centreon-plugin-Network-Beeware-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Xtremio-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Tcp-20190704-123459.el7.centos.noarch centreon-common-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Radius-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Windows-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Solr-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Map4-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Udp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Lenovo-Sseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Audiocodes-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Aix-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-H3c-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Citrix-Appacceleration-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Lefthand-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Pfsense-Snmp-20190704-123459.el7.centos.noarch centreon-connector-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Sensors-Netbotz-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Jboss-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Ilo-Xmlapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ds5000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Sybase-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Elasticache-Api-20190704-123459.el7.centos.noarch centreon-plugin-Network-Generic-Bluecoat-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Nortel-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Radware-Alteon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Netapp-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Netscaler-Mpx8000-Snmp-20170120-1.el7.noarch centreon-plugin-Network-Hp-Vc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ats-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Snmp-20190704-123459.el7.centos.noarch centreon-broker-storage-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Supermicro-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Raisecom-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Firebird-20190704-123459.el7.centos.noarch centreon-auto-discovery-server-19.04.0-4.el7.centos.x86_64 centreon-plugin-Notification-Highsms-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Storeonce-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Fujitsu-Eternus-Dx-Ssh-20190704-123459.el7.centos.noarch centreon-database-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Ssh-20190704-123459.el7.centos.noarch centreon-poller-centreon-engine-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Databases-Informix-20190704-123459.el7.centos.noarch centreon-plugin-Network-Sophos-Es-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Overland-Neo-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dell-N4000-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Compellent-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Aruba-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Alcatel-Omniswitch-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Watchguard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Hpux-Snmp-20190704-123459.el7.centos.noarch centreon-broker-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Switchs-Cisco-Smallbusiness-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Jacarta-Snmp-20190704-123459.el7.centos.noarch centreon-widget-live-top10-memory-usage-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Stormshield-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Weblogic-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Php-Apc-Web-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Apache-Serverstatus-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netasq-Snmp-20170807-1.el7.noarch centreon-plugin-Applications-Redis-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Database-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-A10-AX-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Msl-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Panzura-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Arkoon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Antivirus-Kaspersky-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Device-Safenet-Keysecure-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Ironport-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Wlc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Violin-3000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Devices-Gorgy-Ntpserver-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Ssg-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-broker-cbd-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Ibm-Ds3000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Storagetek-Sl-Snmp-20190704-123459.el7.centos.noarch centreon-cwrapper-perl-0.1-1.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Dell-Equallogic-Snmp-20190704-123459.el7.centos.noarch centreon-widget-live-top10-cpu-usage-19.04.0-5.el7.centos.noarch centreon-plugin-Applications-Protocol-Imap-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Storwize-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Docker-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Freebox-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Vtom-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Billing-Api-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Printers-Generic-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Selenium-20190704-123459.el7.centos.noarch centreon-engine-extcommands-19.04.1-9.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Hp-Storeonce-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-S3-Api-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fritzbox-20190704-123459.el7.centos.noarch centreon-base-config-centreon-engine-19.04.3-6.el7.centos.noarch centreon-plugin-Network-Sonus-SBC-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Efficientip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Isg-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Juniper-Ex-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Fs900-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Riverbed-Steelhead-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Generic-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dlink-Dgs3100-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Nutanix-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Checkpoint-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Zyxel-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Evertz-FC7800-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Haproxy-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Polycom-Rmx-Snmp-20190704-123459.el7.centos.noarch centreon-broker-core-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Dell-Openmanage-Snmp-20190704-123459.el7.centos.noarch centreon-broker-cbmod-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Juniper-Ggsn-Snmp-20190704-123459.el7.centos.noarch centreon-license-manager-common-19.04.0-1.el7.centos.noarch centreon-plugin-Applications-Protocol-Bgp-Snmp-20190704-123459.el7.centos.noarch centreon-license-manager-19.04.0-1.el7.centos.noarch centreon-plugin-Applications-Lmsensors-Snmp-20190704-123459.el7.centos.noarch centreon-widget-grid-map-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Ucopia-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Mssql-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Hibernate-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Mgmtcards-20190704-123459.el7.centos.noarch centreon-plugins-base-1.18-2.el7.centos.noarch centreon-plugin-Virtualization-VMWare-daemon-3.0.3-20190704131142.el7.centos.noarch centreon-plugin-Hardware-Sensors-Sensormetrix-Em01-Web-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Lync-2013-Mssql-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Celerra-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Modbus-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Ec2-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Http-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Dhcp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-IBM-Tsm-20190704-123459.el7.centos.noarch centreon-web-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Github-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3100-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Stormshield-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Ml6000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-AnywhereUSB-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-TrendMicro-Iwsva-20190704-123459.el7.centos.noarch centreon-plugin-Network-Ruggedcom-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Redback-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Hwgste-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Kvm-Avocent-Acs-6000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Srx-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Telephony-Alcatel-OXE-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ospf-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Serverscheck-Sensorgateway-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Synology-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Bgp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Kemp-Snmp-20190704-123459.el7.centos.noarch centreon-connector-perl-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Dell-Sseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Nokia-Timos-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Mge-Snmp-20190704-123459.el7.centos.noarch centreon-widget-tactical-overview-19.04.0-7.el7.centos.noarch centreon-plugin-Applications-Zookeeper-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Cloudfront-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Postgresql-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Vmware2-Esx-Wsman-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-P2000-Xmlapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netasq-Ssh-20170811-1.el7.noarch centreon-plugin-Cloud-Aws-Cloudwatch-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-X509-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Lambda-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Dns-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Vplex-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Video-Zixi-Restapi-20190704-123459.el7.centos.noarch centreon-widget-host-monitoring-19.04.1-1.el7.centos.noarch centreon-plugin-Network-Firewalls-Cisco-Asa-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Fortinet-Fortigate-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Ruckus-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Hirschmann-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oneaccess-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3200-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Sendmail-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Exagrid-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Freebsd-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Sa-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Video-Openheadend-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Stonesoft-20190704-123459.el7.centos.noarch centreon-release-19.04-1.el7.centos.noarch centreon-plugin-Network-Cisco-Meraki-Snmp-20190704-123459.el7.centos.noarch centreon-connector-ssh-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Cisco-Voice-Gateway-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-3par-7000-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Md3000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Clamav-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Vmware2-Connector-Plugin-20190704-123459.el7.centos.noarch centreon-engine-19.04.1-9.el7.centos.x86_64 centreon-plugin-Applications-Pacemaker-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Mysql-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netgear-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Dell-Cmc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Datadomain-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-Sarian-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Dell-TL2000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Huawei-Snmp-20190704-123459.el7.centos.noarch centreon-clib-19.04.0-1.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Dell-IDrac-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Powerware-Snmp-20190704-123459.el7.centos.noarch centreon-widget-hostgroup-monitoring-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Extreme-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Peoplesoft-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Sap-Hana-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Clariion-Navisphere-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-Ami-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Smtp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Jenkins-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Qsan-Nas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Acmepacket-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Juniper-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-3com-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Ibm-Bladecenter-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Oracle-Zs-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Netscaler-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Sensorip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Msa2000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dlink-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-widget-servicegroup-monitoring-19.04.0-7.el7.centos.noarch centreon-plugin-Applications-Redis-Cli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Tomcat-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Netapp-Restapi-20190704-123459.el7.centos.noarch centreon-trap-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Telnet-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ldap-20190704-123459.el7.centos.noarch centreon-widget-service-monitoring-19.04.1-1.el7.centos.noarch centreon-plugin-Hardware-Sensors-Akcp-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Nimble-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dell-6200-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Qnap-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fortinet-Fortimanager-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Isilon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Atrica-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-Portserverts-Snmp-20190704-123459.el7.centos.noarch centreon-pp-manager-19.04.0-3.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts2900-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Oracle-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fritzbox-Snmp-20171013-1.el7.noarch centreon-plugin-Applications-Php-Fpm-Web-20190704-123459.el7.centos.noarch centreon-engine-daemon-19.04.1-9.el7.centos.x86_64 centreon-plugin-Applications-Mail-Bluemind-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Backup-Emc-Recoverypoint-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ftp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Eva-Cli-20190704-123459.el7.centos.noarch centreon-plugin-Network-Citrix-Sdx-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Fluidfs-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3500-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Raritan-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cyberoam-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Emerson-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-servers-ibm-imm-snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ds4000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Cisco-Ucs-Snmp-20190704-123459.el7.centos.noarch centreon-widget-graph-monitoring-19.04.0-5.el7.centos.noarch centreon-plugin-Applications-Biztalk-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Rds-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ntp-20190704-123459.el7.centos.noarch centreon-perl-libs-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Webservers-Nginx-Serverstatus-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Standard-Rfc1628-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Sfxxk-Pssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Central-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Prime-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oracle-Otd-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Clever-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Solaris-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Hp-Procurve-Snmp-20190704-123459.el7.centos.noarch ``` ***Operating System*** CentOS ***Browser used*** - [x] Google Chrome - [x] Firefox - [ ] Internet Explorer IE11 - [ ] Safari Version: -- 19.04.3 OVA Image from your website, on ESXi VMware 6.0 ### Description With the rights credentials, the login form returns the following error (see screens) "The form has not been submitted since 15 minutes. Please retry to resubmit here" ### Steps to Reproduce Browser in "normal" mode with or without adblocker 1. I logged in Centreon ### Describe the received result "The form has not been submitted since 15 minutes. Please retry to resubmit here" ### Describe the expected result Reach the Custom View ### Logs **PHP error logs** ``` tail -f /var/opt/rh/rh-php71/log/php-fpm/centreon-error.log [10-Aug-2019 09:55:52 Europe/Paris] PHP Notice: Undefined index: x-centreon-token in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 99 [10-Aug-2019 09:55:52 Europe/Paris] PHP Warning: in_array() expects parameter 2 to be array, null given in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 99 [10-Aug-2019 09:55:52 Europe/Paris] PHP Notice: Undefined index: x-centreon-token-generated-at in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 132 [10-Aug-2019 09:55:52 Europe/Paris] PHP Warning: Invalid argument supplied for foreach() in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 132 ### Additional relevant information (e.g. frequency, ...) The problem is resolved if i use the "pritate mode" with Firefox or Chrome. ![Capture](https://user-images.githubusercontent.com/2278299/62819430-8d5ff080-bb55-11e9-9f79-2a71d0e9a696.JPG) ![Capture-2](https://user-images.githubusercontent.com/2278299/62819431-8df88700-bb55-11e9-8375-1cf71e7edd99.JPG)
1.0
[Login Form] Only ok in private mode - # BUG REPORT INFORMATION ### Prerequisites > The opened issue, must be code related. GitHub is not meant for support. Feel free to check the CONTRIBUTING section for more details. ***Versions*** For the RPM based systems -- Copy/Paste the result of the following command -- $ rpm -qa | grep centreon centreon-plugin-Network-Loadbalancers-F5-Bigip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hitachi-Hnas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Aerohive-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Blade-Chassis-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oracle-Infiniband-Snmp-20190704-123459.el7.centos.noarch centreon-widget-engine-status-19.04.0-5.el7.centos.noarch centreon-plugin-Operatingsystems-Windows-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Discovery-Nmap-20181019-1.el7.noarch centreon-plugin-Cloud-Ovh-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Alcatel-Omniswitch-6850-20170329-1.el7.noarch centreon-plugin-Applications-Databases-Elasticsearch-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Linux-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Elb-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Poller-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Tomcat-Webmanager-20190704-123459.el7.centos.noarch centreon-19.04.3-6.el7.centos.noarch centreon-plugin-Hardware-Storage-Hitachi-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Mag-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Waas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Quantum-Scalar-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Brocade-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Trapeze-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Paloalto-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-20190704-123459.el7.centos.noarch centreon-plugin-Network-Beeware-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Xtremio-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Tcp-20190704-123459.el7.centos.noarch centreon-common-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Radius-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Windows-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Solr-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Map4-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Udp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Lenovo-Sseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Audiocodes-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Aix-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-H3c-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Citrix-Appacceleration-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Lefthand-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Pfsense-Snmp-20190704-123459.el7.centos.noarch centreon-connector-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Sensors-Netbotz-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Jboss-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Ilo-Xmlapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ds5000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Sybase-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Elasticache-Api-20190704-123459.el7.centos.noarch centreon-plugin-Network-Generic-Bluecoat-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Nortel-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Radware-Alteon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Netapp-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Netscaler-Mpx8000-Snmp-20170120-1.el7.noarch centreon-plugin-Network-Hp-Vc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ats-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Hp-Snmp-20190704-123459.el7.centos.noarch centreon-broker-storage-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Supermicro-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Raisecom-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Apc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Firebird-20190704-123459.el7.centos.noarch centreon-auto-discovery-server-19.04.0-4.el7.centos.x86_64 centreon-plugin-Notification-Highsms-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Storeonce-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Fujitsu-Eternus-Dx-Ssh-20190704-123459.el7.centos.noarch centreon-database-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Ssh-20190704-123459.el7.centos.noarch centreon-poller-centreon-engine-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Databases-Informix-20190704-123459.el7.centos.noarch centreon-plugin-Network-Sophos-Es-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Overland-Neo-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dell-N4000-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Compellent-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Aruba-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Alcatel-Omniswitch-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Watchguard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Hpux-Snmp-20190704-123459.el7.centos.noarch centreon-broker-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Switchs-Cisco-Smallbusiness-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Jacarta-Snmp-20190704-123459.el7.centos.noarch centreon-widget-live-top10-memory-usage-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Stormshield-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Weblogic-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Php-Apc-Web-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Apache-Serverstatus-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netasq-Snmp-20170807-1.el7.noarch centreon-plugin-Applications-Redis-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Database-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-A10-AX-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Msl-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Panzura-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Arkoon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Antivirus-Kaspersky-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Device-Safenet-Keysecure-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Ironport-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Wlc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Violin-3000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Devices-Gorgy-Ntpserver-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Ssg-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-broker-cbd-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Ibm-Ds3000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Storagetek-Sl-Snmp-20190704-123459.el7.centos.noarch centreon-cwrapper-perl-0.1-1.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Dell-Equallogic-Snmp-20190704-123459.el7.centos.noarch centreon-widget-live-top10-cpu-usage-19.04.0-5.el7.centos.noarch centreon-plugin-Applications-Protocol-Imap-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Storwize-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Docker-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Freebox-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Vtom-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Billing-Api-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Printers-Generic-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Selenium-20190704-123459.el7.centos.noarch centreon-engine-extcommands-19.04.1-9.el7.centos.x86_64 centreon-plugin-Hardware-Storage-Hp-Storeonce-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-S3-Api-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fritzbox-20190704-123459.el7.centos.noarch centreon-base-config-centreon-engine-19.04.3-6.el7.centos.noarch centreon-plugin-Network-Sonus-SBC-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Efficientip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Juniper-Isg-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Juniper-Ex-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Fs900-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Riverbed-Steelhead-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Generic-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dlink-Dgs3100-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Nutanix-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Checkpoint-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Zyxel-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Evertz-FC7800-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Haproxy-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Polycom-Rmx-Snmp-20190704-123459.el7.centos.noarch centreon-broker-core-19.04.0-2.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Dell-Openmanage-Snmp-20190704-123459.el7.centos.noarch centreon-broker-cbmod-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Juniper-Ggsn-Snmp-20190704-123459.el7.centos.noarch centreon-license-manager-common-19.04.0-1.el7.centos.noarch centreon-plugin-Applications-Protocol-Bgp-Snmp-20190704-123459.el7.centos.noarch centreon-license-manager-19.04.0-1.el7.centos.noarch centreon-plugin-Applications-Lmsensors-Snmp-20190704-123459.el7.centos.noarch centreon-widget-grid-map-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Ucopia-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Mssql-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Hibernate-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Mgmtcards-20190704-123459.el7.centos.noarch centreon-plugins-base-1.18-2.el7.centos.noarch centreon-plugin-Virtualization-VMWare-daemon-3.0.3-20190704131142.el7.centos.noarch centreon-plugin-Hardware-Sensors-Sensormetrix-Em01-Web-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Lync-2013-Mssql-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Celerra-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Modbus-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Ec2-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Http-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Dhcp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-IBM-Tsm-20190704-123459.el7.centos.noarch centreon-web-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Github-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3100-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Stormshield-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Ml6000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-AnywhereUSB-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-TrendMicro-Iwsva-20190704-123459.el7.centos.noarch centreon-plugin-Network-Ruggedcom-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Redback-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Hwgste-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Kvm-Avocent-Acs-6000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Srx-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Telephony-Alcatel-OXE-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ospf-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Serverscheck-Sensorgateway-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Synology-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Bgp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Kemp-Snmp-20190704-123459.el7.centos.noarch centreon-connector-perl-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Dell-Sseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Nokia-Timos-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Mge-Snmp-20190704-123459.el7.centos.noarch centreon-widget-tactical-overview-19.04.0-7.el7.centos.noarch centreon-plugin-Applications-Zookeeper-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Cloudfront-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Postgresql-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Vmware2-Esx-Wsman-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-P2000-Xmlapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netasq-Ssh-20170811-1.el7.noarch centreon-plugin-Cloud-Aws-Cloudwatch-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-X509-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Lambda-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Dns-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Vplex-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Video-Zixi-Restapi-20190704-123459.el7.centos.noarch centreon-widget-host-monitoring-19.04.1-1.el7.centos.noarch centreon-plugin-Network-Firewalls-Cisco-Asa-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Fortinet-Fortigate-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Ruckus-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Hirschmann-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oneaccess-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3200-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Sendmail-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Exagrid-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Freebsd-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Juniper-Sa-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Video-Openheadend-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Firewalls-Stonesoft-20190704-123459.el7.centos.noarch centreon-release-19.04-1.el7.centos.noarch centreon-plugin-Network-Cisco-Meraki-Snmp-20190704-123459.el7.centos.noarch centreon-connector-ssh-19.04.0-2.el7.centos.x86_64 centreon-plugin-Network-Cisco-Voice-Gateway-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-3par-7000-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Md3000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Clamav-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Virtualization-Vmware2-Connector-Plugin-20190704-123459.el7.centos.noarch centreon-engine-19.04.1-9.el7.centos.x86_64 centreon-plugin-Applications-Pacemaker-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Mysql-20190704-123459.el7.centos.noarch centreon-plugin-Network-Netgear-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Dell-Cmc-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Datadomain-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-Sarian-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Dell-TL2000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Huawei-Snmp-20190704-123459.el7.centos.noarch centreon-clib-19.04.0-1.el7.centos.x86_64 centreon-plugin-Hardware-Servers-Dell-IDrac-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Powerware-Snmp-20190704-123459.el7.centos.noarch centreon-widget-hostgroup-monitoring-19.04.0-5.el7.centos.noarch centreon-plugin-Network-Extreme-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Peoplesoft-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Sap-Hana-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Clariion-Navisphere-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Voip-Asterisk-Ami-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Smtp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Jenkins-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Qsan-Nas-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Acmepacket-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Juniper-Mseries-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-3com-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Ibm-Bladecenter-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Oracle-Zs-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Loadbalancers-Netscaler-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Sensors-Sensorip-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Msa2000-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dlink-Standard-Snmp-20190704-123459.el7.centos.noarch centreon-widget-servicegroup-monitoring-19.04.0-7.el7.centos.noarch centreon-plugin-Applications-Redis-Cli-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Webservers-Tomcat-Jmx-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Netapp-Restapi-20190704-123459.el7.centos.noarch centreon-trap-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Protocol-Telnet-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ldap-20190704-123459.el7.centos.noarch centreon-widget-service-monitoring-19.04.1-1.el7.centos.noarch centreon-plugin-Hardware-Sensors-Akcp-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Nimble-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Dell-6200-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Qnap-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fortinet-Fortimanager-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Emc-Isilon-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Routers-Atrica-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Digi-Portserverts-Snmp-20190704-123459.el7.centos.noarch centreon-pp-manager-19.04.0-3.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts2900-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Databases-Oracle-20190704-123459.el7.centos.noarch centreon-plugin-Network-Fritzbox-Snmp-20171013-1.el7.noarch centreon-plugin-Applications-Php-Fpm-Web-20190704-123459.el7.centos.noarch centreon-engine-daemon-19.04.1-9.el7.centos.x86_64 centreon-plugin-Applications-Mail-Bluemind-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Backup-Emc-Recoverypoint-Ssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ftp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Hp-Eva-Cli-20190704-123459.el7.centos.noarch centreon-plugin-Network-Citrix-Sdx-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Dell-Fluidfs-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ts3500-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Raritan-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cyberoam-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Emerson-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-servers-ibm-imm-snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Storage-Ibm-Ds4000-Smcli-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Cisco-Ucs-Snmp-20190704-123459.el7.centos.noarch centreon-widget-graph-monitoring-19.04.0-5.el7.centos.noarch centreon-plugin-Applications-Biztalk-20190704-123459.el7.centos.noarch centreon-plugin-Cloud-Aws-Rds-Api-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Protocol-Ntp-20190704-123459.el7.centos.noarch centreon-perl-libs-19.04.3-6.el7.centos.noarch centreon-plugin-Applications-Webservers-Nginx-Serverstatus-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Ups-Standard-Rfc1628-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Servers-Sun-Sfxxk-Pssh-20190704-123459.el7.centos.noarch centreon-plugin-Applications-Monitoring-Centreon-Central-20190704-123459.el7.centos.noarch centreon-plugin-Network-Cisco-Prime-Restapi-20190704-123459.el7.centos.noarch centreon-plugin-Network-Oracle-Otd-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Hardware-Pdu-Clever-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Operatingsystems-Solaris-Snmp-20190704-123459.el7.centos.noarch centreon-plugin-Network-Switchs-Hp-Procurve-Snmp-20190704-123459.el7.centos.noarch ``` ***Operating System*** CentOS ***Browser used*** - [x] Google Chrome - [x] Firefox - [ ] Internet Explorer IE11 - [ ] Safari Version: -- 19.04.3 OVA Image from your website, on ESXi VMware 6.0 ### Description With the rights credentials, the login form returns the following error (see screens) "The form has not been submitted since 15 minutes. Please retry to resubmit here" ### Steps to Reproduce Browser in "normal" mode with or without adblocker 1. I logged in Centreon ### Describe the received result "The form has not been submitted since 15 minutes. Please retry to resubmit here" ### Describe the expected result Reach the Custom View ### Logs **PHP error logs** ``` tail -f /var/opt/rh/rh-php71/log/php-fpm/centreon-error.log [10-Aug-2019 09:55:52 Europe/Paris] PHP Notice: Undefined index: x-centreon-token in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 99 [10-Aug-2019 09:55:52 Europe/Paris] PHP Warning: in_array() expects parameter 2 to be array, null given in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 99 [10-Aug-2019 09:55:52 Europe/Paris] PHP Notice: Undefined index: x-centreon-token-generated-at in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 132 [10-Aug-2019 09:55:52 Europe/Paris] PHP Warning: Invalid argument supplied for foreach() in /usr/share/centreon/www/lib/HTML/QuickForm/HTML_QuickFormCustom.php on line 132 ### Additional relevant information (e.g. frequency, ...) The problem is resolved if i use the "pritate mode" with Firefox or Chrome. ![Capture](https://user-images.githubusercontent.com/2278299/62819430-8d5ff080-bb55-11e9-9f79-2a71d0e9a696.JPG) ![Capture-2](https://user-images.githubusercontent.com/2278299/62819431-8df88700-bb55-11e9-8375-1cf71e7edd99.JPG)
non_main
only ok in private mode bug report information prerequisites the opened issue must be code related github is not meant for support feel free to check the contributing section for more details versions for the rpm based systems copy paste the result of the following command rpm qa grep centreon centreon plugin network loadbalancers bigip snmp centos noarch centreon plugin hardware storage hitachi hnas snmp centos noarch centreon plugin network aerohive snmp centos noarch centreon plugin hardware servers hp blade chassis snmp centos noarch centreon plugin network oracle infiniband snmp centos noarch centreon widget engine status centos noarch centreon plugin operatingsystems windows restapi centos noarch centreon plugin discovery nmap noarch centreon plugin cloud ovh restapi centos noarch centreon plugin network switchs alcatel omniswitch noarch centreon plugin applications databases elasticsearch centos noarch centreon plugin operatingsystems linux snmp centos noarch centreon plugin cloud aws elb api centos noarch centreon plugin applications monitoring centreon poller centos noarch centreon plugin applications webservers tomcat webmanager centos noarch centreon centos noarch centreon plugin hardware storage hitachi standard snmp centos noarch centreon plugin network juniper mag snmp centos noarch centreon plugin applications voip asterisk snmp centos noarch centreon plugin network cisco waas snmp centos noarch centreon plugin hardware storage quantum scalar snmp centos noarch centreon plugin network switchs brocade snmp centos noarch centreon plugin network juniper trapeze snmp centos noarch centreon plugin network firewalls paloalto standard snmp centos noarch centreon plugin applications voip asterisk centos noarch centreon plugin network beeware snmp centos noarch centreon plugin hardware storage emc xtremio restapi centos noarch centreon plugin applications protocol tcp centos noarch centreon common centos noarch centreon plugin applications protocol radius centos noarch centreon plugin operatingsystems windows snmp centos noarch centreon plugin applications solr jmx centos noarch centreon plugin applications monitoring centreon jmx centos noarch centreon plugin applications protocol udp centos noarch centreon plugin hardware storage lenovo sseries snmp centos noarch centreon plugin network audiocodes snmp centos noarch centreon plugin operatingsystems aix snmp centos noarch centreon plugin hardware ups apc snmp centos noarch centreon plugin network snmp centos noarch centreon plugin network citrix appacceleration snmp centos noarch centreon plugin hardware storage hp lefthand snmp centos noarch centreon plugin network firewalls pfsense snmp centos noarch centreon connector centos centreon plugin hardware sensors netbotz snmp centos noarch centreon plugin applications jboss jmx centos noarch centreon plugin hardware servers hp ilo xmlapi centos noarch centreon plugin hardware storage ibm smcli centos noarch centreon plugin applications databases sybase centos noarch centreon plugin cloud aws elasticache api centos noarch centreon plugin network generic bluecoat snmp centos noarch centreon plugin network nortel standard snmp centos noarch centreon plugin network radware alteon snmp centos noarch centreon plugin hardware storage netapp snmp centos noarch centreon plugin network loadbalancers netscaler snmp noarch centreon plugin network hp vc snmp centos noarch centreon plugin hardware ats apc snmp centos noarch centreon plugin hardware servers hp snmp centos noarch centreon broker storage centos centreon plugin hardware servers supermicro snmp centos noarch centreon plugin network raisecom snmp centos noarch centreon plugin hardware pdu apc snmp centos noarch centreon plugin applications databases firebird centos noarch centreon auto discovery server centos centreon plugin notification highsms centos noarch centreon plugin hardware storage hp storeonce restapi centos noarch centreon plugin hardware storage fujitsu eternus dx ssh centos noarch centreon database centos noarch centreon plugin applications protocol ssh centos noarch centreon poller centreon engine centos noarch centreon plugin applications databases informix centos noarch centreon plugin network sophos es snmp centos noarch centreon plugin hardware storage overland neo snmp centos noarch centreon plugin network dell centos noarch centreon plugin hardware storage dell compellent snmp centos noarch centreon plugin network switchs aruba standard snmp centos noarch centreon plugin network switchs alcatel omniswitch snmp centos noarch centreon plugin network watchguard snmp centos noarch centreon plugin operatingsystems hpux snmp centos noarch centreon broker centos centreon plugin network switchs cisco smallbusiness standard snmp centos noarch centreon plugin hardware sensors jacarta snmp centos noarch centreon widget live memory usage centos noarch centreon plugin network stormshield ssh centos noarch centreon plugin applications weblogic jmx centos noarch centreon plugin applications php apc web centos noarch centreon plugin applications webservers apache serverstatus centos noarch centreon plugin network netasq snmp noarch centreon plugin applications redis restapi centos noarch centreon plugin applications monitoring centreon database centos noarch centreon plugin network loadbalancers ax snmp centos noarch centreon plugin hardware storage hp msl snmp centos noarch centreon plugin hardware storage panzura snmp centos noarch centreon plugin network firewalls arkoon snmp centos noarch centreon plugin applications antivirus kaspersky snmp centos noarch centreon plugin hardware device safenet keysecure snmp centos noarch centreon plugin network cisco ironport snmp centos noarch centreon plugin network cisco wlc snmp centos noarch centreon plugin hardware storage violin snmp centos noarch centreon plugin hardware devices gorgy ntpserver snmp centos noarch centreon plugin network firewalls juniper ssg snmp centos noarch centreon plugin hardware servers sun mseries snmp centos noarch centreon broker cbd centos centreon plugin hardware storage ibm smcli centos noarch centreon plugin hardware storage storagetek sl snmp centos noarch centreon cwrapper perl centos centreon plugin hardware storage dell equallogic snmp centos noarch centreon widget live cpu usage centos noarch centreon plugin applications protocol imap centos noarch centreon plugin hardware storage ibm storwize ssh centos noarch centreon plugin applications docker restapi centos noarch centreon plugin network freebox restapi centos noarch centreon plugin applications vtom restapi centos noarch centreon plugin cloud aws billing api centos noarch centreon plugin hardware printers generic snmp centos noarch centreon plugin applications selenium centos noarch centreon engine extcommands centos centreon plugin hardware storage hp storeonce ssh centos noarch centreon plugin cloud aws api centos noarch centreon plugin network fritzbox centos noarch centreon base config centreon engine centos noarch centreon plugin network sonus sbc snmp centos noarch centreon plugin network efficientip snmp centos noarch centreon plugin network juniper isg snmp centos noarch centreon plugin network switchs juniper ex snmp centos noarch centreon plugin hardware storage ibm snmp centos noarch centreon plugin network riverbed steelhead snmp centos noarch centreon plugin generic snmp centos noarch centreon plugin network dlink snmp centos noarch centreon plugin virtualization nutanix snmp centos noarch centreon plugin network firewalls checkpoint snmp centos noarch centreon plugin network zyxel snmp centos noarch centreon plugin network evertz snmp centos noarch centreon plugin applications haproxy snmp centos noarch centreon plugin network polycom rmx snmp centos noarch centreon broker core centos centreon plugin hardware servers dell openmanage snmp centos noarch centreon broker cbmod centos centreon plugin network juniper ggsn snmp centos noarch centreon license manager common centos noarch centreon plugin applications protocol bgp snmp centos noarch centreon license manager centos noarch centreon plugin applications lmsensors snmp centos noarch centreon widget grid map centos noarch centreon plugin network ucopia snmp centos noarch centreon plugin applications databases mssql centos noarch centreon plugin applications hibernate jmx centos noarch centreon plugin hardware servers sun mgmtcards centos noarch centreon plugins base centos noarch centreon plugin virtualization vmware daemon centos noarch centreon plugin hardware sensors sensormetrix web centos noarch centreon plugin applications lync mssql centos noarch centreon plugin hardware storage emc celerra ssh centos noarch centreon plugin applications protocol modbus centos noarch centreon plugin cloud aws api centos noarch centreon plugin applications protocol http centos noarch centreon plugin applications protocol dhcp centos noarch centreon plugin applications ibm tsm centos noarch centreon web centos noarch centreon plugin applications github restapi centos noarch centreon plugin hardware storage ibm snmp centos noarch centreon plugin network stormshield snmp centos noarch centreon plugin hardware storage dell snmp centos noarch centreon plugin network digi anywhereusb snmp centos noarch centreon plugin applications trendmicro iwsva centos noarch centreon plugin network ruggedcom centos noarch centreon plugin network routers redback snmp centos noarch centreon plugin hardware sensors hwgste snmp centos noarch centreon plugin hardware kvm avocent acs snmp centos noarch centreon plugin network firewalls juniper srx snmp centos noarch centreon plugin hardware telephony alcatel oxe snmp centos noarch centreon plugin applications protocol ospf centos noarch centreon plugin hardware sensors serverscheck sensorgateway snmp centos noarch centreon plugin hardware storage synology snmp centos noarch centreon plugin applications protocol bgp centos noarch centreon plugin network loadbalancers kemp snmp centos noarch centreon connector perl centos centreon plugin network dell sseries snmp centos noarch centreon plugin network nokia timos snmp centos noarch centreon plugin hardware ups mge snmp centos noarch centreon widget tactical overview centos noarch centreon plugin applications zookeeper jmx centos noarch centreon plugin cloud aws cloudfront api centos noarch centreon plugin applications databases postgresql centos noarch centreon plugin virtualization esx wsman centos noarch centreon plugin hardware storage hp xmlapi centos noarch centreon plugin network netasq ssh noarch centreon plugin cloud aws cloudwatch api centos noarch centreon plugin applications protocol centos noarch centreon plugin network cisco standard snmp centos noarch centreon plugin applications protocol jmx centos noarch centreon plugin cloud aws lambda api centos noarch centreon plugin applications protocol dns centos noarch centreon plugin hardware storage emc vplex restapi centos noarch centreon plugin applications video zixi restapi centos noarch centreon widget host monitoring centos noarch centreon plugin network firewalls cisco asa snmp centos noarch centreon plugin network firewalls fortinet fortigate snmp centos noarch centreon plugin network ruckus snmp centos noarch centreon plugin network switchs hirschmann centos noarch centreon plugin network oneaccess snmp centos noarch centreon plugin hardware storage ibm snmp centos noarch centreon plugin applications sendmail snmp centos noarch centreon plugin hardware storage exagrid snmp centos noarch centreon plugin operatingsystems freebsd snmp centos noarch centreon plugin network firewalls juniper sa snmp centos noarch centreon plugin applications video openheadend snmp centos noarch centreon plugin network firewalls stonesoft centos noarch centreon release centos noarch centreon plugin network cisco meraki snmp centos noarch centreon connector ssh centos centreon plugin network cisco voice gateway snmp centos noarch centreon plugin hardware storage hp ssh centos noarch centreon plugin hardware storage dell smcli centos noarch centreon plugin applications clamav ssh centos noarch centreon plugin virtualization connector plugin centos noarch centreon engine centos centreon plugin applications pacemaker ssh centos noarch centreon plugin applications databases mysql centos noarch centreon plugin network netgear mseries snmp centos noarch centreon plugin hardware servers dell cmc snmp centos noarch centreon plugin hardware storage emc datadomain snmp centos noarch centreon plugin network digi sarian snmp centos noarch centreon plugin hardware dell snmp centos noarch centreon plugin network huawei snmp centos noarch centreon clib centos centreon plugin hardware servers dell idrac snmp centos noarch centreon plugin hardware ups powerware snmp centos noarch centreon widget hostgroup monitoring centos noarch centreon plugin network extreme snmp centos noarch centreon plugin applications peoplesoft jmx centos noarch centreon plugin applications databases sap hana centos noarch centreon plugin hardware storage emc clariion navisphere centos noarch centreon plugin applications voip asterisk ami centos noarch centreon plugin applications protocol smtp centos noarch centreon plugin applications jenkins centos noarch centreon plugin hardware storage qsan nas snmp centos noarch centreon plugin network acmepacket snmp centos noarch centreon plugin network routers juniper mseries snmp centos noarch centreon plugin network snmp centos noarch centreon plugin hardware servers ibm bladecenter snmp centos noarch centreon plugin hardware storage oracle zs snmp centos noarch centreon plugin network loadbalancers netscaler snmp centos noarch centreon plugin hardware sensors sensorip snmp centos noarch centreon plugin hardware storage hp snmp centos noarch centreon plugin network dlink standard snmp centos noarch centreon widget servicegroup monitoring centos noarch centreon plugin applications redis cli centos noarch centreon plugin applications webservers tomcat jmx centos noarch centreon plugin hardware storage netapp restapi centos noarch centreon trap centos noarch centreon plugin applications protocol telnet centos noarch centreon plugin applications protocol ldap centos noarch centreon widget service monitoring centos noarch centreon plugin hardware sensors akcp snmp centos noarch centreon plugin hardware storage nimble snmp centos noarch centreon plugin network dell centos noarch centreon plugin hardware storage qnap snmp centos noarch centreon plugin network fortinet fortimanager snmp centos noarch centreon plugin hardware storage emc isilon snmp centos noarch centreon plugin network routers atrica snmp centos noarch centreon plugin network digi portserverts snmp centos noarch centreon pp manager centos noarch centreon plugin hardware storage ibm snmp centos noarch centreon plugin applications databases oracle centos noarch centreon plugin network fritzbox snmp noarch centreon plugin applications php fpm web centos noarch centreon engine daemon centos centreon plugin applications mail bluemind centos noarch centreon plugin applications backup emc recoverypoint ssh centos noarch centreon plugin applications protocol ftp centos noarch centreon plugin hardware storage hp eva cli centos noarch centreon plugin network citrix sdx snmp centos noarch centreon plugin hardware storage dell fluidfs snmp centos noarch centreon plugin hardware storage ibm snmp centos noarch centreon plugin hardware pdu raritan snmp centos noarch centreon plugin network cyberoam snmp centos noarch centreon plugin hardware pdu emerson snmp centos noarch centreon plugin hardware servers ibm imm snmp centos noarch centreon plugin hardware storage ibm smcli centos noarch centreon plugin hardware servers cisco ucs snmp centos noarch centreon widget graph monitoring centos noarch centreon plugin applications biztalk centos noarch centreon plugin cloud aws rds api centos noarch centreon plugin applications protocol ntp centos noarch centreon perl libs centos noarch centreon plugin applications webservers nginx serverstatus centos noarch centreon plugin hardware ups standard snmp centos noarch centreon plugin hardware servers sun sfxxk pssh centos noarch centreon plugin applications monitoring centreon central centos noarch centreon plugin network cisco prime restapi centos noarch centreon plugin network oracle otd snmp centos noarch centreon plugin hardware pdu clever snmp centos noarch centreon plugin operatingsystems solaris snmp centos noarch centreon plugin network switchs hp procurve snmp centos noarch operating system centos browser used google chrome firefox internet explorer safari version ova image from your website on esxi vmware description with the rights credentials the login form returns the following error see screens the form has not been submitted since minutes please retry to resubmit here steps to reproduce browser in normal mode with or without adblocker i logged in centreon describe the received result the form has not been submitted since minutes please retry to resubmit here describe the expected result reach the custom view logs php error logs tail f var opt rh rh log php fpm centreon error log php notice undefined index x centreon token in usr share centreon www lib html quickform html quickformcustom php on line php warning in array expects parameter to be array null given in usr share centreon www lib html quickform html quickformcustom php on line php notice undefined index x centreon token generated at in usr share centreon www lib html quickform html quickformcustom php on line php warning invalid argument supplied for foreach in usr share centreon www lib html quickform html quickformcustom php on line additional relevant information e g frequency the problem is resolved if i use the pritate mode with firefox or chrome
0
82,917
15,681,282,565
IssuesEvent
2021-03-25 04:59:14
AlexRogalskiy/typescript-tools
https://api.github.com/repos/AlexRogalskiy/typescript-tools
opened
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: typescript-tools/package.json</p> <p>Path to vulnerable library: typescript-tools/node_modules/hosted-git-info/package.json,typescript-tools/node_modules/npm/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - npm-7.0.10.tgz (Root Library) - npm-6.14.11.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/a8e1ca24ae35a815510b92405c93535d11987fa1">a8e1ca24ae35a815510b92405c93535d11987fa1</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl(). <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 3.0.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: typescript-tools/package.json</p> <p>Path to vulnerable library: typescript-tools/node_modules/hosted-git-info/package.json,typescript-tools/node_modules/npm/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - npm-7.0.10.tgz (Root Library) - npm-6.14.11.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/a8e1ca24ae35a815510b92405c93535d11987fa1">a8e1ca24ae35a815510b92405c93535d11987fa1</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl(). <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 3.0.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file typescript tools package json path to vulnerable library typescript tools node modules hosted git info package json typescript tools node modules npm node modules hosted git info package json dependency hierarchy npm tgz root library npm tgz x hosted git info tgz vulnerable library found in head commit a href vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via shortcutmatch in fromurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info step up your open source security game with whitesource
0
4,579
23,776,615,108
IssuesEvent
2022-09-01 21:41:42
pyOpenSci/software-review
https://api.github.com/repos/pyOpenSci/software-review
closed
hudpy: A Python interface for the US Department of Housing and Urban Development APIs
1/editor-checks ⌛ pending-maintainer-response New Submission!
Submitting Author: @etam4260 Package Name: hudpy One-Line Description of Package: A Python Interface to the US Department of Housing and Urban Development APIs Repository Link: https://github.com/etam4260/hudpy Version submitted: 0.2.0 Editor: TBD Reviewer 1: TBD Reviewer 2: TBD Archive: TBD Version accepted: TBD --- ## Description - Include a brief paragraph describing what your package does: Currently the R version is under peer review at ROpenSci: https://github.com/ropensci/software-review/issues/524 This package is a mapped over version from R to Python. This package gives users access to the US Department of Housing and Urban Development APIs available at HUD User: https://www.huduser.gov/hudapi/public/login This also has additional wrappers onto of the data provided from these APIs to determine if US geographies intersect based on addresses and (cross walking a US dataset) explained in this paper below: Wilson, Ron and Din, Alexander, 2018. “Understanding and Enhancing the U.S. Department of Housing and Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 20 Number 2, 277 https://www.huduser.gov/portal/periodicals/cityscpe/vol20num2/ch16.pdf ## Scope - Please indicate which [category or categories][PackageCategories] this package falls under: - [X] Data retrieval - [ ] Data extraction - [X] Data munging - [ ] Data deposition - [ ] Reproducibility - [X] Geospatial - [ ] Education - [ ] Data visualization* > *Please fill out a pre-submission inquiry before submitting a data visualization package. For more info, see [notes on categories][NotesOnCategories] of our guidebook.* NA - **For all submissions**, explain how the and why the package falls under the categories you indicated above. In your explanation, please address the following points (briefly, 1-2 sentences for each): It is a data retrieval package because it retrieves data from an API. It is data munging package because it implements the concept of cross walking a file (converting a dataset described in one US geographic identifier into another) explained in the paper mentioned above. - Who is the target audience and what are scientific applications of this package? - I am hoping to reach professors, researchers, and students with this package. This gives access to the crosswalk files which is a geospatial technique described very well in these journal articles: Din, Alexander and Wilson, Ron, 2020. “Crosswalking ZIP Codes to Census Geographies: Geoprocessing the U.S. Department of Housing & Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 22, Number 1, https://www.huduser.gov/portal/periodicals/cityscpe/vol22num1/ch12.pdf Wilson, Ron and Din, Alexander, 2018. “Understanding and Enhancing the U.S. Department of Housing and Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 20 Number 2, 277 – 294. - Are there other Python packages that accomplish the same thing? If so, how does yours differ? Not any that I know of. - If you made a pre-submission enquiry, please paste the link to the corresponding issue, forum post, or other discussion, or `@tag` the editor you contacted: NA ## Technical checks For details about the pyOpenSci packaging requirements, see our [packaging guide][PackagingGuide]. Confirm each of the following by checking the box. This package: - [X] does not violate the Terms of Service of any service it interacts with. - [X] has an [OSI approved license][OsiApprovedLicense]. - [X] contains a README with instructions for installing the development version. - [X] includes documentation with examples for all functions. - [X] contains a vignette with examples of its essential functions and uses. - [X] has a test suite. - [X] has continuous integration, such as Travis CI, AppVeyor, CircleCI, and/or others. In terms of vignette, I am assuming you are referring to a Jupyter notebook style documentation. I don't have this, but there is a website with vignette like documentation: https://etam4260.github.io/hudpy/build/html/index.html ## Publication options - [ ] Do you wish to automatically submit to the [Journal of Open Source Software][JournalOfOpenSourceSoftware]? If so: <details> <summary>JOSS Checks</summary> - [ ] The package has an **obvious research application** according to JOSS's definition in their [submission requirements][JossSubmissionRequirements]. Be aware that completing the pyOpenSci review process **does not** guarantee acceptance to JOSS. Be sure to read their submission requirements (linked above) if you are interested in submitting to JOSS. - [ ] The package is not a "minor utility" as defined by JOSS's [submission requirements][JossSubmissionRequirements]: "Minor ‘utility’ packages, including ‘thin’ API clients, are not acceptable." pyOpenSci welcomes these packages under "Data Retrieval", but JOSS has slightly different criteria. - [ ] The package contains a `paper.md` matching [JOSS's requirements][JossPaperRequirements] with a high-level description in the package root or in `inst/`. - [ ] The package is deposited in a long-term repository with the DOI: *Note: Do not submit your package separately to JOSS* </details> ## Are you OK with Reviewers Submitting Issues and/or pull requests to your Repo Directly? This option will allow reviewers to open smaller issues that can then be linked to PR's rather than submitting a more dense text based review. It will also allow you to demonstrate addressing the issue via PR links. - [x] Yes I am OK with reviewers submitting requested changes as issues to my repo. Reviewers will then link to the issues in their submitted review. ## Code of conduct - [X] I agree to abide by [pyOpenSci's Code of Conduct][PyOpenSciCodeOfConduct] during the review process and in maintaining my package should it be accepted. **P.S.** *Have feedback/comments about our review process? Leave a comment [here][Comments] I feel like there should be a section for details that might not fit into any of the questions above. Currently my package does not implement caching, but the skeleton code is currently in the module for future implementation. It will work similar to the caching implemented in the rhud package currently being reviewed by ROpenSci. I have not looked into the PEP 8 style guide yet, but will make sure to review that. ## Editor and Review Templates [Editor and review templates can be found here][Templates] [PackagingGuide]: https://www.pyopensci.org/contributing-guide/authoring/index.html#packaging-guide [PackageCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#package-categories [NotesOnCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#notes-on-categories [JournalOfOpenSourceSoftware]: http://joss.theoj.org/ [JossSubmissionRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#submission-requirements [JossPaperRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#what-should-my-paper-contain [PyOpenSciCodeOfConduct]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/code-of-conduct.html?highlight=code%20conduct [OsiApprovedLicense]: https://opensource.org/licenses [Templates]: https://www.pyopensci.org/contributing-guide/appendices/templates.html [Comments]: https://github.com/pyOpenSci/governance/issues/8
True
hudpy: A Python interface for the US Department of Housing and Urban Development APIs - Submitting Author: @etam4260 Package Name: hudpy One-Line Description of Package: A Python Interface to the US Department of Housing and Urban Development APIs Repository Link: https://github.com/etam4260/hudpy Version submitted: 0.2.0 Editor: TBD Reviewer 1: TBD Reviewer 2: TBD Archive: TBD Version accepted: TBD --- ## Description - Include a brief paragraph describing what your package does: Currently the R version is under peer review at ROpenSci: https://github.com/ropensci/software-review/issues/524 This package is a mapped over version from R to Python. This package gives users access to the US Department of Housing and Urban Development APIs available at HUD User: https://www.huduser.gov/hudapi/public/login This also has additional wrappers onto of the data provided from these APIs to determine if US geographies intersect based on addresses and (cross walking a US dataset) explained in this paper below: Wilson, Ron and Din, Alexander, 2018. “Understanding and Enhancing the U.S. Department of Housing and Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 20 Number 2, 277 https://www.huduser.gov/portal/periodicals/cityscpe/vol20num2/ch16.pdf ## Scope - Please indicate which [category or categories][PackageCategories] this package falls under: - [X] Data retrieval - [ ] Data extraction - [X] Data munging - [ ] Data deposition - [ ] Reproducibility - [X] Geospatial - [ ] Education - [ ] Data visualization* > *Please fill out a pre-submission inquiry before submitting a data visualization package. For more info, see [notes on categories][NotesOnCategories] of our guidebook.* NA - **For all submissions**, explain how the and why the package falls under the categories you indicated above. In your explanation, please address the following points (briefly, 1-2 sentences for each): It is a data retrieval package because it retrieves data from an API. It is data munging package because it implements the concept of cross walking a file (converting a dataset described in one US geographic identifier into another) explained in the paper mentioned above. - Who is the target audience and what are scientific applications of this package? - I am hoping to reach professors, researchers, and students with this package. This gives access to the crosswalk files which is a geospatial technique described very well in these journal articles: Din, Alexander and Wilson, Ron, 2020. “Crosswalking ZIP Codes to Census Geographies: Geoprocessing the U.S. Department of Housing & Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 22, Number 1, https://www.huduser.gov/portal/periodicals/cityscpe/vol22num1/ch12.pdf Wilson, Ron and Din, Alexander, 2018. “Understanding and Enhancing the U.S. Department of Housing and Urban Development’s ZIP Code Crosswalk Files,” Cityscape: A Journal of Policy Development and Research, Volume 20 Number 2, 277 – 294. - Are there other Python packages that accomplish the same thing? If so, how does yours differ? Not any that I know of. - If you made a pre-submission enquiry, please paste the link to the corresponding issue, forum post, or other discussion, or `@tag` the editor you contacted: NA ## Technical checks For details about the pyOpenSci packaging requirements, see our [packaging guide][PackagingGuide]. Confirm each of the following by checking the box. This package: - [X] does not violate the Terms of Service of any service it interacts with. - [X] has an [OSI approved license][OsiApprovedLicense]. - [X] contains a README with instructions for installing the development version. - [X] includes documentation with examples for all functions. - [X] contains a vignette with examples of its essential functions and uses. - [X] has a test suite. - [X] has continuous integration, such as Travis CI, AppVeyor, CircleCI, and/or others. In terms of vignette, I am assuming you are referring to a Jupyter notebook style documentation. I don't have this, but there is a website with vignette like documentation: https://etam4260.github.io/hudpy/build/html/index.html ## Publication options - [ ] Do you wish to automatically submit to the [Journal of Open Source Software][JournalOfOpenSourceSoftware]? If so: <details> <summary>JOSS Checks</summary> - [ ] The package has an **obvious research application** according to JOSS's definition in their [submission requirements][JossSubmissionRequirements]. Be aware that completing the pyOpenSci review process **does not** guarantee acceptance to JOSS. Be sure to read their submission requirements (linked above) if you are interested in submitting to JOSS. - [ ] The package is not a "minor utility" as defined by JOSS's [submission requirements][JossSubmissionRequirements]: "Minor ‘utility’ packages, including ‘thin’ API clients, are not acceptable." pyOpenSci welcomes these packages under "Data Retrieval", but JOSS has slightly different criteria. - [ ] The package contains a `paper.md` matching [JOSS's requirements][JossPaperRequirements] with a high-level description in the package root or in `inst/`. - [ ] The package is deposited in a long-term repository with the DOI: *Note: Do not submit your package separately to JOSS* </details> ## Are you OK with Reviewers Submitting Issues and/or pull requests to your Repo Directly? This option will allow reviewers to open smaller issues that can then be linked to PR's rather than submitting a more dense text based review. It will also allow you to demonstrate addressing the issue via PR links. - [x] Yes I am OK with reviewers submitting requested changes as issues to my repo. Reviewers will then link to the issues in their submitted review. ## Code of conduct - [X] I agree to abide by [pyOpenSci's Code of Conduct][PyOpenSciCodeOfConduct] during the review process and in maintaining my package should it be accepted. **P.S.** *Have feedback/comments about our review process? Leave a comment [here][Comments] I feel like there should be a section for details that might not fit into any of the questions above. Currently my package does not implement caching, but the skeleton code is currently in the module for future implementation. It will work similar to the caching implemented in the rhud package currently being reviewed by ROpenSci. I have not looked into the PEP 8 style guide yet, but will make sure to review that. ## Editor and Review Templates [Editor and review templates can be found here][Templates] [PackagingGuide]: https://www.pyopensci.org/contributing-guide/authoring/index.html#packaging-guide [PackageCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#package-categories [NotesOnCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#notes-on-categories [JournalOfOpenSourceSoftware]: http://joss.theoj.org/ [JossSubmissionRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#submission-requirements [JossPaperRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#what-should-my-paper-contain [PyOpenSciCodeOfConduct]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/code-of-conduct.html?highlight=code%20conduct [OsiApprovedLicense]: https://opensource.org/licenses [Templates]: https://www.pyopensci.org/contributing-guide/appendices/templates.html [Comments]: https://github.com/pyOpenSci/governance/issues/8
main
hudpy a python interface for the us department of housing and urban development apis submitting author package name hudpy one line description of package a python interface to the us department of housing and urban development apis repository link version submitted editor tbd reviewer tbd reviewer tbd archive tbd version accepted tbd description include a brief paragraph describing what your package does currently the r version is under peer review at ropensci this package is a mapped over version from r to python this package gives users access to the us department of housing and urban development apis available at hud user this also has additional wrappers onto of the data provided from these apis to determine if us geographies intersect based on addresses and cross walking a us dataset explained in this paper below wilson ron and din alexander “understanding and enhancing the u s department of housing and urban development’s zip code crosswalk files ” cityscape a journal of policy development and research volume number scope please indicate which this package falls under data retrieval data extraction data munging data deposition reproducibility geospatial education data visualization please fill out a pre submission inquiry before submitting a data visualization package for more info see of our guidebook na for all submissions explain how the and why the package falls under the categories you indicated above in your explanation please address the following points briefly sentences for each it is a data retrieval package because it retrieves data from an api it is data munging package because it implements the concept of cross walking a file converting a dataset described in one us geographic identifier into another explained in the paper mentioned above who is the target audience and what are scientific applications of this package i am hoping to reach professors researchers and students with this package this gives access to the crosswalk files which is a geospatial technique described very well in these journal articles din alexander and wilson ron “crosswalking zip codes to census geographies geoprocessing the u s department of housing urban development’s zip code crosswalk files ” cityscape a journal of policy development and research volume number wilson ron and din alexander “understanding and enhancing the u s department of housing and urban development’s zip code crosswalk files ” cityscape a journal of policy development and research volume number – are there other python packages that accomplish the same thing if so how does yours differ not any that i know of if you made a pre submission enquiry please paste the link to the corresponding issue forum post or other discussion or tag the editor you contacted na technical checks for details about the pyopensci packaging requirements see our confirm each of the following by checking the box this package does not violate the terms of service of any service it interacts with has an contains a readme with instructions for installing the development version includes documentation with examples for all functions contains a vignette with examples of its essential functions and uses has a test suite has continuous integration such as travis ci appveyor circleci and or others in terms of vignette i am assuming you are referring to a jupyter notebook style documentation i don t have this but there is a website with vignette like documentation publication options do you wish to automatically submit to the if so joss checks the package has an obvious research application according to joss s definition in their be aware that completing the pyopensci review process does not guarantee acceptance to joss be sure to read their submission requirements linked above if you are interested in submitting to joss the package is not a minor utility as defined by joss s minor ‘utility’ packages including ‘thin’ api clients are not acceptable pyopensci welcomes these packages under data retrieval but joss has slightly different criteria the package contains a paper md matching with a high level description in the package root or in inst the package is deposited in a long term repository with the doi note do not submit your package separately to joss are you ok with reviewers submitting issues and or pull requests to your repo directly this option will allow reviewers to open smaller issues that can then be linked to pr s rather than submitting a more dense text based review it will also allow you to demonstrate addressing the issue via pr links yes i am ok with reviewers submitting requested changes as issues to my repo reviewers will then link to the issues in their submitted review code of conduct i agree to abide by during the review process and in maintaining my package should it be accepted p s have feedback comments about our review process leave a comment i feel like there should be a section for details that might not fit into any of the questions above currently my package does not implement caching but the skeleton code is currently in the module for future implementation it will work similar to the caching implemented in the rhud package currently being reviewed by ropensci i have not looked into the pep style guide yet but will make sure to review that editor and review templates
1
205,738
16,007,648,617
IssuesEvent
2021-04-20 06:24:41
dankamongmen/notcurses
https://api.github.com/repos/dankamongmen/notcurses
opened
add some sprixel stats
bitmaps documentation enhancement
Currently we have no stats regarding sprixels. I'd like to be able to know how many had been drawn, and how many draws had been elided.
1.0
add some sprixel stats - Currently we have no stats regarding sprixels. I'd like to be able to know how many had been drawn, and how many draws had been elided.
non_main
add some sprixel stats currently we have no stats regarding sprixels i d like to be able to know how many had been drawn and how many draws had been elided
0
726,438
24,999,387,576
IssuesEvent
2022-11-03 05:58:35
GeoChatter/GeoChatter-Client
https://api.github.com/repos/GeoChatter/GeoChatter-Client
closed
Some UserScripts can't access game start screen settings container
bug -priority-level-3 lang-C# lang-TS non-critical
Scripts like Unity and Blink mode can't locate the game settings container
1.0
Some UserScripts can't access game start screen settings container - Scripts like Unity and Blink mode can't locate the game settings container
non_main
some userscripts can t access game start screen settings container scripts like unity and blink mode can t locate the game settings container
0
118,969
10,020,845,157
IssuesEvent
2019-07-16 13:30:17
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
opened
Create EJB Custom Binding Property Tests
in:EJB Container team:Blizzard test delivery
Create tests for the new properties to implemented for the custom bindings work.
1.0
Create EJB Custom Binding Property Tests - Create tests for the new properties to implemented for the custom bindings work.
non_main
create ejb custom binding property tests create tests for the new properties to implemented for the custom bindings work
0
531
3,926,806,906
IssuesEvent
2016-04-23 04:46:10
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
opened
Forecast: Wrong result due to timezone
Maintainer Input Requested
My timezone is IST and the IA is giving forecast of `Phoenixville, PA` which is wrong. See the snaps below: <img width="1436" alt="screen shot 2016-04-23 at 10 09 00 am" src="https://cloud.githubusercontent.com/assets/915277/14759309/3a005caa-093c-11e6-86e0-09ef0d20d7d7.png"> <img width="734" alt="screen shot 2016-04-23 at 10 11 59 am" src="https://cloud.githubusercontent.com/assets/915277/14759310/583e0afa-093c-11e6-8e37-2d53d40e8613.png"> IA Page: http://duck.co/ia/view/forecast
True
Forecast: Wrong result due to timezone - My timezone is IST and the IA is giving forecast of `Phoenixville, PA` which is wrong. See the snaps below: <img width="1436" alt="screen shot 2016-04-23 at 10 09 00 am" src="https://cloud.githubusercontent.com/assets/915277/14759309/3a005caa-093c-11e6-86e0-09ef0d20d7d7.png"> <img width="734" alt="screen shot 2016-04-23 at 10 11 59 am" src="https://cloud.githubusercontent.com/assets/915277/14759310/583e0afa-093c-11e6-8e37-2d53d40e8613.png"> IA Page: http://duck.co/ia/view/forecast
main
forecast wrong result due to timezone my timezone is ist and the ia is giving forecast of phoenixville pa which is wrong see the snaps below img width alt screen shot at am src img width alt screen shot at am src ia page
1
125,694
10,349,805,699
IssuesEvent
2019-09-04 23:59:57
Vachok/ftpplus
https://api.github.com/repos/Vachok/ftpplus
closed
testToString - Fail [D22] [D66]
Low TestQuality bug resolution_Wont Do
Execute DBMessengerTest::testToString**testToString** *DBMessengerTest* *did not expect to find [true] but found [false]* *java.lang.AssertionError*
1.0
testToString - Fail [D22] [D66] - Execute DBMessengerTest::testToString**testToString** *DBMessengerTest* *did not expect to find [true] but found [false]* *java.lang.AssertionError*
non_main
testtostring fail execute dbmessengertest testtostring testtostring dbmessengertest did not expect to find but found java lang assertionerror
0
2,059
6,977,595,507
IssuesEvent
2017-12-12 15:07:19
ansible/ansible
https://api.github.com/repos/ansible/ansible
reopened
AWS_PROFILE is no longer respected by aws modules in ansible 2.4
affects_2.4 aws bug_report cloud module needs_maintainer support:community
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME all aws modules (I guess). Observed for `ec2_vpc`, `ec2_group`, `ec2`, `ec2_eni` etc. ##### ANSIBLE VERSION ``` ansible 2.4.1.0 config file = /home/stefan/mediapeers/mpx-infrastructure/ansible.cfg configured module search path = [u'/home/stefan/mediapeers/mpx-infrastructure/library'] ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible executable location = /usr/local/bin/ansible python version = 2.7.12 (default, Nov 20 2017, 18:23:56) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION ``` ANSIBLE_FORCE_COLOR(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True ANSIBLE_NOCOWS(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True ANSIBLE_PIPELINING(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True DEFAULT_BECOME_METHOD(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = sudo DEFAULT_CALLBACK_WHITELIST(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = ['profile_tasks'] DEFAULT_FORKS(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 5 DEFAULT_HOST_LIST(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/inventory/staging'] DEFAULT_LOG_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/log/ansible.log DEFAULT_MANAGED_STR(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = Ansible managed: {file} modified on %Y-%m-%d %H:%M:%S by {uid} on {host} DEFAULT_MODULE_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/library'] DEFAULT_POLL_INTERVAL(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 15 DEFAULT_PRIVATE_KEY_FILE(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/settings/ssh_keys/MPX-Staging- DEFAULT_REMOTE_TMP(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = $HOME/.ansible/tmp DEFAULT_REMOTE_USER(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = ubuntu DEFAULT_ROLES_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/roles'] DEFAULT_STDOUT_CALLBACK(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = skippy DEFAULT_TIMEOUT(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 60 DEFAULT_VAULT_PASSWORD_FILE(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/.vault-password HOST_KEY_CHECKING(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = False RETRY_FILES_ENABLED(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = False RETRY_FILES_SAVE_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan ``` ##### OS / ENVIRONMENT Ubuntu 14.04 / 16.04 ##### SUMMARY I always used `AWS_PROFILE` to switch between AWS profiles configures in `~/.aws/credentials` or `~/.boto`. After switching to Ansible 2.4 the ENV variable is ignored and always the default profile is taken, leading to things being created in the wrong AWS account. I didn't find anything about this in the CHANGELOG nor did I find an open ticket for this, so I guess that is a regression bug. ##### STEPS TO REPRODUCE - Have AWS credentials configured for multiple aws profiles (I have `~./boto` and `~/.aws/credentials` not sure which is used by Ansible those days) - Run ansible-playbook command with `AWS_PROFILE=something_else_then_default` - aws modules such as ec2_* are using default profile / wrong credentials set ##### EXPECTED RESULTS Correct credentials get used as in Ansible < 2.4 ##### ACTUAL RESULTS Wrong credentials being used, leading to AWS resources being created under the wrong AWS account. Please advise on how to work arround it, or confirm that this is a bug. Thank you.
True
AWS_PROFILE is no longer respected by aws modules in ansible 2.4 - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME all aws modules (I guess). Observed for `ec2_vpc`, `ec2_group`, `ec2`, `ec2_eni` etc. ##### ANSIBLE VERSION ``` ansible 2.4.1.0 config file = /home/stefan/mediapeers/mpx-infrastructure/ansible.cfg configured module search path = [u'/home/stefan/mediapeers/mpx-infrastructure/library'] ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible executable location = /usr/local/bin/ansible python version = 2.7.12 (default, Nov 20 2017, 18:23:56) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION ``` ANSIBLE_FORCE_COLOR(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True ANSIBLE_NOCOWS(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True ANSIBLE_PIPELINING(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = True DEFAULT_BECOME_METHOD(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = sudo DEFAULT_CALLBACK_WHITELIST(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = ['profile_tasks'] DEFAULT_FORKS(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 5 DEFAULT_HOST_LIST(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/inventory/staging'] DEFAULT_LOG_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/log/ansible.log DEFAULT_MANAGED_STR(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = Ansible managed: {file} modified on %Y-%m-%d %H:%M:%S by {uid} on {host} DEFAULT_MODULE_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/library'] DEFAULT_POLL_INTERVAL(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 15 DEFAULT_PRIVATE_KEY_FILE(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/settings/ssh_keys/MPX-Staging- DEFAULT_REMOTE_TMP(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = $HOME/.ansible/tmp DEFAULT_REMOTE_USER(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = ubuntu DEFAULT_ROLES_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = [u'/home/stefan/mediapeers/mpx-infrastructure/roles'] DEFAULT_STDOUT_CALLBACK(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = skippy DEFAULT_TIMEOUT(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = 60 DEFAULT_VAULT_PASSWORD_FILE(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan/mediapeers/mpx-infrastructure/.vault-password HOST_KEY_CHECKING(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = False RETRY_FILES_ENABLED(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = False RETRY_FILES_SAVE_PATH(/home/stefan/mediapeers/mpx-infrastructure/ansible.cfg) = /home/stefan ``` ##### OS / ENVIRONMENT Ubuntu 14.04 / 16.04 ##### SUMMARY I always used `AWS_PROFILE` to switch between AWS profiles configures in `~/.aws/credentials` or `~/.boto`. After switching to Ansible 2.4 the ENV variable is ignored and always the default profile is taken, leading to things being created in the wrong AWS account. I didn't find anything about this in the CHANGELOG nor did I find an open ticket for this, so I guess that is a regression bug. ##### STEPS TO REPRODUCE - Have AWS credentials configured for multiple aws profiles (I have `~./boto` and `~/.aws/credentials` not sure which is used by Ansible those days) - Run ansible-playbook command with `AWS_PROFILE=something_else_then_default` - aws modules such as ec2_* are using default profile / wrong credentials set ##### EXPECTED RESULTS Correct credentials get used as in Ansible < 2.4 ##### ACTUAL RESULTS Wrong credentials being used, leading to AWS resources being created under the wrong AWS account. Please advise on how to work arround it, or confirm that this is a bug. Thank you.
main
aws profile is no longer respected by aws modules in ansible issue type bug report component name all aws modules i guess observed for vpc group eni etc ansible version ansible config file home stefan mediapeers mpx infrastructure ansible cfg configured module search path ansible python module location usr local lib dist packages ansible executable location usr local bin ansible python version default nov configuration ansible force color home stefan mediapeers mpx infrastructure ansible cfg true ansible nocows home stefan mediapeers mpx infrastructure ansible cfg true ansible pipelining home stefan mediapeers mpx infrastructure ansible cfg true default become method home stefan mediapeers mpx infrastructure ansible cfg sudo default callback whitelist home stefan mediapeers mpx infrastructure ansible cfg default forks home stefan mediapeers mpx infrastructure ansible cfg default host list home stefan mediapeers mpx infrastructure ansible cfg default log path home stefan mediapeers mpx infrastructure ansible cfg home stefan mediapeers mpx infrastructure log ansible log default managed str home stefan mediapeers mpx infrastructure ansible cfg ansible managed file modified on y m d h m s by uid on host default module path home stefan mediapeers mpx infrastructure ansible cfg default poll interval home stefan mediapeers mpx infrastructure ansible cfg default private key file home stefan mediapeers mpx infrastructure ansible cfg home stefan mediapeers mpx infrastructure settings ssh keys mpx staging default remote tmp home stefan mediapeers mpx infrastructure ansible cfg home ansible tmp default remote user home stefan mediapeers mpx infrastructure ansible cfg ubuntu default roles path home stefan mediapeers mpx infrastructure ansible cfg default stdout callback home stefan mediapeers mpx infrastructure ansible cfg skippy default timeout home stefan mediapeers mpx infrastructure ansible cfg default vault password file home stefan mediapeers mpx infrastructure ansible cfg home stefan mediapeers mpx infrastructure vault password host key checking home stefan mediapeers mpx infrastructure ansible cfg false retry files enabled home stefan mediapeers mpx infrastructure ansible cfg false retry files save path home stefan mediapeers mpx infrastructure ansible cfg home stefan os environment ubuntu summary i always used aws profile to switch between aws profiles configures in aws credentials or boto after switching to ansible the env variable is ignored and always the default profile is taken leading to things being created in the wrong aws account i didn t find anything about this in the changelog nor did i find an open ticket for this so i guess that is a regression bug steps to reproduce have aws credentials configured for multiple aws profiles i have boto and aws credentials not sure which is used by ansible those days run ansible playbook command with aws profile something else then default aws modules such as are using default profile wrong credentials set expected results correct credentials get used as in ansible actual results wrong credentials being used leading to aws resources being created under the wrong aws account please advise on how to work arround it or confirm that this is a bug thank you
1
2,565
8,712,611,266
IssuesEvent
2018-12-06 22:50:02
clearlinux/swupd-client
https://api.github.com/repos/clearlinux/swupd-client
closed
Make it easier to add a global option
maintainability
Swupd documentation states that there are some options that are applicable to most subcommands. We could have a better way to handle parameters so we wouldn't need to duplicate them in all subcommands that use them. Currently even help documentation on what that parameter is doing is duplicated. Commit 15b987c is one example on how it's hard to maintain global options.
True
Make it easier to add a global option - Swupd documentation states that there are some options that are applicable to most subcommands. We could have a better way to handle parameters so we wouldn't need to duplicate them in all subcommands that use them. Currently even help documentation on what that parameter is doing is duplicated. Commit 15b987c is one example on how it's hard to maintain global options.
main
make it easier to add a global option swupd documentation states that there are some options that are applicable to most subcommands we could have a better way to handle parameters so we wouldn t need to duplicate them in all subcommands that use them currently even help documentation on what that parameter is doing is duplicated commit is one example on how it s hard to maintain global options
1
1,292
5,475,877,810
IssuesEvent
2017-03-11 15:37:01
WhitestormJS/whitestorm.js
https://api.github.com/repos/WhitestormJS/whitestorm.js
closed
Make Travis CI builds run on linux
ENHANCEMENT MAINTAINANCE
Currently, all builds use OSX. We use it because of [headless-gl] works with it out of the box. **See:** - https://github.com/stackgl/headless-gl#how-can-i-use-headless-gl-with-a-continuous-integration-service **Solution.** Add those lines to `travis.yml` ```yml language: node_js os: linux sudo: required dist: trusty addons: apt: packages: - mesa-utils - xvfb - libgl1-mesa-dri - libglapi-mesa - libosmesa6 node_js: - '6' before_script: - export DISPLAY=:99.0; sh -e /etc/init.d/xvfb start ``` ###### Version: - [x] v2.x.x - [ ] v1.x.x ###### Issue type: - [ ] Bug - [x] Proposal/Enhancement - [ ] Question ------ <details> <summary> <b>Tested on: </b> </summary> ###### --- Desktop - [ ] Chrome - [ ] Chrome Canary - [ ] Chrome dev-channel - [ ] Firefox - [ ] Opera - [ ] Microsoft IE - [ ] Microsoft Edge ###### --- Android - [ ] Chrome - [ ] Firefox - [ ] Opera ###### --- IOS - [ ] Chrome - [ ] Firefox - [ ] Opera </details>
True
Make Travis CI builds run on linux - Currently, all builds use OSX. We use it because of [headless-gl] works with it out of the box. **See:** - https://github.com/stackgl/headless-gl#how-can-i-use-headless-gl-with-a-continuous-integration-service **Solution.** Add those lines to `travis.yml` ```yml language: node_js os: linux sudo: required dist: trusty addons: apt: packages: - mesa-utils - xvfb - libgl1-mesa-dri - libglapi-mesa - libosmesa6 node_js: - '6' before_script: - export DISPLAY=:99.0; sh -e /etc/init.d/xvfb start ``` ###### Version: - [x] v2.x.x - [ ] v1.x.x ###### Issue type: - [ ] Bug - [x] Proposal/Enhancement - [ ] Question ------ <details> <summary> <b>Tested on: </b> </summary> ###### --- Desktop - [ ] Chrome - [ ] Chrome Canary - [ ] Chrome dev-channel - [ ] Firefox - [ ] Opera - [ ] Microsoft IE - [ ] Microsoft Edge ###### --- Android - [ ] Chrome - [ ] Firefox - [ ] Opera ###### --- IOS - [ ] Chrome - [ ] Firefox - [ ] Opera </details>
main
make travis ci builds run on linux currently all builds use osx we use it because of works with it out of the box see solution add those lines to travis yml yml language node js os linux sudo required dist trusty addons apt packages mesa utils xvfb mesa dri libglapi mesa node js before script export display sh e etc init d xvfb start version x x x x issue type bug proposal enhancement question tested on desktop chrome chrome canary chrome dev channel firefox opera microsoft ie microsoft edge android chrome firefox opera ios chrome firefox opera
1
269,073
8,425,450,586
IssuesEvent
2018-10-16 02:37:20
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Make --compiler=python3.7 available for run_tests.py
kind/enhancement lang/Python priority/P1
`run_tests.py` currently does not support the latest Python 3 (3.7)
1.0
Make --compiler=python3.7 available for run_tests.py - `run_tests.py` currently does not support the latest Python 3 (3.7)
non_main
make compiler available for run tests py run tests py currently does not support the latest python
0
464,534
13,326,429,937
IssuesEvent
2020-08-27 11:37:41
Uninett/Argus
https://api.github.com/repos/Uninett/Argus
closed
Rename "active"->"open" and "inactive"->"closed" (incidents)
data model priority: low task
"Closed" is in the ITIL terminology list, and "open" could be a fitting antonym.
1.0
Rename "active"->"open" and "inactive"->"closed" (incidents) - "Closed" is in the ITIL terminology list, and "open" could be a fitting antonym.
non_main
rename active open and inactive closed incidents closed is in the itil terminology list and open could be a fitting antonym
0
2,128
7,268,949,795
IssuesEvent
2018-02-20 11:57:33
dgets/DANT2a
https://api.github.com/repos/dgets/DANT2a
closed
Move entry processing code from _Tick to Utility
enhancement maintainability
The code for processing the list of each type of entries should be relocated to _Utility.cs_.
True
Move entry processing code from _Tick to Utility - The code for processing the list of each type of entries should be relocated to _Utility.cs_.
main
move entry processing code from tick to utility the code for processing the list of each type of entries should be relocated to utility cs
1
1,806
6,575,943,409
IssuesEvent
2017-09-11 17:55:40
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
feature request: rhn_channel, and other rhn_* modules, should have validate_certs parameter
affects_2.1 feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> rhn_channel rhn_* ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` [snemirovsky@SN-WS-Fedora24 patch]$ ansible --version [WARNING]: log file at /var/log/ansible.log is not writeable and we cannot create it, aborting ansible 2.1.1.0 config file = /ansible/conf/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> ##### STEPS TO REPRODUCE <!--- run the module agains rh satellite api URL with self-signed or expired cert. --> <!--- Paste example playbooks or commands between quotes below --> ``` ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> excected my code to work ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [RHN_CHANNEL_configure : Configure Channel on Satellite] ****************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:590) ```
True
feature request: rhn_channel, and other rhn_* modules, should have validate_certs parameter - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> rhn_channel rhn_* ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` [snemirovsky@SN-WS-Fedora24 patch]$ ansible --version [WARNING]: log file at /var/log/ansible.log is not writeable and we cannot create it, aborting ansible 2.1.1.0 config file = /ansible/conf/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> ##### STEPS TO REPRODUCE <!--- run the module agains rh satellite api URL with self-signed or expired cert. --> <!--- Paste example playbooks or commands between quotes below --> ``` ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> excected my code to work ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [RHN_CHANNEL_configure : Configure Channel on Satellite] ****************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:590) ```
main
feature request rhn channel and other rhn modules should have validate certs parameter issue type feature idea component name rhn channel rhn ansible version ansible version log file at var log ansible log is not writeable and we cannot create it aborting ansible config file ansible conf ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific summary steps to reproduce run the module agains rh satellite api url with self signed or expired cert expected results excected my code to work actual results task an exception occurred during task execution to see the full traceback use vvv the error was ssl sslerror certificate verify failed ssl c
1
464
3,689,779,242
IssuesEvent
2016-02-25 17:35:04
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
closed
Rename IF_X86_X64 to IF_X86_64
GoodContrib Maintainability
The macros IF_X86_X64, IF_X86_64_ELSE, and so on, are defined in core/lib/globals_shared.h and used in four other files. The names seem confusingly inconsistent. I think you should probably do something like: perl -i -pe 's/_X86_X64/_X86_64/g;' `find core ext -type f`
True
Rename IF_X86_X64 to IF_X86_64 - The macros IF_X86_X64, IF_X86_64_ELSE, and so on, are defined in core/lib/globals_shared.h and used in four other files. The names seem confusingly inconsistent. I think you should probably do something like: perl -i -pe 's/_X86_X64/_X86_64/g;' `find core ext -type f`
main
rename if to if the macros if if else and so on are defined in core lib globals shared h and used in four other files the names seem confusingly inconsistent i think you should probably do something like perl i pe s g find core ext type f
1
4,787
24,628,453,134
IssuesEvent
2022-10-16 20:17:20
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
RecursionError in records endpoint
type: bug work: backend status: ready restricted: maintainers
## Description I've been getting this error for a few tables in my environment. * These tables have been created using the 'Create new table' button, not using file import. * These tables have no rows. ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/2/records/ Django Version: 3.1.14 Python Version: 3.9.14 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 67, in list records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 82, in paginate_queryset preview_metadata, preview_columns = get_preview_info(table.id) File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 70, in get_preview_info fk_constraints = [ File "/code/mathesar/utils/preview.py", line 73, in <listcomp> if table_constraint.type == ConstraintType.FOREIGN_KEY.value File "/code/mathesar/models/base.py", line 774, in type return constraint_utils.get_constraint_type_from_char(self._constraint_record['contype']) File "/code/mathesar/models/base.py", line 766, in _constraint_record return get_constraint_record_from_oid(self.oid, engine) File "/code/db/constraints/operations/select.py", line 33, in get_constraint_record_from_oid pg_constraint = get_pg_catalog_table("pg_constraint", engine, metadata=metadata) File "/code/db/utils.py", line 92, in warning_ignored_func return f(*args, **kwargs) File "/code/db/utils.py", line 99, in get_pg_catalog_table return sqlalchemy.Table(table_name, metadata, autoload_with=engine, schema='pg_catalog') File "<string>", line 2, in __new__ <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/deprecations.py", line 298, in warned return fn(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 600, in __new__ metadata._remove_table(name, schema) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ compat.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 595, in __new__ table._init(name, metadata, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 670, in _init self._autoload( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 705, in _autoload conn_insp.reflect_table( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 774, in reflect_table for col_d in self.get_columns( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 497, in get_columns col_defs = self.dialect.get_columns( File "<string>", line 2, in get_columns <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 55, in cache ret = fn(self, con, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 3585, in get_columns table_oid = self.get_table_oid( File "<string>", line 2, in get_table_oid <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 55, in cache ret = fn(self, con, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 3462, in get_table_oid c = connection.execute(s, dict(table_name=table_name, schema=schema)) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/future/engine.py", line 280, in execute return self._execute_20( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1582, in _execute_20 return meth(self, args_10style, kwargs_10style, execution_options) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 324, in _execute_on_connection return connection._execute_clauseelement( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1451, in _execute_clauseelement ret = self._execute_context( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1813, in _execute_context self._handle_dbapi_exception( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1998, in _handle_dbapi_exception util.raise_(exc_info[1], with_traceback=exc_info[2]) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1786, in _execute_context result = context._setup_result_proxy() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 1406, in _setup_result_proxy result = self._setup_dml_or_text_result() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 1494, in _setup_dml_or_text_result result = _cursor.CursorResult(self, strategy, cursor_description) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 1253, in __init__ metadata = self._init_metadata(context, cursor_description) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 1310, in _init_metadata metadata = metadata._adapt_to_context(context) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 136, in _adapt_to_context invoked_statement._exported_columns_iterator() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 126, in _exported_columns_iterator return iter(self.exported_columns) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 2870, in exported_columns return self.selected_columns File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 1180, in __get__ obj.__dict__[self.__name__] = result = self.fget(obj) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 6354, in selected_columns return ColumnCollection( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1128, in __init__ self._initial_populate(columns) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1131, in _initial_populate self._populate_separate_keys(iter_) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1227, in _populate_separate_keys self._colset.update(c for k, c in self._collection) Exception Type: RecursionError at /api/db/v0/tables/2/records/ Exception Value: maximum recursion depth exceeded ``` I'm not sure about the cause and it's occuring consistently for me but unable to reproduce it on staging.
True
RecursionError in records endpoint - ## Description I've been getting this error for a few tables in my environment. * These tables have been created using the 'Create new table' button, not using file import. * These tables have no rows. ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/2/records/ Django Version: 3.1.14 Python Version: 3.9.14 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 67, in list records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 82, in paginate_queryset preview_metadata, preview_columns = get_preview_info(table.id) File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 81, in get_preview_info preview_info, columns = _preview_info_by_column_id( File "/code/mathesar/utils/preview.py", line 22, in _preview_info_by_column_id referent_preview_info, referent_preview_columns = get_preview_info( File "/code/mathesar/utils/preview.py", line 70, in get_preview_info fk_constraints = [ File "/code/mathesar/utils/preview.py", line 73, in <listcomp> if table_constraint.type == ConstraintType.FOREIGN_KEY.value File "/code/mathesar/models/base.py", line 774, in type return constraint_utils.get_constraint_type_from_char(self._constraint_record['contype']) File "/code/mathesar/models/base.py", line 766, in _constraint_record return get_constraint_record_from_oid(self.oid, engine) File "/code/db/constraints/operations/select.py", line 33, in get_constraint_record_from_oid pg_constraint = get_pg_catalog_table("pg_constraint", engine, metadata=metadata) File "/code/db/utils.py", line 92, in warning_ignored_func return f(*args, **kwargs) File "/code/db/utils.py", line 99, in get_pg_catalog_table return sqlalchemy.Table(table_name, metadata, autoload_with=engine, schema='pg_catalog') File "<string>", line 2, in __new__ <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/deprecations.py", line 298, in warned return fn(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 600, in __new__ metadata._remove_table(name, schema) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ compat.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 595, in __new__ table._init(name, metadata, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 670, in _init self._autoload( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/schema.py", line 705, in _autoload conn_insp.reflect_table( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 774, in reflect_table for col_d in self.get_columns( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 497, in get_columns col_defs = self.dialect.get_columns( File "<string>", line 2, in get_columns <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 55, in cache ret = fn(self, con, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 3585, in get_columns table_oid = self.get_table_oid( File "<string>", line 2, in get_table_oid <source code not available> File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/reflection.py", line 55, in cache ret = fn(self, con, *args, **kw) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/dialects/postgresql/base.py", line 3462, in get_table_oid c = connection.execute(s, dict(table_name=table_name, schema=schema)) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/future/engine.py", line 280, in execute return self._execute_20( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1582, in _execute_20 return meth(self, args_10style, kwargs_10style, execution_options) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 324, in _execute_on_connection return connection._execute_clauseelement( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1451, in _execute_clauseelement ret = self._execute_context( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1813, in _execute_context self._handle_dbapi_exception( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1998, in _handle_dbapi_exception util.raise_(exc_info[1], with_traceback=exc_info[2]) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1786, in _execute_context result = context._setup_result_proxy() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 1406, in _setup_result_proxy result = self._setup_dml_or_text_result() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 1494, in _setup_dml_or_text_result result = _cursor.CursorResult(self, strategy, cursor_description) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 1253, in __init__ metadata = self._init_metadata(context, cursor_description) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 1310, in _init_metadata metadata = metadata._adapt_to_context(context) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/cursor.py", line 136, in _adapt_to_context invoked_statement._exported_columns_iterator() File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 126, in _exported_columns_iterator return iter(self.exported_columns) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 2870, in exported_columns return self.selected_columns File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 1180, in __get__ obj.__dict__[self.__name__] = result = self.fget(obj) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/selectable.py", line 6354, in selected_columns return ColumnCollection( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1128, in __init__ self._initial_populate(columns) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1131, in _initial_populate self._populate_separate_keys(iter_) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1227, in _populate_separate_keys self._colset.update(c for k, c in self._collection) Exception Type: RecursionError at /api/db/v0/tables/2/records/ Exception Value: maximum recursion depth exceeded ``` I'm not sure about the cause and it's occuring consistently for me but unable to reproduce it on staging.
main
recursionerror in records endpoint description i ve been getting this error for a few tables in my environment these tables have been created using the create new table button not using file import these tables have no rows environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware traceback most recent call last file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets records py line in list records paginator paginate queryset file code mathesar api pagination py line in paginate queryset preview metadata preview columns get preview info table id file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info preview info columns preview info by column id file code mathesar utils preview py line in preview info by column id referent preview info referent preview columns get preview info file code mathesar utils preview py line in get preview info fk constraints file code mathesar utils preview py line in if table constraint type constrainttype foreign key value file code mathesar models base py line in type return constraint utils get constraint type from char self constraint record file code mathesar models base py line in constraint record return get constraint record from oid self oid engine file code db constraints operations select py line in get constraint record from oid pg constraint get pg catalog table pg constraint engine metadata metadata file code db utils py line in warning ignored func return f args kwargs file code db utils py line in get pg catalog table return sqlalchemy table table name metadata autoload with engine schema pg catalog file line in new file usr local lib site packages sqlalchemy util deprecations py line in warned return fn args kwargs file usr local lib site packages sqlalchemy sql schema py line in new metadata remove table name schema file usr local lib site packages sqlalchemy util langhelpers py line in exit compat raise file usr local lib site packages sqlalchemy util compat py line in raise raise exception file usr local lib site packages sqlalchemy sql schema py line in new table init name metadata args kw file usr local lib site packages sqlalchemy sql schema py line in init self autoload file usr local lib site packages sqlalchemy sql schema py line in autoload conn insp reflect table file usr local lib site packages sqlalchemy engine reflection py line in reflect table for col d in self get columns file usr local lib site packages sqlalchemy engine reflection py line in get columns col defs self dialect get columns file line in get columns file usr local lib site packages sqlalchemy engine reflection py line in cache ret fn self con args kw file usr local lib site packages sqlalchemy dialects postgresql base py line in get columns table oid self get table oid file line in get table oid file usr local lib site packages sqlalchemy engine reflection py line in cache ret fn self con args kw file usr local lib site packages sqlalchemy dialects postgresql base py line in get table oid c connection execute s dict table name table name schema schema file usr local lib site packages sqlalchemy future engine py line in execute return self execute file usr local lib site packages sqlalchemy engine base py line in execute return meth self args kwargs execution options file usr local lib site packages sqlalchemy sql elements py line in execute on connection return connection execute clauseelement file usr local lib site packages sqlalchemy engine base py line in execute clauseelement ret self execute context file usr local lib site packages sqlalchemy engine base py line in execute context self handle dbapi exception file usr local lib site packages sqlalchemy engine base py line in handle dbapi exception util raise exc info with traceback exc info file usr local lib site packages sqlalchemy util compat py line in raise raise exception file usr local lib site packages sqlalchemy engine base py line in execute context result context setup result proxy file usr local lib site packages sqlalchemy engine default py line in setup result proxy result self setup dml or text result file usr local lib site packages sqlalchemy engine default py line in setup dml or text result result cursor cursorresult self strategy cursor description file usr local lib site packages sqlalchemy engine cursor py line in init metadata self init metadata context cursor description file usr local lib site packages sqlalchemy engine cursor py line in init metadata metadata metadata adapt to context context file usr local lib site packages sqlalchemy engine cursor py line in adapt to context invoked statement exported columns iterator file usr local lib site packages sqlalchemy sql selectable py line in exported columns iterator return iter self exported columns file usr local lib site packages sqlalchemy sql selectable py line in exported columns return self selected columns file usr local lib site packages sqlalchemy util langhelpers py line in get obj dict result self fget obj file usr local lib site packages sqlalchemy sql selectable py line in selected columns return columncollection file usr local lib site packages sqlalchemy sql base py line in init self initial populate columns file usr local lib site packages sqlalchemy sql base py line in initial populate self populate separate keys iter file usr local lib site packages sqlalchemy sql base py line in populate separate keys self colset update c for k c in self collection exception type recursionerror at api db tables records exception value maximum recursion depth exceeded i m not sure about the cause and it s occuring consistently for me but unable to reproduce it on staging
1
2,457
3,869,769,367
IssuesEvent
2016-04-10 20:01:33
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
23249938: System Integrity Protection applied out of scope.
classification:security reproducible:always status:open
#### Description Summary: System Integrity protection is applied in e.g. Download folder so user cannot delete System app (copied from other OS X version) within Download folder wou Steps to Reproduce: 1. Copy Preview.app from other disk or System into Downloads folder 2. Try to trash the Preview.app Expected Results: Trashed app Actual Results: Fail to trash the app Version: 10.11.1 Notes: Configuration: Attachments: - Product Version: 10.11 Created: 2015-10-25T09:40:56.785720 Originated: 2015-10-25T00:00:00 Open Radar Link: http://www.openradar.me/23249938
True
23249938: System Integrity Protection applied out of scope. - #### Description Summary: System Integrity protection is applied in e.g. Download folder so user cannot delete System app (copied from other OS X version) within Download folder wou Steps to Reproduce: 1. Copy Preview.app from other disk or System into Downloads folder 2. Try to trash the Preview.app Expected Results: Trashed app Actual Results: Fail to trash the app Version: 10.11.1 Notes: Configuration: Attachments: - Product Version: 10.11 Created: 2015-10-25T09:40:56.785720 Originated: 2015-10-25T00:00:00 Open Radar Link: http://www.openradar.me/23249938
non_main
system integrity protection applied out of scope description summary system integrity protection is applied in e g download folder so user cannot delete system app copied from other os x version within download folder wou steps to reproduce copy preview app from other disk or system into downloads folder try to trash the preview app expected results trashed app actual results fail to trash the app version notes configuration attachments product version created originated open radar link
0
4,795
24,698,663,223
IssuesEvent
2022-10-19 13:53:07
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Hide grouping separators by default in Number columns
type: enhancement work: frontend status: ready restricted: maintainers
## Current behavior - We default to displaying grouping separators (e.g. `1,234`) in _all_ numbers. - Users can turn off grouping separators by customizing their display options. - This "on-by-default" behavior, feels a bit strange for columns like primary keys and years, where users typically don't see grouping separators. - During import, Mathesar uses the `numeric` type by default -- even if all values within the import data can be cast to integers. @mathemancer explained that we made this decision early on in order to be flexible, and I think it's worth sticking with it. But it means that years don't get imported as integers by default, making it hard to take the approach were heading towards with the discussion in #1527. - The back end does not set the grouping separator display option during import. It only gets set via the front end if the user manually adjust it. ## Desired behavior - When the user has not configured any display options: - Columns with a UI type of "Number" will be rendered without grouping separators (regardless of their Postgres type). This means that the user will need to manually turn them on sometimes. - Columns with a UI type of "Money" will be rendered _with_ grouping separators. The user can turn them off if needed.
True
Hide grouping separators by default in Number columns - ## Current behavior - We default to displaying grouping separators (e.g. `1,234`) in _all_ numbers. - Users can turn off grouping separators by customizing their display options. - This "on-by-default" behavior, feels a bit strange for columns like primary keys and years, where users typically don't see grouping separators. - During import, Mathesar uses the `numeric` type by default -- even if all values within the import data can be cast to integers. @mathemancer explained that we made this decision early on in order to be flexible, and I think it's worth sticking with it. But it means that years don't get imported as integers by default, making it hard to take the approach were heading towards with the discussion in #1527. - The back end does not set the grouping separator display option during import. It only gets set via the front end if the user manually adjust it. ## Desired behavior - When the user has not configured any display options: - Columns with a UI type of "Number" will be rendered without grouping separators (regardless of their Postgres type). This means that the user will need to manually turn them on sometimes. - Columns with a UI type of "Money" will be rendered _with_ grouping separators. The user can turn them off if needed.
main
hide grouping separators by default in number columns current behavior we default to displaying grouping separators e g in all numbers users can turn off grouping separators by customizing their display options this on by default behavior feels a bit strange for columns like primary keys and years where users typically don t see grouping separators during import mathesar uses the numeric type by default even if all values within the import data can be cast to integers mathemancer explained that we made this decision early on in order to be flexible and i think it s worth sticking with it but it means that years don t get imported as integers by default making it hard to take the approach were heading towards with the discussion in the back end does not set the grouping separator display option during import it only gets set via the front end if the user manually adjust it desired behavior when the user has not configured any display options columns with a ui type of number will be rendered without grouping separators regardless of their postgres type this means that the user will need to manually turn them on sometimes columns with a ui type of money will be rendered with grouping separators the user can turn them off if needed
1
1,873
6,577,499,468
IssuesEvent
2017-09-12 01:20:29
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
setting password when creating user causes job state to always be changed
affects_2.0 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> user module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> nothing ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> OSX Yosemite, El Cap ##### SUMMARY <!--- Explain the problem briefly --> when specifying a user, if a password is specified the user play is always changed unless update_password: on_create is specified. when providing the hash explicitly, vs the plaintext password, this should not be necessary. this behavior does not exist on RHEL5/6/7. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ``` - name: create user user (osx) user: name: newuser state: present password: "{{ user_hash }}" #update_password: on_create groups: 'admin' append: yes shell: /bin/bash ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Similar behavior to rhel application of this play, where the play only reports changed when there is a change to the system state. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> Play results in state changed unless update_password: on_create is specified. ``` TASK [testuser_account_setup : create testuser user (osx)] ************************* task path: /Users/brad8328/repos/ansible-repo/roles/testuser_account_setup/tasks/main.yml:27 <testhost2> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost2> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost2 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746 `" )'"'"'' <testhost1> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost1> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127 `" )'"'"'' <testhost1> PUT /var/folders/d9/7d_sb4mj1bz1jywsy8rrpj61n4xnb5/T/tmpBEbtI3 TO /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/user <testhost1> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r '[testhost1]' <testhost2> PUT /var/folders/d9/7d_sb4mj1bz1jywsy8rrpj61n4xnb5/T/tmp8mMX8j TO /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/user <testhost2> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r '[testhost2]' <testhost1> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost1> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost1 '/bin/sh -c '"'"'sudo -H -S -p "[sudo via ansible, key=ixheeyzjjhmvtkxfyujaeglkekyibvqs] password: " -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-ixheeyzjjhmvtkxfyujaeglkekyibvqs; /bin/sh -c '"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/user; rm -rf "/Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/" > /dev/null 2>&1'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"''"'"'"'"'"'"'"'"''"'"'' <testhost2> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost2> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost2 '/bin/sh -c '"'"'sudo -H -S -p "[sudo via ansible, key=nckvjyuwnrxjapcanzwalcynwzwjeuft] password: " -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-nckvjyuwnrxjapcanzwalcynwzwjeuft; /bin/sh -c '"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/user; rm -rf "/Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/" > /dev/null 2>&1'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"''"'"'"'"'"'"'"'"''"'"'' changed: [testhost1] => {"append": true, "changed": true, "comment": "", "group": 4294967295, "groups": "admin", "home": "/Users/testuser", "invocation": {"module_args": {"append": true, "comment": null, "createhome": true, "expires": null, "force": false, "generate_ssh_key": null, "group": null, "groups": "admin", "home": null, "login_class": null, "move_home": false, "name": "testuser", "non_unique": false, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "remove": false, "shell": "/bin/bash", "skeleton": null, "ssh_key_bits": "2048", "ssh_key_comment": "ansible-generated on ansibleusers-imac.esri.com", "ssh_key_file": null, "ssh_key_passphrase": null, "ssh_key_type": "rsa", "state": "present", "system": false, "uid": null, "update_password": "always"}, "module_name": "user"}, "move_home": false, "name": "testuser", "password": "NOT_LOGGING_PASSWORD", "shell": "/bin/bash", "state": "present", "uid": 502} changed: [testhost2] => {"append": true, "changed": true, "comment": "", "group": 4294967295, "groups": "admin", "home": "/Users/testuser", "invocation": {"module_args": {"append": true, "comment": null, "createhome": true, "expires": null, "force": false, "generate_ssh_key": null, "group": null, "groups": "admin", "home": null, "login_class": null, "move_home": false, "name": "testuser", "non_unique": false, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "remove": false, "shell": "/bin/bash", "skeleton": null, "ssh_key_bits": "2048", "ssh_key_comment": "ansible-generated on nerumoancer-vm.esri.com", "ssh_key_file": null, "ssh_key_passphrase": null, "ssh_key_type": "rsa", "state": "present", "system": false, "uid": null, "update_password": "always"}, "module_name": "user"}, "move_home": false, "name": "testuser", "password": "NOT_LOGGING_PASSWORD", "shell": "/bin/bash", "state": "present", "uid": 503} ```
True
setting password when creating user causes job state to always be changed - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> user module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> nothing ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> OSX Yosemite, El Cap ##### SUMMARY <!--- Explain the problem briefly --> when specifying a user, if a password is specified the user play is always changed unless update_password: on_create is specified. when providing the hash explicitly, vs the plaintext password, this should not be necessary. this behavior does not exist on RHEL5/6/7. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ``` - name: create user user (osx) user: name: newuser state: present password: "{{ user_hash }}" #update_password: on_create groups: 'admin' append: yes shell: /bin/bash ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Similar behavior to rhel application of this play, where the play only reports changed when there is a change to the system state. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> Play results in state changed unless update_password: on_create is specified. ``` TASK [testuser_account_setup : create testuser user (osx)] ************************* task path: /Users/brad8328/repos/ansible-repo/roles/testuser_account_setup/tasks/main.yml:27 <testhost2> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost2> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost2 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746 `" )'"'"'' <testhost1> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost1> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127 `" )'"'"'' <testhost1> PUT /var/folders/d9/7d_sb4mj1bz1jywsy8rrpj61n4xnb5/T/tmpBEbtI3 TO /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/user <testhost1> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r '[testhost1]' <testhost2> PUT /var/folders/d9/7d_sb4mj1bz1jywsy8rrpj61n4xnb5/T/tmp8mMX8j TO /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/user <testhost2> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r '[testhost2]' <testhost1> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost1> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost1 '/bin/sh -c '"'"'sudo -H -S -p "[sudo via ansible, key=ixheeyzjjhmvtkxfyujaeglkekyibvqs] password: " -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-ixheeyzjjhmvtkxfyujaeglkekyibvqs; /bin/sh -c '"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/user; rm -rf "/Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-221933915191127/" > /dev/null 2>&1'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"''"'"'"'"'"'"'"'"''"'"'' <testhost2> ESTABLISH SSH CONNECTION FOR USER: ansibleuser <testhost2> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ansibleuser -o ConnectTimeout=10 -o ControlPath=/Users/brad8328/.ansible/cp/ansible-ssh-%h-%p-%r -tt testhost2 '/bin/sh -c '"'"'sudo -H -S -p "[sudo via ansible, key=nckvjyuwnrxjapcanzwalcynwzwjeuft] password: " -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-nckvjyuwnrxjapcanzwalcynwzwjeuft; /bin/sh -c '"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/user; rm -rf "/Users/ansibleuser/.ansible/tmp/ansible-tmp-1459977514.94-56562468891746/" > /dev/null 2>&1'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"'"''"'"'"'"'"'"'"'"''"'"'' changed: [testhost1] => {"append": true, "changed": true, "comment": "", "group": 4294967295, "groups": "admin", "home": "/Users/testuser", "invocation": {"module_args": {"append": true, "comment": null, "createhome": true, "expires": null, "force": false, "generate_ssh_key": null, "group": null, "groups": "admin", "home": null, "login_class": null, "move_home": false, "name": "testuser", "non_unique": false, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "remove": false, "shell": "/bin/bash", "skeleton": null, "ssh_key_bits": "2048", "ssh_key_comment": "ansible-generated on ansibleusers-imac.esri.com", "ssh_key_file": null, "ssh_key_passphrase": null, "ssh_key_type": "rsa", "state": "present", "system": false, "uid": null, "update_password": "always"}, "module_name": "user"}, "move_home": false, "name": "testuser", "password": "NOT_LOGGING_PASSWORD", "shell": "/bin/bash", "state": "present", "uid": 502} changed: [testhost2] => {"append": true, "changed": true, "comment": "", "group": 4294967295, "groups": "admin", "home": "/Users/testuser", "invocation": {"module_args": {"append": true, "comment": null, "createhome": true, "expires": null, "force": false, "generate_ssh_key": null, "group": null, "groups": "admin", "home": null, "login_class": null, "move_home": false, "name": "testuser", "non_unique": false, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "remove": false, "shell": "/bin/bash", "skeleton": null, "ssh_key_bits": "2048", "ssh_key_comment": "ansible-generated on nerumoancer-vm.esri.com", "ssh_key_file": null, "ssh_key_passphrase": null, "ssh_key_type": "rsa", "state": "present", "system": false, "uid": null, "update_password": "always"}, "module_name": "user"}, "move_home": false, "name": "testuser", "password": "NOT_LOGGING_PASSWORD", "shell": "/bin/bash", "state": "present", "uid": 503} ```
main
setting password when creating user causes job state to always be changed issue type bug report component name user module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables nothing os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific osx yosemite el cap summary when specifying a user if a password is specified the user play is always changed unless update password on create is specified when providing the hash explicitly vs the plaintext password this should not be necessary this behavior does not exist on steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name create user user osx user name newuser state present password user hash update password on create groups admin append yes shell bin bash expected results similar behavior to rhel application of this play where the play only reports changed when there is a change to the system state actual results play results in state changed unless update password on create is specified task task path users repos ansible repo roles testuser account setup tasks main yml establish ssh connection for user ansibleuser ssh exec ssh c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r tt bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp establish ssh connection for user ansibleuser ssh exec ssh c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r tt bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp put var folders t to users ansibleuser ansible tmp ansible tmp user ssh exec sftp b c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r put var folders t to users ansibleuser ansible tmp ansible tmp user ssh exec sftp b c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r establish ssh connection for user ansibleuser ssh exec ssh c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r tt bin sh c sudo h s p password u root bin sh c echo become success ixheeyzjjhmvtkxfyujaeglkekyibvqs bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ansibleuser ansible tmp ansible tmp user rm rf users ansibleuser ansible tmp ansible tmp dev null establish ssh connection for user ansibleuser ssh exec ssh c vvv o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ansibleuser o connecttimeout o controlpath users ansible cp ansible ssh h p r tt bin sh c sudo h s p password u root bin sh c echo become success nckvjyuwnrxjapcanzwalcynwzwjeuft bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ansibleuser ansible tmp ansible tmp user rm rf users ansibleuser ansible tmp ansible tmp dev null changed append true changed true comment group groups admin home users testuser invocation module args append true comment null createhome true expires null force false generate ssh key null group null groups admin home null login class null move home false name testuser non unique false password value specified in no log parameter remove false shell bin bash skeleton null ssh key bits ssh key comment ansible generated on ansibleusers imac esri com ssh key file null ssh key passphrase null ssh key type rsa state present system false uid null update password always module name user move home false name testuser password not logging password shell bin bash state present uid changed append true changed true comment group groups admin home users testuser invocation module args append true comment null createhome true expires null force false generate ssh key null group null groups admin home null login class null move home false name testuser non unique false password value specified in no log parameter remove false shell bin bash skeleton null ssh key bits ssh key comment ansible generated on nerumoancer vm esri com ssh key file null ssh key passphrase null ssh key type rsa state present system false uid null update password always module name user move home false name testuser password not logging password shell bin bash state present uid
1
788,895
27,772,323,301
IssuesEvent
2023-03-16 15:09:44
AY2223S2-CS2103T-T15-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-T15-1/tp
closed
Implement / Adapt Make command to new command format
priority.High type.Story
As a TTRPG facilitator / player, I can add an entity's default template so that I can track various elements of my TTRPG experience.
1.0
Implement / Adapt Make command to new command format - As a TTRPG facilitator / player, I can add an entity's default template so that I can track various elements of my TTRPG experience.
non_main
implement adapt make command to new command format as a ttrpg facilitator player i can add an entity s default template so that i can track various elements of my ttrpg experience
0
4,969
25,520,686,253
IssuesEvent
2022-11-28 20:11:24
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Question]: Divergent values from documentation and implementation
type: question ❓ status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬
### Question for Carbon Hi folks! I'm working in a project that is using Carbon V11 as the base for our Design System and I noticed a thing related to the typography. Our design team is creating the screens and components on Figma using a Carbon V11 template, so they are basing everything on the typography values there (which are the same as the [documentation](https://carbondesignsystem.com/guidelines/typography/type-sets)). But when implementing the values are different, for instance the `h1` tag: It maps to the `heading-06`, so the line should have the value 3.125rem for height. As in: ![image](https://user-images.githubusercontent.com/105936322/203381552-48ebeba1-b326-4748-9221-e48203cde802.png) But, in reality, it has a height of 1.199rem as defined on @carbon\type\scss\_styles.scss: ![image](https://user-images.githubusercontent.com/105936322/203381973-982726cc-deb7-4b87-9906-b16fa27c2358.png) Also, this 1.199 value doesn't come from any other fluid typography token. This is just an example to bring attention to the problem. So, my question is: am I using something wrong here? Maybe left something behind on the documentation? Is there someone who can help me? If this is a real issue, it either needs to update the documentation or the code, to avoid wrongly designed pages (i. e. design team designs a page expecting typography to behavior a certain way and when the dev team implements it, it will be off). Thanks so far! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
True
[Question]: Divergent values from documentation and implementation - ### Question for Carbon Hi folks! I'm working in a project that is using Carbon V11 as the base for our Design System and I noticed a thing related to the typography. Our design team is creating the screens and components on Figma using a Carbon V11 template, so they are basing everything on the typography values there (which are the same as the [documentation](https://carbondesignsystem.com/guidelines/typography/type-sets)). But when implementing the values are different, for instance the `h1` tag: It maps to the `heading-06`, so the line should have the value 3.125rem for height. As in: ![image](https://user-images.githubusercontent.com/105936322/203381552-48ebeba1-b326-4748-9221-e48203cde802.png) But, in reality, it has a height of 1.199rem as defined on @carbon\type\scss\_styles.scss: ![image](https://user-images.githubusercontent.com/105936322/203381973-982726cc-deb7-4b87-9906-b16fa27c2358.png) Also, this 1.199 value doesn't come from any other fluid typography token. This is just an example to bring attention to the problem. So, my question is: am I using something wrong here? Maybe left something behind on the documentation? Is there someone who can help me? If this is a real issue, it either needs to update the documentation or the code, to avoid wrongly designed pages (i. e. design team designs a page expecting typography to behavior a certain way and when the dev team implements it, it will be off). Thanks so far! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
main
divergent values from documentation and implementation question for carbon hi folks i m working in a project that is using carbon as the base for our design system and i noticed a thing related to the typography our design team is creating the screens and components on figma using a carbon template so they are basing everything on the typography values there which are the same as the but when implementing the values are different for instance the tag it maps to the heading so the line should have the value for height as in but in reality it has a height of as defined on carbon type scss styles scss also this value doesn t come from any other fluid typography token this is just an example to bring attention to the problem so my question is am i using something wrong here maybe left something behind on the documentation is there someone who can help me if this is a real issue it either needs to update the documentation or the code to avoid wrongly designed pages i e design team designs a page expecting typography to behavior a certain way and when the dev team implements it it will be off thanks so far code of conduct i agree to follow this project s
1
259,988
22,583,305,056
IssuesEvent
2022-06-28 13:26:41
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
Follow-ups for test_installAddon migration to react-testing-library
component: testing qa: not needed priority: p3
Follow-ups identified during https://github.com/mozilla/addons-frontend/pull/11612: - Once all tests have been migrated, replace old `getFakeAddonManagerWrapper` implementation with `getFakeAddonManagerWrapperWithJest`, which we will then be able to remove. - Refactor `setCurrentStatus` status to use a parameterize style of testing using `it.each()`: > I'm wondering if we could combine all these tests that set up a state and then check that `setInstallState` has been called with a particular `status`, using `it.each()`? It seems like the setup is close to identical, and then the assertion is also nearly the same, just that we are looking for a different `status`, so they seem like a good candidate for a set of parameterized tests. > > (...) the override for `getAddon` would be part of the parameterized data. You'd only have to include what gets returned by the Promise, I think, so the data would look like `{ isActive: true, isEnabled: false }`, for example. I can see how that might be a bit ugly, but I still think getting rid of all of this duplicate code would make sense. - Migrate tests still marked as `FIXME` that modify the add-on after it has been loaded. > It's definitely not going to work by dispatching actions. You'll need to render a page and then either use `onLocationChanged` or dispatch a `loadAddon` (I generally use the former as it's more like what a user would do). The problem is that `addon` is a direct prop of `InstallButtonWrapper`, and we cannot set a direct prop of an already loaded component. We need to have the parent re-render the component with the new prop value, which is what happens when the test runs from a page. This can be tested from either the `Addon` page or the `AddonVersions` page, so these tests should live in the test suite for one of those - Remove `test_installAddon.js` once all of the above is done.
1.0
Follow-ups for test_installAddon migration to react-testing-library - Follow-ups identified during https://github.com/mozilla/addons-frontend/pull/11612: - Once all tests have been migrated, replace old `getFakeAddonManagerWrapper` implementation with `getFakeAddonManagerWrapperWithJest`, which we will then be able to remove. - Refactor `setCurrentStatus` status to use a parameterize style of testing using `it.each()`: > I'm wondering if we could combine all these tests that set up a state and then check that `setInstallState` has been called with a particular `status`, using `it.each()`? It seems like the setup is close to identical, and then the assertion is also nearly the same, just that we are looking for a different `status`, so they seem like a good candidate for a set of parameterized tests. > > (...) the override for `getAddon` would be part of the parameterized data. You'd only have to include what gets returned by the Promise, I think, so the data would look like `{ isActive: true, isEnabled: false }`, for example. I can see how that might be a bit ugly, but I still think getting rid of all of this duplicate code would make sense. - Migrate tests still marked as `FIXME` that modify the add-on after it has been loaded. > It's definitely not going to work by dispatching actions. You'll need to render a page and then either use `onLocationChanged` or dispatch a `loadAddon` (I generally use the former as it's more like what a user would do). The problem is that `addon` is a direct prop of `InstallButtonWrapper`, and we cannot set a direct prop of an already loaded component. We need to have the parent re-render the component with the new prop value, which is what happens when the test runs from a page. This can be tested from either the `Addon` page or the `AddonVersions` page, so these tests should live in the test suite for one of those - Remove `test_installAddon.js` once all of the above is done.
non_main
follow ups for test installaddon migration to react testing library follow ups identified during once all tests have been migrated replace old getfakeaddonmanagerwrapper implementation with getfakeaddonmanagerwrapperwithjest which we will then be able to remove refactor setcurrentstatus status to use a parameterize style of testing using it each i m wondering if we could combine all these tests that set up a state and then check that setinstallstate has been called with a particular status using it each it seems like the setup is close to identical and then the assertion is also nearly the same just that we are looking for a different status so they seem like a good candidate for a set of parameterized tests the override for getaddon would be part of the parameterized data you d only have to include what gets returned by the promise i think so the data would look like isactive true isenabled false for example i can see how that might be a bit ugly but i still think getting rid of all of this duplicate code would make sense migrate tests still marked as fixme that modify the add on after it has been loaded it s definitely not going to work by dispatching actions you ll need to render a page and then either use onlocationchanged or dispatch a loadaddon i generally use the former as it s more like what a user would do the problem is that addon is a direct prop of installbuttonwrapper and we cannot set a direct prop of an already loaded component we need to have the parent re render the component with the new prop value which is what happens when the test runs from a page this can be tested from either the addon page or the addonversions page so these tests should live in the test suite for one of those remove test installaddon js once all of the above is done
0
3,518
13,781,550,963
IssuesEvent
2020-10-08 16:19:53
NaluKit/nalu
https://api.github.com/repos/NaluKit/nalu
closed
redirectTo-method of filter gets called twice
maintainance
The `redirect`-method of a Filter-class gets called twice and the 'getParameter'-method four times. This might be confusing, in case you set a breakpoint inside the method. To avoid confusing, optimize the code, so that the methed gets called only one time.
True
redirectTo-method of filter gets called twice - The `redirect`-method of a Filter-class gets called twice and the 'getParameter'-method four times. This might be confusing, in case you set a breakpoint inside the method. To avoid confusing, optimize the code, so that the methed gets called only one time.
main
redirectto method of filter gets called twice the redirect method of a filter class gets called twice and the getparameter method four times this might be confusing in case you set a breakpoint inside the method to avoid confusing optimize the code so that the methed gets called only one time
1
36,102
9,745,265,560
IssuesEvent
2019-06-03 09:12:54
groupe-sii/ogham
https://api.github.com/repos/groupe-sii/ogham
opened
Provide configurers for mainly used service providers
API Builder P4 enhancement
Provide a simple automatic way to configure email sending through Gmail, Yahoo, ... It could be something like this with fluent API: ```java .email() .gmail() .account() .username() .applicationPassword() ``` An using properties: ``` ogham.email.gmail.account.username= ogham.email.gmail.account.application-password= ``` And nothing else to do. It will automatically configure correctly the SMTP configuration
1.0
Provide configurers for mainly used service providers - Provide a simple automatic way to configure email sending through Gmail, Yahoo, ... It could be something like this with fluent API: ```java .email() .gmail() .account() .username() .applicationPassword() ``` An using properties: ``` ogham.email.gmail.account.username= ogham.email.gmail.account.application-password= ``` And nothing else to do. It will automatically configure correctly the SMTP configuration
non_main
provide configurers for mainly used service providers provide a simple automatic way to configure email sending through gmail yahoo it could be something like this with fluent api java email gmail account username applicationpassword an using properties ogham email gmail account username ogham email gmail account application password and nothing else to do it will automatically configure correctly the smtp configuration
0
462,204
13,243,011,871
IssuesEvent
2020-08-19 10:41:38
magento/magento2
https://api.github.com/repos/magento/magento2
closed
[Issue] improve exception handling in Layout render
Component: View Fixed in 2.4.x Issue: Confirmed Issue: Format is valid Issue: Ready for Work Priority: P2 Progress: ready for dev Reproduced on 2.4.x Severity: S3
This issue is automatically created based on existing pull request: magento/magento2#27478: improve exception handling in Layout render --------- ### Preconditions (*) When layout is being rendered in production mode, all exception messages are logged to system log as critical issues. However exception stack is lost. If we pass exception to `LoggerInterface::critical` instead this data will be saved to `var/report`. ### Steps to reproduce (*) 1. Add `throw \Exception('test');` in a template. 2. Validate that after rendering such a template, exception stack is being saved to `var/report` in addition to its message being logged. ### Actual Result: (*) :heavy_multiplication_x: All exception messages are logged to system log ### Expected Result: :heavy_check_mark: Messages should be logged to the exception log file ### Questions or comments 1. Is passing `Exception` to `LoggerInterface::critical` a preferred way of logging issues? (I see it being used this way in other places) 2. Was saving non translated message in case of `LocalizedException` adding any value? ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] All automated tests passed successfully (all builds are green)
1.0
[Issue] improve exception handling in Layout render - This issue is automatically created based on existing pull request: magento/magento2#27478: improve exception handling in Layout render --------- ### Preconditions (*) When layout is being rendered in production mode, all exception messages are logged to system log as critical issues. However exception stack is lost. If we pass exception to `LoggerInterface::critical` instead this data will be saved to `var/report`. ### Steps to reproduce (*) 1. Add `throw \Exception('test');` in a template. 2. Validate that after rendering such a template, exception stack is being saved to `var/report` in addition to its message being logged. ### Actual Result: (*) :heavy_multiplication_x: All exception messages are logged to system log ### Expected Result: :heavy_check_mark: Messages should be logged to the exception log file ### Questions or comments 1. Is passing `Exception` to `LoggerInterface::critical` a preferred way of logging issues? (I see it being used this way in other places) 2. Was saving non translated message in case of `LocalizedException` adding any value? ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] All automated tests passed successfully (all builds are green)
non_main
improve exception handling in layout render this issue is automatically created based on existing pull request magento improve exception handling in layout render preconditions when layout is being rendered in production mode all exception messages are logged to system log as critical issues however exception stack is lost if we pass exception to loggerinterface critical instead this data will be saved to var report steps to reproduce add throw exception test in a template validate that after rendering such a template exception stack is being saved to var report in addition to its message being logged actual result heavy multiplication x all exception messages are logged to system log expected result heavy check mark messages should be logged to the exception log file questions or comments is passing exception to loggerinterface critical a preferred way of logging issues i see it being used this way in other places was saving non translated message in case of localizedexception adding any value contribution checklist pull request has a meaningful description of its purpose all commits are accompanied by meaningful commit messages all new or changed code is covered with unit integration tests if applicable all automated tests passed successfully all builds are green
0
233
2,910,269,419
IssuesEvent
2015-06-21 15:50:20
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
closed
can't run win_post_install from pip
Maintainability Windows
please add this snippet to spyder_win_post_install.py to mimic bdist_wininst. I cribbed it from pywin32_post_install.py. Thanks! try: # When this script is run from inside the bdist_wininst installer, # file_created() and directory_created() are additional builtin # functions which write lines to Python23\pywin32-install.log. This is # a list of actions for the uninstaller, the format is inspired by what # the Wise installer also creates. file_created is_bdist_wininst = True except NameError: is_bdist_wininst = False # we know what it is not - but not what it is :) def file_created(file): pass def directory_created(directory): pass def get_root_hkey(): try: winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, root_key_name, 0, winreg.KEY_CREATE_SUB_KEY) return winreg.HKEY_LOCAL_MACHINE except OSError, details: # Either not exist, or no permissions to create subkey means # must be HKCU return winreg.HKEY_CURRENT_USER try: create_shortcut except NameError: # Create a function with the same signature as create_shortcut provided # by bdist_wininst def create_shortcut(path, description, filename, arguments="", workdir="", iconpath="", iconindex=0): import pythoncom from win32com.shell import shell, shellcon ilink = pythoncom.CoCreateInstance(shell.CLSID_ShellLink, None, pythoncom.CLSCTX_INPROC_SERVER, shell.IID_IShellLink) ilink.SetPath(path) ilink.SetDescription(description) if arguments: ilink.SetArguments(arguments) if workdir: ilink.SetWorkingDirectory(workdir) if iconpath or iconindex: ilink.SetIconLocation(iconpath, iconindex) # now save it. ipf = ilink.QueryInterface(pythoncom.IID_IPersistFile) ipf.Save(filename, 0) # Support the same list of "path names" as bdist_wininst. def get_special_folder_path(path_name): import pythoncom from win32com.shell import shell, shellcon for maybe in """ CSIDL_COMMON_STARTMENU CSIDL_STARTMENU CSIDL_COMMON_APPDATA CSIDL_LOCAL_APPDATA CSIDL_APPDATA CSIDL_COMMON_DESKTOPDIRECTORY CSIDL_DESKTOPDIRECTORY CSIDL_COMMON_STARTUP CSIDL_STARTUP CSIDL_COMMON_PROGRAMS CSIDL_PROGRAMS CSIDL_PROGRAM_FILES_COMMON CSIDL_PROGRAM_FILES CSIDL_FONTS""".split(): if maybe == path_name: csidl = getattr(shellcon, maybe) return shell.SHGetSpecialFolderPath(0, csidl, False) raise ValueError("%s is an unknown path ID" % (path_name,))
True
can't run win_post_install from pip - please add this snippet to spyder_win_post_install.py to mimic bdist_wininst. I cribbed it from pywin32_post_install.py. Thanks! try: # When this script is run from inside the bdist_wininst installer, # file_created() and directory_created() are additional builtin # functions which write lines to Python23\pywin32-install.log. This is # a list of actions for the uninstaller, the format is inspired by what # the Wise installer also creates. file_created is_bdist_wininst = True except NameError: is_bdist_wininst = False # we know what it is not - but not what it is :) def file_created(file): pass def directory_created(directory): pass def get_root_hkey(): try: winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, root_key_name, 0, winreg.KEY_CREATE_SUB_KEY) return winreg.HKEY_LOCAL_MACHINE except OSError, details: # Either not exist, or no permissions to create subkey means # must be HKCU return winreg.HKEY_CURRENT_USER try: create_shortcut except NameError: # Create a function with the same signature as create_shortcut provided # by bdist_wininst def create_shortcut(path, description, filename, arguments="", workdir="", iconpath="", iconindex=0): import pythoncom from win32com.shell import shell, shellcon ilink = pythoncom.CoCreateInstance(shell.CLSID_ShellLink, None, pythoncom.CLSCTX_INPROC_SERVER, shell.IID_IShellLink) ilink.SetPath(path) ilink.SetDescription(description) if arguments: ilink.SetArguments(arguments) if workdir: ilink.SetWorkingDirectory(workdir) if iconpath or iconindex: ilink.SetIconLocation(iconpath, iconindex) # now save it. ipf = ilink.QueryInterface(pythoncom.IID_IPersistFile) ipf.Save(filename, 0) # Support the same list of "path names" as bdist_wininst. def get_special_folder_path(path_name): import pythoncom from win32com.shell import shell, shellcon for maybe in """ CSIDL_COMMON_STARTMENU CSIDL_STARTMENU CSIDL_COMMON_APPDATA CSIDL_LOCAL_APPDATA CSIDL_APPDATA CSIDL_COMMON_DESKTOPDIRECTORY CSIDL_DESKTOPDIRECTORY CSIDL_COMMON_STARTUP CSIDL_STARTUP CSIDL_COMMON_PROGRAMS CSIDL_PROGRAMS CSIDL_PROGRAM_FILES_COMMON CSIDL_PROGRAM_FILES CSIDL_FONTS""".split(): if maybe == path_name: csidl = getattr(shellcon, maybe) return shell.SHGetSpecialFolderPath(0, csidl, False) raise ValueError("%s is an unknown path ID" % (path_name,))
main
can t run win post install from pip please add this snippet to spyder win post install py to mimic bdist wininst i cribbed it from post install py thanks try when this script is run from inside the bdist wininst installer file created and directory created are additional builtin functions which write lines to install log this is a list of actions for the uninstaller the format is inspired by what the wise installer also creates file created is bdist wininst true except nameerror is bdist wininst false we know what it is not but not what it is def file created file pass def directory created directory pass def get root hkey try winreg openkey winreg hkey local machine root key name winreg key create sub key return winreg hkey local machine except oserror details either not exist or no permissions to create subkey means must be hkcu return winreg hkey current user try create shortcut except nameerror create a function with the same signature as create shortcut provided by bdist wininst def create shortcut path description filename arguments workdir iconpath iconindex import pythoncom from shell import shell shellcon ilink pythoncom cocreateinstance shell clsid shelllink none pythoncom clsctx inproc server shell iid ishelllink ilink setpath path ilink setdescription description if arguments ilink setarguments arguments if workdir ilink setworkingdirectory workdir if iconpath or iconindex ilink seticonlocation iconpath iconindex now save it ipf ilink queryinterface pythoncom iid ipersistfile ipf save filename support the same list of path names as bdist wininst def get special folder path path name import pythoncom from shell import shell shellcon for maybe in csidl common startmenu csidl startmenu csidl common appdata csidl local appdata csidl appdata csidl common desktopdirectory csidl desktopdirectory csidl common startup csidl startup csidl common programs csidl programs csidl program files common csidl program files csidl fonts split if maybe path name csidl getattr shellcon maybe return shell shgetspecialfolderpath csidl false raise valueerror s is an unknown path id path name
1
931
4,643,416,228
IssuesEvent
2016-09-30 13:26:16
duckduckgo/zeroclickinfo-fathead
https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead
closed
Python: show built-in functions
Maintainer Input Requested Mission: Programming
In #270 we removed some entries to avoid namespace conflicts. We might want to white list the built-in functions. https://docs.python.org/3.5/library/functions.html#min For example, even though there are likely conflicts for `open()` users are most likely looking for the built-in. https://duckduckgo.com/?q=python+open&t=canonical&atb=v23&ia=qa https://docs.python.org/3.5/library/functions.html#open ------ IA Page: http://duck.co/ia/view/python [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @gaulrobe
True
Python: show built-in functions - In #270 we removed some entries to avoid namespace conflicts. We might want to white list the built-in functions. https://docs.python.org/3.5/library/functions.html#min For example, even though there are likely conflicts for `open()` users are most likely looking for the built-in. https://duckduckgo.com/?q=python+open&t=canonical&atb=v23&ia=qa https://docs.python.org/3.5/library/functions.html#open ------ IA Page: http://duck.co/ia/view/python [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @gaulrobe
main
python show built in functions in we removed some entries to avoid namespace conflicts we might want to white list the built in functions for example even though there are likely conflicts for open users are most likely looking for the built in ia page gaulrobe
1
253,606
19,143,117,050
IssuesEvent
2021-12-02 02:42:45
Peedrooo/PortfolioAI
https://api.github.com/repos/Peedrooo/PortfolioAI
closed
00.05 - Pull request template
documentation
## Create pull request template - Details of all changes and objectives with that - Conflict - Title must been associete to theme - Fields
1.0
00.05 - Pull request template - ## Create pull request template - Details of all changes and objectives with that - Conflict - Title must been associete to theme - Fields
non_main
pull request template create pull request template details of all changes and objectives with that conflict title must been associete to theme fields
0
378,798
11,208,790,456
IssuesEvent
2020-01-06 08:53:47
yalla-coop/earwig
https://api.github.com/repos/yalla-coop/earwig
closed
Updates to level 2 > level 3 flow
backlog priority-2
@thejoefriel We need to make the following changes to when a level 2 user (non-worker) tries to give a review and has to get verified. Please refer to the page called 'Updated giving review journeys' in the wireframes. - [x] Add new 'Verification required' page - [x] Remove 'Important message for workers' page - [x] After verification, ensure the giving review page follows - [x] If user cancels at any point, return them to the entity profile page - [x] If user cancels after uploading verification photo, keep them verified but return them to the entity profile page Ensure consistency across both mobile and desktop
1.0
Updates to level 2 > level 3 flow - @thejoefriel We need to make the following changes to when a level 2 user (non-worker) tries to give a review and has to get verified. Please refer to the page called 'Updated giving review journeys' in the wireframes. - [x] Add new 'Verification required' page - [x] Remove 'Important message for workers' page - [x] After verification, ensure the giving review page follows - [x] If user cancels at any point, return them to the entity profile page - [x] If user cancels after uploading verification photo, keep them verified but return them to the entity profile page Ensure consistency across both mobile and desktop
non_main
updates to level level flow thejoefriel we need to make the following changes to when a level user non worker tries to give a review and has to get verified please refer to the page called updated giving review journeys in the wireframes add new verification required page remove important message for workers page after verification ensure the giving review page follows if user cancels at any point return them to the entity profile page if user cancels after uploading verification photo keep them verified but return them to the entity profile page ensure consistency across both mobile and desktop
0
147,859
11,810,553,723
IssuesEvent
2020-03-19 16:40:25
vmware/tern
https://api.github.com/repos/vmware/tern
closed
Run tests using `python3 setup.py test`
tests tools
**Problem Statement** The command `python3 setup.py test` is not running the tests. **Describe the Proposal** `python3 setup.py test` should be able to run all the tests. Such functionality will help the developers to know if they are breaking anything with their new features. **Steps to Implement Proposal** 1. Reorganise import statements in the test files. 2. Add a function in setup.py to load test_suite
1.0
Run tests using `python3 setup.py test` - **Problem Statement** The command `python3 setup.py test` is not running the tests. **Describe the Proposal** `python3 setup.py test` should be able to run all the tests. Such functionality will help the developers to know if they are breaking anything with their new features. **Steps to Implement Proposal** 1. Reorganise import statements in the test files. 2. Add a function in setup.py to load test_suite
non_main
run tests using setup py test problem statement the command setup py test is not running the tests describe the proposal setup py test should be able to run all the tests such functionality will help the developers to know if they are breaking anything with their new features steps to implement proposal reorganise import statements in the test files add a function in setup py to load test suite
0
3,563
14,263,444,274
IssuesEvent
2020-11-20 14:27:01
precice/precice
https://api.github.com/repos/precice/precice
closed
Convert vector and map access pattern in tests
maintainability
As pointed out by @fsimonis here: https://github.com/precice/precice/pull/894#issuecomment-722626356 > We should definitely convert the vector and map access from c[i] to c.at(i). I guess we should do this for all tests. Not sure what extent this will have.
True
Convert vector and map access pattern in tests - As pointed out by @fsimonis here: https://github.com/precice/precice/pull/894#issuecomment-722626356 > We should definitely convert the vector and map access from c[i] to c.at(i). I guess we should do this for all tests. Not sure what extent this will have.
main
convert vector and map access pattern in tests as pointed out by fsimonis here we should definitely convert the vector and map access from c to c at i i guess we should do this for all tests not sure what extent this will have
1
3,003
11,045,512,383
IssuesEvent
2019-12-09 15:14:39
18F/cg-product
https://api.github.com/repos/18F/cg-product
closed
As a cloud.gov operator, I want to be able to deploy the AWS service broker to staging.
contractor-3-maintainability contractor-4-resilience
In order to test the AWS service broker for redis and elastic search as a replacement for Kubernetes, we need the broker deployed to staging. ## Acceptance Criteria * the AWS service broker at https://github.com/awslabs/aws-servicebroker is deployed to the staging environment via a concourse pipeline * the redis and elastic search offerings are made available in the marketplace via a test org. * the names of the redis and elastic search offerings do not clash with existing offerings in the marketplace. ## Implementation sketch Pipeline: * [ ] ingests the broker release as a resource: https://github.com/awslabs/aws-servicebroker/releases * [ ] deploys the broker as an application to cloud.gov (including necessary services) * [ ] registers the service broker with the cloud controller * [ ] makes the redis and elastic search offerings available in a test organization Repository: https://github.com/18F/cg-deploy-aws-servicebroker
True
As a cloud.gov operator, I want to be able to deploy the AWS service broker to staging. - In order to test the AWS service broker for redis and elastic search as a replacement for Kubernetes, we need the broker deployed to staging. ## Acceptance Criteria * the AWS service broker at https://github.com/awslabs/aws-servicebroker is deployed to the staging environment via a concourse pipeline * the redis and elastic search offerings are made available in the marketplace via a test org. * the names of the redis and elastic search offerings do not clash with existing offerings in the marketplace. ## Implementation sketch Pipeline: * [ ] ingests the broker release as a resource: https://github.com/awslabs/aws-servicebroker/releases * [ ] deploys the broker as an application to cloud.gov (including necessary services) * [ ] registers the service broker with the cloud controller * [ ] makes the redis and elastic search offerings available in a test organization Repository: https://github.com/18F/cg-deploy-aws-servicebroker
main
as a cloud gov operator i want to be able to deploy the aws service broker to staging in order to test the aws service broker for redis and elastic search as a replacement for kubernetes we need the broker deployed to staging acceptance criteria the aws service broker at is deployed to the staging environment via a concourse pipeline the redis and elastic search offerings are made available in the marketplace via a test org the names of the redis and elastic search offerings do not clash with existing offerings in the marketplace implementation sketch pipeline ingests the broker release as a resource deploys the broker as an application to cloud gov including necessary services registers the service broker with the cloud controller makes the redis and elastic search offerings available in a test organization repository
1
169,825
14,233,463,415
IssuesEvent
2020-11-18 12:12:06
MaibornWolff/clean-air
https://api.github.com/repos/MaibornWolff/clean-air
closed
Restructure documentation
documentation
**Is your feature request related to a problem? Please describe.** As a clean air user I want have a structured documentation section so I can build the software and the hardware
1.0
Restructure documentation - **Is your feature request related to a problem? Please describe.** As a clean air user I want have a structured documentation section so I can build the software and the hardware
non_main
restructure documentation is your feature request related to a problem please describe as a clean air user i want have a structured documentation section so i can build the software and the hardware
0
33,300
9,098,500,899
IssuesEvent
2019-02-20 00:12:37
gini/dexter
https://api.github.com/repos/gini/dexter
closed
ID and Secret-id is still required even when the binary has been built with the said parameters
build question
Hey Team, I tried out Dexter on an Ubuntu 18 host and looks to be id & secret id are still required to get a successful login. Built the binary with following parameters: `CLIENT_ID=529839178945-toh9cj642someid7hvu7id9n5.apps.googleusercontent.com CLIENT_SECRET=VzK7vrksome_secret3ts6AfZa6Diz OS=linux make` When a login attempt being made via dexter as follows getting a http 400. `$dexter auth ` However when all the parameters are specified as follows can log in just fine. `dexter auth -i xxxxxx -s ffdfdff` Appreciate your feedback guys, could be something that I might have missed.
1.0
ID and Secret-id is still required even when the binary has been built with the said parameters - Hey Team, I tried out Dexter on an Ubuntu 18 host and looks to be id & secret id are still required to get a successful login. Built the binary with following parameters: `CLIENT_ID=529839178945-toh9cj642someid7hvu7id9n5.apps.googleusercontent.com CLIENT_SECRET=VzK7vrksome_secret3ts6AfZa6Diz OS=linux make` When a login attempt being made via dexter as follows getting a http 400. `$dexter auth ` However when all the parameters are specified as follows can log in just fine. `dexter auth -i xxxxxx -s ffdfdff` Appreciate your feedback guys, could be something that I might have missed.
non_main
id and secret id is still required even when the binary has been built with the said parameters hey team i tried out dexter on an ubuntu host and looks to be id secret id are still required to get a successful login built the binary with following parameters client id apps googleusercontent com client secret os linux make when a login attempt being made via dexter as follows getting a http dexter auth however when all the parameters are specified as follows can log in just fine dexter auth i xxxxxx s ffdfdff appreciate your feedback guys could be something that i might have missed
0
774,369
27,193,436,796
IssuesEvent
2023-02-20 01:39:24
KevCaz/KevCaz.github.io
https://api.github.com/repos/KevCaz/KevCaz.github.io
closed
New icons/emoticons to categorize movies watched
enhancement low priority
e.g. - :trophy: : favorite - :laughing: : fun - :repeat: to be watched more than once ...
1.0
New icons/emoticons to categorize movies watched - e.g. - :trophy: : favorite - :laughing: : fun - :repeat: to be watched more than once ...
non_main
new icons emoticons to categorize movies watched e g trophy favorite laughing fun repeat to be watched more than once
0
269,861
20,509,304,691
IssuesEvent
2022-03-01 03:30:00
UToledo-SeniorDesign/Diabetes-Management-Mobile-App
https://api.github.com/repos/UToledo-SeniorDesign/Diabetes-Management-Mobile-App
opened
Discuss how to handle updating meal(s)
bug documentation
## Overview Currently we do not know how we handle updating a meal. Example: if the user selects a past meal to use `now` as their current meal, and they submit (without making any changes to it), we have no way of know that this meal's data is already in the table and therefore we are creating a new meal entry instead of updating the meal they selected in the DB. We need to discuss what would be a good way to handle this case, maybe even create security tests to ensure this logic works, and document any edge cases we can think of in some google doc/notion/readme file.
1.0
Discuss how to handle updating meal(s) - ## Overview Currently we do not know how we handle updating a meal. Example: if the user selects a past meal to use `now` as their current meal, and they submit (without making any changes to it), we have no way of know that this meal's data is already in the table and therefore we are creating a new meal entry instead of updating the meal they selected in the DB. We need to discuss what would be a good way to handle this case, maybe even create security tests to ensure this logic works, and document any edge cases we can think of in some google doc/notion/readme file.
non_main
discuss how to handle updating meal s overview currently we do not know how we handle updating a meal example if the user selects a past meal to use now as their current meal and they submit without making any changes to it we have no way of know that this meal s data is already in the table and therefore we are creating a new meal entry instead of updating the meal they selected in the db we need to discuss what would be a good way to handle this case maybe even create security tests to ensure this logic works and document any edge cases we can think of in some google doc notion readme file
0
4,536
23,616,490,067
IssuesEvent
2022-08-24 16:18:16
freedomofpress/securedrop-client
https://api.github.com/repos/freedomofpress/securedrop-client
closed
Identifying test failures in the CI pipeline is attention-consuming
maintainer quality of life :gear: Tooling :beach_umbrella: Summer cleanup
## Description When any test fails in the CI pipeline, the `test` job fails. That's expected. However, if you want to determine which test failed (or which requirement wasn't met, it could be a linting issue), you need to scroll through the output of one single long step. I propose splitting the testing tasks into multiple CI **jobs**. Current: - build - test - setup - lint-and-test-all-the-things Proposed: - build - lint, type check, etc. - setup - lint, type check, etc. - unit test - setup - test (`make test`) - integration test - setup - test (`make test-integration`) - functional test - setup - test (`make test-functional`) ## Considerations ### Trade-offs CI builds duration can be measured using "wall clock time", and "CPU time". One advantage of grouping all tasks as we do currently is keeping the "CPU time" minimal. That's good for the planet in terms or energy consumption :earth_africa:, and may be cheaper. Splitting the tasks into more jobs may (or may not, it depends) result in a decrease in "wall clock time", which means people have to wait less for CI results. However it often results in an increase of "CPU time" because some setup has to be repeated across jobs (which are typically run independently). It's main value lays in the readability of the results, and a decrease of the time spent looking for answers. _I don't take increasing CPU time lightly (:earth_americas:), but I would give a try to splitting jobs, see how we like it before assuming that the trade-off it not worth it._ ### Impact Developers could rely on CI builds more effectively to narrow down troubleshooting. It's a day-to-day quality of life improvement. ### Security There are no implications for the threat model, because we'd still be running the exact same checks, and those are not order-dependent.
True
Identifying test failures in the CI pipeline is attention-consuming - ## Description When any test fails in the CI pipeline, the `test` job fails. That's expected. However, if you want to determine which test failed (or which requirement wasn't met, it could be a linting issue), you need to scroll through the output of one single long step. I propose splitting the testing tasks into multiple CI **jobs**. Current: - build - test - setup - lint-and-test-all-the-things Proposed: - build - lint, type check, etc. - setup - lint, type check, etc. - unit test - setup - test (`make test`) - integration test - setup - test (`make test-integration`) - functional test - setup - test (`make test-functional`) ## Considerations ### Trade-offs CI builds duration can be measured using "wall clock time", and "CPU time". One advantage of grouping all tasks as we do currently is keeping the "CPU time" minimal. That's good for the planet in terms or energy consumption :earth_africa:, and may be cheaper. Splitting the tasks into more jobs may (or may not, it depends) result in a decrease in "wall clock time", which means people have to wait less for CI results. However it often results in an increase of "CPU time" because some setup has to be repeated across jobs (which are typically run independently). It's main value lays in the readability of the results, and a decrease of the time spent looking for answers. _I don't take increasing CPU time lightly (:earth_americas:), but I would give a try to splitting jobs, see how we like it before assuming that the trade-off it not worth it._ ### Impact Developers could rely on CI builds more effectively to narrow down troubleshooting. It's a day-to-day quality of life improvement. ### Security There are no implications for the threat model, because we'd still be running the exact same checks, and those are not order-dependent.
main
identifying test failures in the ci pipeline is attention consuming description when any test fails in the ci pipeline the test job fails that s expected however if you want to determine which test failed or which requirement wasn t met it could be a linting issue you need to scroll through the output of one single long step i propose splitting the testing tasks into multiple ci jobs current build test setup lint and test all the things proposed build lint type check etc setup lint type check etc unit test setup test make test integration test setup test make test integration functional test setup test make test functional considerations trade offs ci builds duration can be measured using wall clock time and cpu time one advantage of grouping all tasks as we do currently is keeping the cpu time minimal that s good for the planet in terms or energy consumption earth africa and may be cheaper splitting the tasks into more jobs may or may not it depends result in a decrease in wall clock time which means people have to wait less for ci results however it often results in an increase of cpu time because some setup has to be repeated across jobs which are typically run independently it s main value lays in the readability of the results and a decrease of the time spent looking for answers i don t take increasing cpu time lightly earth americas but i would give a try to splitting jobs see how we like it before assuming that the trade off it not worth it impact developers could rely on ci builds more effectively to narrow down troubleshooting it s a day to day quality of life improvement security there are no implications for the threat model because we d still be running the exact same checks and those are not order dependent
1
1,967
6,694,168,257
IssuesEvent
2017-10-10 00:03:19
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Reddit Search: option to filter results - "top", "new", "relevance" reddit results
Maintainer Input Requested
Currently, IA shows results which are sorted according to "relevancy" by default (if I understood it correct). It would be awesome if we could make an option on IA results to sort according to "top", "new" reddit results. --- IA Page: http://duck.co/ia/view/reddit_search [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mattr555
True
Reddit Search: option to filter results - "top", "new", "relevance" reddit results - Currently, IA shows results which are sorted according to "relevancy" by default (if I understood it correct). It would be awesome if we could make an option on IA results to sort according to "top", "new" reddit results. --- IA Page: http://duck.co/ia/view/reddit_search [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mattr555
main
reddit search option to filter results top new relevance reddit results currently ia shows results which are sorted according to relevancy by default if i understood it correct it would be awesome if we could make an option on ia results to sort according to top new reddit results ia page
1
1,308
5,557,372,032
IssuesEvent
2017-03-24 11:51:35
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Stat tracking is completely unmaintainable
Maintainability/Hinders improvements Not a bug
Still hosted by erro, nobody is quite sure how it works, can't add new stuff to it, Cheridan still has to manually update it. Stat tracking is the only thing that injects the barr minimum of objectivety and rationality into balance shouting matches, it really needs to be overhauled badly.
True
Stat tracking is completely unmaintainable - Still hosted by erro, nobody is quite sure how it works, can't add new stuff to it, Cheridan still has to manually update it. Stat tracking is the only thing that injects the barr minimum of objectivety and rationality into balance shouting matches, it really needs to be overhauled badly.
main
stat tracking is completely unmaintainable still hosted by erro nobody is quite sure how it works can t add new stuff to it cheridan still has to manually update it stat tracking is the only thing that injects the barr minimum of objectivety and rationality into balance shouting matches it really needs to be overhauled badly
1
625,590
19,758,158,410
IssuesEvent
2022-01-16 00:29:50
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
usb: stm32: testusb fails and often hangs forever
bug priority: low area: Drivers area: USB platform: STM32 Stale
**Describe the bug** On the Nucleo F767zi and Nucleo F070rb boards, many test cases from the testusb Linux kernel module fail. Failure rates and their consequences can take various forms: most failures only result in a timeout for a specific tranfer, but some of the failures can render the device unusable, forcing the user to reset the MCU in order to use the USB device again. The following table summarize the status of each test case for the two boards. | Test case | nucleo_f767zi | Failure Rate | nucleo_f070rb | Failure Rate | |:---------:|:--------------------:|:------------:|:--------------------:|:------------:| | 0 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 1 | :x: | 10-20% | :heavy_check_mark: | N/A | | 2 | :heavy_check_mark: | N/A | :x: | 100% | | 3 | :x: | 10-20% | :heavy_check_mark: | N/A | | 4 | :heavy_check_mark: | N/A | :x: | 100% | | 5 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 6 | :heavy_check_mark: | N/A | :x: | 100% | | 7 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 8 | :heavy_check_mark: | N/A | :x: | 100% | | 9 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 10 | :heavy_check_mark: | N/A | :x: | 100% | | 11 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 12 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 13 | :x: | 100% | :x: | 100% | | 14 | :x: | < 10% | :x: | 100% | | 17 | :x: | 10-20% | :heavy_check_mark: | N/A | | 18 | :heavy_check_mark: | N/A | :x: | 100% | | 19 | :x: | 10-20% | :heavy_check_mark: | N/A | | 20 | :heavy_check_mark: | N/A | :x: | 100% | | 21 | :x: | < 10% | :x: | 100% | | 24 | :x: | 100% | :heavy_check_mark: | N/A | | 27 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 28 | :heavy_check_mark: | N/A | :x: | 100% | | 29 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | I used the following command line for all the tests: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` On the nucleo_f767zi board: - test cases 1, 3, 17 and 19 fail between 10 and 20% of the time. When it fails, the test end with "Connection timed out". The device is still usable after that failure. - test case 13 always fail with "Invalid argument". It doesn't have any other impact. - test cases 14 and 21 seem to trigger issue #27992 . Indeed, after enough time, these tests fail and the control endpoints become unresponsive. Unplugging the cable and plugging it again won't solve the issue: the device is unable to enumerate and the MCU has to be reset in order to have a functional USB device. - test case 24 always fails and is unable to complete. The testusb program stays stuck in the kernel in an uninterruptible sleep state. The device is still operational after plugging it again. On the nucleo_f070rb board: - test cases 2, 4, 6, 8, 14, 18, 20, 21 always fail and return "Connection timed out". The device can operate normally after that. - test case 13 always fail with "Invalid argument". - test cases 10 and 28 always fail and are stuck in the kernel indefinitely. The device needs to plugged again in order to On both boards, whenever a test fails, the log of the device shows: ``` <err> usb_dc_stm32: Unable to get write lock (-16) <err> usb_dc_stm32: Unable to get write lock (-16) ``` Changing the value of `USB_NUMOF_EP_WRITE_RETRIES` from 3 to 30 doesn't seem to have any impact on the tests failures. **To Reproduce** For the nucleo_f767zi board: - Compile and flash the testusb sample - On the Linux host, follow the instructions from `sample/subsys/usb/testusb/README.rst` - For each test case, run the following command: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` Since some test cases only fail around 10% of the time, this command may need to be executed multiple times before the failure occurs. For the nucleo_f070rb board: - Since that board doesn't have a USB connector, you'll need an external micro USB connector (the PA11 and PA12 pins can be used as USB D+/D-) - Once you are able to connect a Linux host to the board, you need to set the status of the "usb" node of the device tree to "okay": ``` diff --git a/boards/arm/nucleo_f070rb/nucleo_f070rb.dts b/boards/arm/nucleo_f070rb/nucleo_f070rb.dts index 9f5923d36d57..dcdf20912124 100644 --- a/boards/arm/nucleo_f070rb/nucleo_f070rb.dts +++ b/boards/arm/nucleo_f070rb/nucleo_f070rb.dts @@ -41,6 +41,10 @@ }; }; +&usb { + status = "okay"; +}; + &usart1 { current-speed = <115200>; status = "okay"; ``` - Compile and flash the testusb sample - On the Linux host, follow the instructions from `sample/subsys/usb/testusb/README.rst` - Use the command: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` **Expected behavior** All tests should pass. **Impact** There are multiple consequences to these failures: - The USB device might not pass USB certification - Some transfers randomly time out which compromise the communication with a host. - In some cases, USB is completely unusable without a MCU hardware reset, which is not acceptable or even possible in production. **Environment (please complete the following information):** - Host: Ubuntu 19.10 with Linux kernel 5.3.0-64-generic - Toolchain : Zephyr SDK - Commit SHA: c8cca16c204543b8f6dd7261c643b48ba8c95a94 - Boards: nucleo_f767zi and nucleo_f070rb (with external usb connector)
1.0
usb: stm32: testusb fails and often hangs forever - **Describe the bug** On the Nucleo F767zi and Nucleo F070rb boards, many test cases from the testusb Linux kernel module fail. Failure rates and their consequences can take various forms: most failures only result in a timeout for a specific tranfer, but some of the failures can render the device unusable, forcing the user to reset the MCU in order to use the USB device again. The following table summarize the status of each test case for the two boards. | Test case | nucleo_f767zi | Failure Rate | nucleo_f070rb | Failure Rate | |:---------:|:--------------------:|:------------:|:--------------------:|:------------:| | 0 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 1 | :x: | 10-20% | :heavy_check_mark: | N/A | | 2 | :heavy_check_mark: | N/A | :x: | 100% | | 3 | :x: | 10-20% | :heavy_check_mark: | N/A | | 4 | :heavy_check_mark: | N/A | :x: | 100% | | 5 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 6 | :heavy_check_mark: | N/A | :x: | 100% | | 7 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 8 | :heavy_check_mark: | N/A | :x: | 100% | | 9 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 10 | :heavy_check_mark: | N/A | :x: | 100% | | 11 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 12 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 13 | :x: | 100% | :x: | 100% | | 14 | :x: | < 10% | :x: | 100% | | 17 | :x: | 10-20% | :heavy_check_mark: | N/A | | 18 | :heavy_check_mark: | N/A | :x: | 100% | | 19 | :x: | 10-20% | :heavy_check_mark: | N/A | | 20 | :heavy_check_mark: | N/A | :x: | 100% | | 21 | :x: | < 10% | :x: | 100% | | 24 | :x: | 100% | :heavy_check_mark: | N/A | | 27 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | | 28 | :heavy_check_mark: | N/A | :x: | 100% | | 29 | :heavy_check_mark: | N/A | :heavy_check_mark: | N/A | I used the following command line for all the tests: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` On the nucleo_f767zi board: - test cases 1, 3, 17 and 19 fail between 10 and 20% of the time. When it fails, the test end with "Connection timed out". The device is still usable after that failure. - test case 13 always fail with "Invalid argument". It doesn't have any other impact. - test cases 14 and 21 seem to trigger issue #27992 . Indeed, after enough time, these tests fail and the control endpoints become unresponsive. Unplugging the cable and plugging it again won't solve the issue: the device is unable to enumerate and the MCU has to be reset in order to have a functional USB device. - test case 24 always fails and is unable to complete. The testusb program stays stuck in the kernel in an uninterruptible sleep state. The device is still operational after plugging it again. On the nucleo_f070rb board: - test cases 2, 4, 6, 8, 14, 18, 20, 21 always fail and return "Connection timed out". The device can operate normally after that. - test case 13 always fail with "Invalid argument". - test cases 10 and 28 always fail and are stuck in the kernel indefinitely. The device needs to plugged again in order to On both boards, whenever a test fails, the log of the device shows: ``` <err> usb_dc_stm32: Unable to get write lock (-16) <err> usb_dc_stm32: Unable to get write lock (-16) ``` Changing the value of `USB_NUMOF_EP_WRITE_RETRIES` from 3 to 30 doesn't seem to have any impact on the tests failures. **To Reproduce** For the nucleo_f767zi board: - Compile and flash the testusb sample - On the Linux host, follow the instructions from `sample/subsys/usb/testusb/README.rst` - For each test case, run the following command: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` Since some test cases only fail around 10% of the time, this command may need to be executed multiple times before the failure occurs. For the nucleo_f070rb board: - Since that board doesn't have a USB connector, you'll need an external micro USB connector (the PA11 and PA12 pins can be used as USB D+/D-) - Once you are able to connect a Linux host to the board, you need to set the status of the "usb" node of the device tree to "okay": ``` diff --git a/boards/arm/nucleo_f070rb/nucleo_f070rb.dts b/boards/arm/nucleo_f070rb/nucleo_f070rb.dts index 9f5923d36d57..dcdf20912124 100644 --- a/boards/arm/nucleo_f070rb/nucleo_f070rb.dts +++ b/boards/arm/nucleo_f070rb/nucleo_f070rb.dts @@ -41,6 +41,10 @@ }; }; +&usb { + status = "okay"; +}; + &usart1 { current-speed = <115200>; status = "okay"; ``` - Compile and flash the testusb sample - On the Linux host, follow the instructions from `sample/subsys/usb/testusb/README.rst` - Use the command: ``` sudo ./testusb -D /dev/usb/xxx/yyy -v 512 -t <test_case> ``` **Expected behavior** All tests should pass. **Impact** There are multiple consequences to these failures: - The USB device might not pass USB certification - Some transfers randomly time out which compromise the communication with a host. - In some cases, USB is completely unusable without a MCU hardware reset, which is not acceptable or even possible in production. **Environment (please complete the following information):** - Host: Ubuntu 19.10 with Linux kernel 5.3.0-64-generic - Toolchain : Zephyr SDK - Commit SHA: c8cca16c204543b8f6dd7261c643b48ba8c95a94 - Boards: nucleo_f767zi and nucleo_f070rb (with external usb connector)
non_main
usb testusb fails and often hangs forever describe the bug on the nucleo and nucleo boards many test cases from the testusb linux kernel module fail failure rates and their consequences can take various forms most failures only result in a timeout for a specific tranfer but some of the failures can render the device unusable forcing the user to reset the mcu in order to use the usb device again the following table summarize the status of each test case for the two boards test case nucleo failure rate nucleo failure rate heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a x x heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a heavy check mark n a heavy check mark n a x x x x x heavy check mark n a heavy check mark n a x x heavy check mark n a heavy check mark n a x x x x heavy check mark n a heavy check mark n a heavy check mark n a heavy check mark n a x heavy check mark n a heavy check mark n a i used the following command line for all the tests sudo testusb d dev usb xxx yyy v t on the nucleo board test cases and fail between and of the time when it fails the test end with connection timed out the device is still usable after that failure test case always fail with invalid argument it doesn t have any other impact test cases and seem to trigger issue indeed after enough time these tests fail and the control endpoints become unresponsive unplugging the cable and plugging it again won t solve the issue the device is unable to enumerate and the mcu has to be reset in order to have a functional usb device test case always fails and is unable to complete the testusb program stays stuck in the kernel in an uninterruptible sleep state the device is still operational after plugging it again on the nucleo board test cases always fail and return connection timed out the device can operate normally after that test case always fail with invalid argument test cases and always fail and are stuck in the kernel indefinitely the device needs to plugged again in order to on both boards whenever a test fails the log of the device shows usb dc unable to get write lock usb dc unable to get write lock changing the value of usb numof ep write retries from to doesn t seem to have any impact on the tests failures to reproduce for the nucleo board compile and flash the testusb sample on the linux host follow the instructions from sample subsys usb testusb readme rst for each test case run the following command sudo testusb d dev usb xxx yyy v t since some test cases only fail around of the time this command may need to be executed multiple times before the failure occurs for the nucleo board since that board doesn t have a usb connector you ll need an external micro usb connector the and pins can be used as usb d d once you are able to connect a linux host to the board you need to set the status of the usb node of the device tree to okay diff git a boards arm nucleo nucleo dts b boards arm nucleo nucleo dts index a boards arm nucleo nucleo dts b boards arm nucleo nucleo dts usb status okay current speed status okay compile and flash the testusb sample on the linux host follow the instructions from sample subsys usb testusb readme rst use the command sudo testusb d dev usb xxx yyy v t expected behavior all tests should pass impact there are multiple consequences to these failures the usb device might not pass usb certification some transfers randomly time out which compromise the communication with a host in some cases usb is completely unusable without a mcu hardware reset which is not acceptable or even possible in production environment please complete the following information host ubuntu with linux kernel generic toolchain zephyr sdk commit sha boards nucleo and nucleo with external usb connector
0
325,258
27,860,680,396
IssuesEvent
2023-03-21 05:51:50
ladybirdweb/faveo-helpdesk
https://api.github.com/repos/ladybirdweb/faveo-helpdesk
closed
ErrorException in GET /public/step1
bugsnag Testing
## Error in Faveo Community **ErrorException** in **GET /public/step1** Breadcrumb not found with name "licence" (View: /home/hostlh/domains/faveo.linkhouse.pl/public_html/resources/views/themes/default1/client/layout/client.blade.php) (View: /home/hostlh/domains/faveo.linkhouse.pl/public_html/resources/views/themes/default1/client/layout/client.blade.php) [View on Bugsnag](https://app.bugsnag.com/ladybird-web-solution-pvt-ltd/faveo-community/errors/61a6a48171ba480008266088?event_id=61a6a481008958de3a040000&i=gh&m=ci) ## Stacktrace app/Exceptions/Handler.php:117 - App\Exceptions\Handler::render500 app/Exceptions/Handler.php:201 - App\Exceptions\Handler::common app/Exceptions/Handler.php:91 - App\Exceptions\Handler::render app/Http/Middleware/IsInstalled.php:20 - App\Http\Middleware\IsInstalled::handle app/Http/Middleware/LanguageMiddleware.php:33 - App\Http\Middleware\LanguageMiddleware::handle app/Http/Middleware/VerifyCsrfToken.php:33 - App\Http\Middleware\VerifyCsrfToken::handle [View full stacktrace](https://app.bugsnag.com/ladybird-web-solution-pvt-ltd/faveo-community/errors/61a6a48171ba480008266088?event_id=61a6a481008958de3a040000&i=gh&m=ci) *Created automatically via Bugsnag*
1.0
ErrorException in GET /public/step1 - ## Error in Faveo Community **ErrorException** in **GET /public/step1** Breadcrumb not found with name "licence" (View: /home/hostlh/domains/faveo.linkhouse.pl/public_html/resources/views/themes/default1/client/layout/client.blade.php) (View: /home/hostlh/domains/faveo.linkhouse.pl/public_html/resources/views/themes/default1/client/layout/client.blade.php) [View on Bugsnag](https://app.bugsnag.com/ladybird-web-solution-pvt-ltd/faveo-community/errors/61a6a48171ba480008266088?event_id=61a6a481008958de3a040000&i=gh&m=ci) ## Stacktrace app/Exceptions/Handler.php:117 - App\Exceptions\Handler::render500 app/Exceptions/Handler.php:201 - App\Exceptions\Handler::common app/Exceptions/Handler.php:91 - App\Exceptions\Handler::render app/Http/Middleware/IsInstalled.php:20 - App\Http\Middleware\IsInstalled::handle app/Http/Middleware/LanguageMiddleware.php:33 - App\Http\Middleware\LanguageMiddleware::handle app/Http/Middleware/VerifyCsrfToken.php:33 - App\Http\Middleware\VerifyCsrfToken::handle [View full stacktrace](https://app.bugsnag.com/ladybird-web-solution-pvt-ltd/faveo-community/errors/61a6a48171ba480008266088?event_id=61a6a481008958de3a040000&i=gh&m=ci) *Created automatically via Bugsnag*
non_main
errorexception in get public error in faveo community errorexception in get public breadcrumb not found with name licence view home hostlh domains faveo linkhouse pl public html resources views themes client layout client blade php view home hostlh domains faveo linkhouse pl public html resources views themes client layout client blade php stacktrace app exceptions handler php app exceptions handler app exceptions handler php app exceptions handler common app exceptions handler php app exceptions handler render app http middleware isinstalled php app http middleware isinstalled handle app http middleware languagemiddleware php app http middleware languagemiddleware handle app http middleware verifycsrftoken php app http middleware verifycsrftoken handle created automatically via bugsnag
0
47,470
13,236,205,426
IssuesEvent
2020-08-18 19:23:19
ExchangeUnion/xud
https://api.github.com/repos/ExchangeUnion/xud
closed
SwapRecovery stops monitoring swap after xud restart
P1 duplicate security swaps
This is from makers perspective. What happened: taker filled our `sell eth/btc` order, taker's eth side needed to be collateralized which caused the connext eth payment to timeout (60s) as well as the xud swap (90s) and the eth payment was never sent. Due to lack of https://github.com/ExchangeUnion/xud/issues/1708, `xud` currently doesn't know about the status of the connext htlc, thus assumes it's up and throws the swap into swap recovery. So far correct. But then when I went ahead and restarted `xud`, swap recovery was not monitoring the swap anymore. Checking the xud db, the swap state changed to `1` (errored). We need to find out what changed the swap state to `1` and prevent that from happening so that swap recovery continues monitoring the swap. ![image](https://user-images.githubusercontent.com/19181985/90397399-129ab100-e098-11ea-9a8c-0ec689e50ce3.png) xud log: https://paste.ubuntu.com/p/Y8ZzWhwZJk/
True
SwapRecovery stops monitoring swap after xud restart - This is from makers perspective. What happened: taker filled our `sell eth/btc` order, taker's eth side needed to be collateralized which caused the connext eth payment to timeout (60s) as well as the xud swap (90s) and the eth payment was never sent. Due to lack of https://github.com/ExchangeUnion/xud/issues/1708, `xud` currently doesn't know about the status of the connext htlc, thus assumes it's up and throws the swap into swap recovery. So far correct. But then when I went ahead and restarted `xud`, swap recovery was not monitoring the swap anymore. Checking the xud db, the swap state changed to `1` (errored). We need to find out what changed the swap state to `1` and prevent that from happening so that swap recovery continues monitoring the swap. ![image](https://user-images.githubusercontent.com/19181985/90397399-129ab100-e098-11ea-9a8c-0ec689e50ce3.png) xud log: https://paste.ubuntu.com/p/Y8ZzWhwZJk/
non_main
swaprecovery stops monitoring swap after xud restart this is from makers perspective what happened taker filled our sell eth btc order taker s eth side needed to be collateralized which caused the connext eth payment to timeout as well as the xud swap and the eth payment was never sent due to lack of xud currently doesn t know about the status of the connext htlc thus assumes it s up and throws the swap into swap recovery so far correct but then when i went ahead and restarted xud swap recovery was not monitoring the swap anymore checking the xud db the swap state changed to errored we need to find out what changed the swap state to and prevent that from happening so that swap recovery continues monitoring the swap xud log
0
2,981
10,732,725,728
IssuesEvent
2019-10-28 22:42:46
wtfd-tech/wtfd
https://api.github.com/repos/wtfd-tech/wtfd
opened
Gereral Fix Branch
Idea Maintainability
Use e.g. branch [general-fixes](https://github.com/wtfd-tech/wtfd/tree/general-fixes) for minor changes to keep them apart from other branches, where they don't belong. Then merge this branch with regularity.
True
Gereral Fix Branch - Use e.g. branch [general-fixes](https://github.com/wtfd-tech/wtfd/tree/general-fixes) for minor changes to keep them apart from other branches, where they don't belong. Then merge this branch with regularity.
main
gereral fix branch use e g branch for minor changes to keep them apart from other branches where they don t belong then merge this branch with regularity
1
285,379
24,662,497,186
IssuesEvent
2022-10-18 07:51:55
pingcap/tiflow
https://api.github.com/repos/pingcap/tiflow
closed
Add savepoint test
component/test type/enhancement
### Which jobs are flaking? N/A ### Which test(s) are flaking? N/A ### Jenkins logs or GitHub Actions link ```log (paste logs here) ``` ### Anything else we need to know - Does this test exist for other branches as well? - Has there been a high frequency of failure lately?
1.0
Add savepoint test - ### Which jobs are flaking? N/A ### Which test(s) are flaking? N/A ### Jenkins logs or GitHub Actions link ```log (paste logs here) ``` ### Anything else we need to know - Does this test exist for other branches as well? - Has there been a high frequency of failure lately?
non_main
add savepoint test which jobs are flaking n a which test s are flaking n a jenkins logs or github actions link log paste logs here anything else we need to know does this test exist for other branches as well has there been a high frequency of failure lately
0
3,953
17,930,647,890
IssuesEvent
2021-09-10 08:47:43
restqa/restqapi
https://api.github.com/repos/restqa/restqapi
closed
Translate the test scenario into a curl command
enhancement good first issue pair with maintainer
Hello 👋, ### 👀 Background During the test scenario execution, if the user is enabling the html report he will be able to access to the detail description of the api ran behind the scene Example: ![Screen+Recording+2564-06-27+at+02 10 51+PM](https://user-images.githubusercontent.com/4768226/123535943-a4ac2500-d751-11eb-82c2-791074596cb3.gif) This is very helpful for the user to know more about the actual behavior of the api during the test execution, however in order to debug or to reproduce the scenario behavior it would be great to be able to run rerun the scenario manually through another tool like postman or the terminal. ### ✌️ What is the actual behavior? When the user is run the test, the last step will add an [attachement](https://github.com/cucumber/cucumber-js/blob/main/docs/support_files/attachments.md) as an after hook to print the recorded apis. File : https://github.com/restqa/restqapi/blob/master/src/restqapi/hooks.js#L52 ### 🕵️‍♀️ How to reproduce the current behavior? 1. Clone the current project: `git clone git@github.com:restqa/restqapi.git` 2. Install the dependencies: `npm i` 3. Run the command: `npm run example -- -f json:result.json` 4. Check the attachment step in the result.json The content of the attachment step should be the detail of each api calls. ### 🤞 What is the expected behavior? The actual behavior is good enough to know more about the api call that are executed under the scene, but in term of developer experience we should have an easy way to replicate the api call through the tool of our choice. ### 😎 Proposed solution. The current attachment is composed of the api detail detailed composed of : ``` apis: [{ request: { "hostname": "jsonplaceholder.typicode.com", "protocol": "https:", "pathname": "/todos/300000", }, response: {} }] ``` We want to propose to add the curl command of each api in the attachement: ``` apis: [{ curl: "curl --url https://jsonplaceholder.typicode.com/todos/3000000" request: { "hostname": "jsonplaceholder.typicode.com", "protocol": "https:", "pathname": "/todos/300000", }, response: {} }] ``` Then the user will be able to copy the curl command and use it in the terminal, postman, insomnia, etc... This will guarantee an easy way for the user to reproduce the api call. Cheers.
True
Translate the test scenario into a curl command - Hello 👋, ### 👀 Background During the test scenario execution, if the user is enabling the html report he will be able to access to the detail description of the api ran behind the scene Example: ![Screen+Recording+2564-06-27+at+02 10 51+PM](https://user-images.githubusercontent.com/4768226/123535943-a4ac2500-d751-11eb-82c2-791074596cb3.gif) This is very helpful for the user to know more about the actual behavior of the api during the test execution, however in order to debug or to reproduce the scenario behavior it would be great to be able to run rerun the scenario manually through another tool like postman or the terminal. ### ✌️ What is the actual behavior? When the user is run the test, the last step will add an [attachement](https://github.com/cucumber/cucumber-js/blob/main/docs/support_files/attachments.md) as an after hook to print the recorded apis. File : https://github.com/restqa/restqapi/blob/master/src/restqapi/hooks.js#L52 ### 🕵️‍♀️ How to reproduce the current behavior? 1. Clone the current project: `git clone git@github.com:restqa/restqapi.git` 2. Install the dependencies: `npm i` 3. Run the command: `npm run example -- -f json:result.json` 4. Check the attachment step in the result.json The content of the attachment step should be the detail of each api calls. ### 🤞 What is the expected behavior? The actual behavior is good enough to know more about the api call that are executed under the scene, but in term of developer experience we should have an easy way to replicate the api call through the tool of our choice. ### 😎 Proposed solution. The current attachment is composed of the api detail detailed composed of : ``` apis: [{ request: { "hostname": "jsonplaceholder.typicode.com", "protocol": "https:", "pathname": "/todos/300000", }, response: {} }] ``` We want to propose to add the curl command of each api in the attachement: ``` apis: [{ curl: "curl --url https://jsonplaceholder.typicode.com/todos/3000000" request: { "hostname": "jsonplaceholder.typicode.com", "protocol": "https:", "pathname": "/todos/300000", }, response: {} }] ``` Then the user will be able to copy the curl command and use it in the terminal, postman, insomnia, etc... This will guarantee an easy way for the user to reproduce the api call. Cheers.
main
translate the test scenario into a curl command hello 👋 👀 background during the test scenario execution if the user is enabling the html report he will be able to access to the detail description of the api ran behind the scene example this is very helpful for the user to know more about the actual behavior of the api during the test execution however in order to debug or to reproduce the scenario behavior it would be great to be able to run rerun the scenario manually through another tool like postman or the terminal ✌️ what is the actual behavior when the user is run the test the last step will add an as an after hook to print the recorded apis file 🕵️‍♀️ how to reproduce the current behavior clone the current project git clone git github com restqa restqapi git install the dependencies npm i run the command npm run example f json result json check the attachment step in the result json the content of the attachment step should be the detail of each api calls 🤞 what is the expected behavior the actual behavior is good enough to know more about the api call that are executed under the scene but in term of developer experience we should have an easy way to replicate the api call through the tool of our choice 😎 proposed solution the current attachment is composed of the api detail detailed composed of apis request hostname jsonplaceholder typicode com protocol https pathname todos response we want to propose to add the curl command of each api in the attachement apis curl curl url request hostname jsonplaceholder typicode com protocol https pathname todos response then the user will be able to copy the curl command and use it in the terminal postman insomnia etc this will guarantee an easy way for the user to reproduce the api call cheers
1
4,101
19,359,197,281
IssuesEvent
2021-12-16 01:43:31
aws/aws-sam-build-images
https://api.github.com/repos/aws/aws-sam-build-images
closed
Add .NET runtimes
runtime dotnet maintainer/need-response
### Description Currently there are not provided docker images for the .NET runtime, and the [sam build command does not support the `--use-container` flag](https://github.com/aws/aws-sam-cli/pull/1096). This makes typical build and deploy workflows with SAM not working consistently for .NET projects. It would be great if a Docker image could be provided and the [sam cli](https://github.com/aws/aws-sam-cli) updated to work with the container. ### Benefits Ensures consistent behavior for all [AWS Lambda supported runtimes](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html) by the AWS SAM cli and build/deploy workflow. ### Request Add .NET Dockerfiles supporting AWS runtimes - Dockerfile-dotnetcore31 - Dockerfile-dotnetcore21
True
Add .NET runtimes - ### Description Currently there are not provided docker images for the .NET runtime, and the [sam build command does not support the `--use-container` flag](https://github.com/aws/aws-sam-cli/pull/1096). This makes typical build and deploy workflows with SAM not working consistently for .NET projects. It would be great if a Docker image could be provided and the [sam cli](https://github.com/aws/aws-sam-cli) updated to work with the container. ### Benefits Ensures consistent behavior for all [AWS Lambda supported runtimes](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html) by the AWS SAM cli and build/deploy workflow. ### Request Add .NET Dockerfiles supporting AWS runtimes - Dockerfile-dotnetcore31 - Dockerfile-dotnetcore21
main
add net runtimes description currently there are not provided docker images for the net runtime and the this makes typical build and deploy workflows with sam not working consistently for net projects it would be great if a docker image could be provided and the updated to work with the container benefits ensures consistent behavior for all by the aws sam cli and build deploy workflow request add net dockerfiles supporting aws runtimes dockerfile dockerfile
1
127,859
18,024,438,523
IssuesEvent
2021-09-17 01:16:52
rgordon95/GordonEnterprisesSite
https://api.github.com/repos/rgordon95/GordonEnterprisesSite
opened
CVE-2021-29059 (High) detected in is-svg-2.1.0.tgz
security vulnerability
## CVE-2021-29059 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-2.1.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz</a></p> <p>Path to dependency file: GordonEnterprisesSite/package.json</p> <p>Path to vulnerable library: GordonEnterprisesSite/node_modules/is-svg/package.json</p> <p> Dependency Hierarchy: - gulp-imagemin-3.4.0.tgz (Root Library) - imagemin-svgo-5.2.4.tgz - :x: **is-svg-2.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was discovered in IS-SVG version 2.1.0 to 4.2.2 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sindresorhus/is-svg/releases/tag/v4.3.0">https://github.com/sindresorhus/is-svg/releases/tag/v4.3.0</a></p> <p>Release Date: 2021-06-21</p> <p>Fix Resolution: is-svg - 4.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29059 (High) detected in is-svg-2.1.0.tgz - ## CVE-2021-29059 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-2.1.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz</a></p> <p>Path to dependency file: GordonEnterprisesSite/package.json</p> <p>Path to vulnerable library: GordonEnterprisesSite/node_modules/is-svg/package.json</p> <p> Dependency Hierarchy: - gulp-imagemin-3.4.0.tgz (Root Library) - imagemin-svgo-5.2.4.tgz - :x: **is-svg-2.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was discovered in IS-SVG version 2.1.0 to 4.2.2 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sindresorhus/is-svg/releases/tag/v4.3.0">https://github.com/sindresorhus/is-svg/releases/tag/v4.3.0</a></p> <p>Release Date: 2021-06-21</p> <p>Fix Resolution: is-svg - 4.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in is svg tgz cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file gordonenterprisessite package json path to vulnerable library gordonenterprisessite node modules is svg package json dependency hierarchy gulp imagemin tgz root library imagemin svgo tgz x is svg tgz vulnerable library found in base branch master vulnerability details a vulnerability was discovered in is svg version to and below where a regular expression denial of service redos occurs if the application is provided and checks a crafted invalid svg string publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is svg step up your open source security game with whitesource
0
375,158
11,100,830,590
IssuesEvent
2019-12-16 20:04:38
dhenry-KCI/FredCo-Post-Go-Live-
https://api.github.com/repos/dhenry-KCI/FredCo-Post-Go-Live-
opened
Pdox- report errors
High Priority
Planning staff is getting errors on a number of reports under “project reports”. ![image](https://user-images.githubusercontent.com/50637067/70938969-4a3af000-2015-11ea-95cd-c1d365216c7d.png)
1.0
Pdox- report errors - Planning staff is getting errors on a number of reports under “project reports”. ![image](https://user-images.githubusercontent.com/50637067/70938969-4a3af000-2015-11ea-95cd-c1d365216c7d.png)
non_main
pdox report errors planning staff is getting errors on a number of reports under “project reports”
0
4,385
3,367,263,020
IssuesEvent
2015-11-22 01:33:08
openhab/openhab
https://api.github.com/repos/openhab/openhab
closed
can't setup IDE
build-or-ide question
https://github.com/openhab/openhab/wiki/IDE-Setup I read this tutorial , i install all required plugins on top of an existing Eclipse ,use this update site:http://yoxos.eclipsesource.com/userdata/profile/c5f3985b62c488f0df0dfbc369f9e057 there is an error while i isntalling it:"The installation cannot be completed as requested."
1.0
can't setup IDE - https://github.com/openhab/openhab/wiki/IDE-Setup I read this tutorial , i install all required plugins on top of an existing Eclipse ,use this update site:http://yoxos.eclipsesource.com/userdata/profile/c5f3985b62c488f0df0dfbc369f9e057 there is an error while i isntalling it:"The installation cannot be completed as requested."
non_main
can t setup ide i read this tutorial i install all required plugins on top of an existing eclipse use this update site there is an error while i isntalling it the installation cannot be completed as requested
0
67,174
9,011,957,756
IssuesEvent
2019-02-05 15:50:09
zeppelinos/zos
https://api.github.com/repos/zeppelinos/zos
opened
Links in documentation pointing to source code are broken
kind:documentation
Many links that intend to point to the actual source code of contracts in the Github repo are broken. As an example, in https://docs.zeppelinos.org/docs/Initializable.html, the link in _Source: Initializable.sol_ redirects to: `git+https://github.com/zeppelinos/zos/blob/v2.0.1/contracts/Initializable.sol`. Removing the `git+` does not fix this issue, since the path to the contract is wrong. In this case, it should be https://github.com/zeppelinos/zos/blob/v2.0.1/packages/lib/contracts/Initializable.sol
1.0
Links in documentation pointing to source code are broken - Many links that intend to point to the actual source code of contracts in the Github repo are broken. As an example, in https://docs.zeppelinos.org/docs/Initializable.html, the link in _Source: Initializable.sol_ redirects to: `git+https://github.com/zeppelinos/zos/blob/v2.0.1/contracts/Initializable.sol`. Removing the `git+` does not fix this issue, since the path to the contract is wrong. In this case, it should be https://github.com/zeppelinos/zos/blob/v2.0.1/packages/lib/contracts/Initializable.sol
non_main
links in documentation pointing to source code are broken many links that intend to point to the actual source code of contracts in the github repo are broken as an example in the link in source initializable sol redirects to git removing the git does not fix this issue since the path to the contract is wrong in this case it should be
0
317,024
23,660,664,390
IssuesEvent
2022-08-26 15:16:03
infrahq/infra
https://api.github.com/repos/infrahq/infra
closed
Docs: the docs for the google workspace id integration are confusing
area/documentation
## Page URL https://infrahq.com/docs/identity-providers/google ## Page Location the whole thing ## Describe the Issue the docs describe the command to run in the cli, and the process to generate everything in the google ui. but the infra ui is different and a bit confusing. ## Screenshots ![CleanShot 2022-07-26 at 17 10 01@2x](https://user-images.githubusercontent.com/633681/181133345-a0c707a2-38e6-4d0f-bcaf-9009028f9869.png) ![CleanShot 2022-07-26 at 17 12 16](https://user-images.githubusercontent.com/633681/181133464-dcf175c5-580f-4f28-a066-02f6d6ac4231.png) <!-- If applicable, add screenshots to help explain what you are seeing. --> ## Suggested Solution <!-- A clear and concise description of what you would prefer. --> ## Your Environment <!-- What OS and browser are you using --> ## Additional Context <!-- Add any other context about the problem here. -->
1.0
Docs: the docs for the google workspace id integration are confusing - ## Page URL https://infrahq.com/docs/identity-providers/google ## Page Location the whole thing ## Describe the Issue the docs describe the command to run in the cli, and the process to generate everything in the google ui. but the infra ui is different and a bit confusing. ## Screenshots ![CleanShot 2022-07-26 at 17 10 01@2x](https://user-images.githubusercontent.com/633681/181133345-a0c707a2-38e6-4d0f-bcaf-9009028f9869.png) ![CleanShot 2022-07-26 at 17 12 16](https://user-images.githubusercontent.com/633681/181133464-dcf175c5-580f-4f28-a066-02f6d6ac4231.png) <!-- If applicable, add screenshots to help explain what you are seeing. --> ## Suggested Solution <!-- A clear and concise description of what you would prefer. --> ## Your Environment <!-- What OS and browser are you using --> ## Additional Context <!-- Add any other context about the problem here. -->
non_main
docs the docs for the google workspace id integration are confusing page url page location the whole thing describe the issue the docs describe the command to run in the cli and the process to generate everything in the google ui but the infra ui is different and a bit confusing screenshots suggested solution your environment additional context
0
417,019
12,154,787,556
IssuesEvent
2020-04-25 10:03:14
tal3898/Hummus
https://api.github.com/repos/tal3898/Hummus
closed
The meta data fields, are not in json until it changed
Done EntityEditor bug priority - high
כל השדות של הבקשה: ישות, מערכת... לא נמצאות בגיסון (כשלוחצים על הכפור שמציג את הגיסון) , עד שמשנים את ערכם. רק כשמשנים את ערכם , אז הם נמצאים
1.0
The meta data fields, are not in json until it changed - כל השדות של הבקשה: ישות, מערכת... לא נמצאות בגיסון (כשלוחצים על הכפור שמציג את הגיסון) , עד שמשנים את ערכם. רק כשמשנים את ערכם , אז הם נמצאים
non_main
the meta data fields are not in json until it changed כל השדות של הבקשה ישות מערכת לא נמצאות בגיסון כשלוחצים על הכפור שמציג את הגיסון עד שמשנים את ערכם רק כשמשנים את ערכם אז הם נמצאים
0
120,966
10,143,890,288
IssuesEvent
2019-08-04 16:00:26
HippieStation/HippieStation
https://api.github.com/repos/HippieStation/HippieStation
closed
No damage done while you have your liver removed
Needs Reproducing/Testing
Pretty simple. You can have no liver in your body and just don't give a damn about chemicals or any toxin damage affecting you.
1.0
No damage done while you have your liver removed - Pretty simple. You can have no liver in your body and just don't give a damn about chemicals or any toxin damage affecting you.
non_main
no damage done while you have your liver removed pretty simple you can have no liver in your body and just don t give a damn about chemicals or any toxin damage affecting you
0
62,253
8,583,632,575
IssuesEvent
2018-11-13 20:18:11
lammps/lammps
https://api.github.com/repos/lammps/lammps
closed
Typo in doc/src/Build_basics.txt -> cmake ignores the variables
bug documentation
The variables documented under "Choice of compiler and compile/link options" should be "FLAGS" not "FlAGS" - the later is ignored by cmake: https://github.com/lammps/lammps/blob/1651a21f921df63c774cd316e24bd8f03bfb6edf/doc/src/Build_basics.txt#L140-L142
1.0
Typo in doc/src/Build_basics.txt -> cmake ignores the variables - The variables documented under "Choice of compiler and compile/link options" should be "FLAGS" not "FlAGS" - the later is ignored by cmake: https://github.com/lammps/lammps/blob/1651a21f921df63c774cd316e24bd8f03bfb6edf/doc/src/Build_basics.txt#L140-L142
non_main
typo in doc src build basics txt cmake ignores the variables the variables documented under choice of compiler and compile link options should be flags not flags the later is ignored by cmake
0
5,603
28,052,891,977
IssuesEvent
2023-03-29 07:24:30
warengonzaga/thirdweb-support-discord-bot
https://api.github.com/repos/warengonzaga/thirdweb-support-discord-bot
closed
convert to forum-based bot
in progress maintainers only tweak
The code for the bot is completely similar to the thread-based one, just needs some tricky updates.
True
convert to forum-based bot - The code for the bot is completely similar to the thread-based one, just needs some tricky updates.
main
convert to forum based bot the code for the bot is completely similar to the thread based one just needs some tricky updates
1
3,905
17,376,851,944
IssuesEvent
2021-07-30 23:28:22
chorman0773/Clever-ISA
https://api.github.com/repos/chorman0773/Clever-ISA
closed
Bulk Register Storage: pushgpr, popgpr, pushar, and popar don't mention when the stack pointer is updated
I-unclear S-blocked-on-maintainer X-main
Bulk Register Storage: pushgpr, popgpr, pushar, and popar don't mention when the stack pointer is updated. pushgpr and pushar should update the stack pointer after storing it, and popgpr and popar should not modify it at all after restoring it.
True
Bulk Register Storage: pushgpr, popgpr, pushar, and popar don't mention when the stack pointer is updated - Bulk Register Storage: pushgpr, popgpr, pushar, and popar don't mention when the stack pointer is updated. pushgpr and pushar should update the stack pointer after storing it, and popgpr and popar should not modify it at all after restoring it.
main
bulk register storage pushgpr popgpr pushar and popar don t mention when the stack pointer is updated bulk register storage pushgpr popgpr pushar and popar don t mention when the stack pointer is updated pushgpr and pushar should update the stack pointer after storing it and popgpr and popar should not modify it at all after restoring it
1
85,760
10,458,946,066
IssuesEvent
2019-09-20 09:44:03
Nurminen-dev/wacs
https://api.github.com/repos/Nurminen-dev/wacs
closed
Bokun detailed product info API
documentation enhancement
### Bokun detailed product info API [https://docs.google.com/document/d/1oavs_UCcUVdcOlkC9obtas21OyntEVMrTEsvjktZLJs/edit#](https://docs.google.com/document/d/1oavs_UCcUVdcOlkC9obtas21OyntEVMrTEsvjktZLJs/edit#) -> Get full details via ID TODO: - **Design and document API for Comgate**, fill 3.1.2 Products: Product information in WACS API .docx (can copy base from 3.1.1. Products: Product list) - endpoint could be, for example: POST /v1/activities/products/{TargetSystem}{DestinationID}/{ProductID} - Request body same as in 3.1.1. Products: Product list - Response body: Return product details .. in similiar format as **2.1.2 Product mapping: Product information** (same if possible, but adapt / leave out parameters if not provided by Booking API) - **WACS internal additions:** - add endpoint to api-v1/activities.js constructor ("Activities/ProductInformation") - route new URL params (DestinationID, ProductID) from activities.js/_activityRequestHandler() to ExternalModulesHandler.js/externalRequest() (in data parameter, combined with auth data from POST body) - add "Activities/ProductInformation" to ExternalModulesHandler.js/externalRequest(), moduleMethodName could be "activitiesProductInformation", for example - **ExternalModules/bokun/bokun.js:** - add activitiesProductInformation method and make bokun implementation, return data should match what's documented in step1 Fork **2-bokun-productinformation** branch and make changes there! :)
1.0
Bokun detailed product info API - ### Bokun detailed product info API [https://docs.google.com/document/d/1oavs_UCcUVdcOlkC9obtas21OyntEVMrTEsvjktZLJs/edit#](https://docs.google.com/document/d/1oavs_UCcUVdcOlkC9obtas21OyntEVMrTEsvjktZLJs/edit#) -> Get full details via ID TODO: - **Design and document API for Comgate**, fill 3.1.2 Products: Product information in WACS API .docx (can copy base from 3.1.1. Products: Product list) - endpoint could be, for example: POST /v1/activities/products/{TargetSystem}{DestinationID}/{ProductID} - Request body same as in 3.1.1. Products: Product list - Response body: Return product details .. in similiar format as **2.1.2 Product mapping: Product information** (same if possible, but adapt / leave out parameters if not provided by Booking API) - **WACS internal additions:** - add endpoint to api-v1/activities.js constructor ("Activities/ProductInformation") - route new URL params (DestinationID, ProductID) from activities.js/_activityRequestHandler() to ExternalModulesHandler.js/externalRequest() (in data parameter, combined with auth data from POST body) - add "Activities/ProductInformation" to ExternalModulesHandler.js/externalRequest(), moduleMethodName could be "activitiesProductInformation", for example - **ExternalModules/bokun/bokun.js:** - add activitiesProductInformation method and make bokun implementation, return data should match what's documented in step1 Fork **2-bokun-productinformation** branch and make changes there! :)
non_main
bokun detailed product info api bokun detailed product info api get full details via id todo design and document api for comgate fill products product information in wacs api docx can copy base from products product list endpoint could be for example post activities products targetsystem destinationid productid request body same as in products product list response body return product details in similiar format as product mapping product information same if possible but adapt leave out parameters if not provided by booking api wacs internal additions add endpoint to api activities js constructor activities productinformation route new url params destinationid productid from activities js activityrequesthandler to externalmoduleshandler js externalrequest in data parameter combined with auth data from post body add activities productinformation to externalmoduleshandler js externalrequest modulemethodname could be activitiesproductinformation for example externalmodules bokun bokun js add activitiesproductinformation method and make bokun implementation return data should match what s documented in fork bokun productinformation branch and make changes there
0
4,975
25,546,448,012
IssuesEvent
2022-11-29 19:14:09
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Customize the filtering functions for FK columns
type: enhancement work: frontend status: ready restricted: new maintainers
## Current behavior - Filtering on an FK column presents the user with different filtering functions that don't quite make sense given our UX with the record selector. For FK columns, the value within the filter condition will be acquired through the (modal) Record Selector and afterwards displayed within the input using the Record Summary for the acquired record. The numerical primary key value used for the filtering condition won't be displayed to the user unless it happens to be contained within the Record Summary. Thus, for example, selecting all Publications with and author _greater than_ "Octavia Butler" doesn't quite make sense -- in the quantitative sense, at least 😜. ## Desired behavior - Filtering functions for FK columns look more like the functions for boolean columns. We should only have "is empty" and "is equal to".
True
Customize the filtering functions for FK columns - ## Current behavior - Filtering on an FK column presents the user with different filtering functions that don't quite make sense given our UX with the record selector. For FK columns, the value within the filter condition will be acquired through the (modal) Record Selector and afterwards displayed within the input using the Record Summary for the acquired record. The numerical primary key value used for the filtering condition won't be displayed to the user unless it happens to be contained within the Record Summary. Thus, for example, selecting all Publications with and author _greater than_ "Octavia Butler" doesn't quite make sense -- in the quantitative sense, at least 😜. ## Desired behavior - Filtering functions for FK columns look more like the functions for boolean columns. We should only have "is empty" and "is equal to".
main
customize the filtering functions for fk columns current behavior filtering on an fk column presents the user with different filtering functions that don t quite make sense given our ux with the record selector for fk columns the value within the filter condition will be acquired through the modal record selector and afterwards displayed within the input using the record summary for the acquired record the numerical primary key value used for the filtering condition won t be displayed to the user unless it happens to be contained within the record summary thus for example selecting all publications with and author greater than octavia butler doesn t quite make sense in the quantitative sense at least 😜 desired behavior filtering functions for fk columns look more like the functions for boolean columns we should only have is empty and is equal to
1
2,027
6,777,490,892
IssuesEvent
2017-10-27 22:35:42
christoff-buerger/racr
https://api.github.com/repos/christoff-buerger/racr
closed
portable record-based unique identities
core logic example high maintainability security
Discussing issue #85 with the _Chez Scheme_ developers (cisco/ChezScheme#226), it turns out uniqueness of immutable records is an editorial mistake of _R6RS_; regarding consistency with other immutable entities and equality tests on such, it should not be part of the standard and be removed to open opportunity to optimise constructions of immutable records with equal constant arguments or without fields to a single instance derived at compile time. The _Chez Scheme_ compiler is implementing such a folding of immutable records. This is relevant for _RACR_, as it uses empty records for unique, internal keys that can be checked without risking that such checks are flawed by users inference (as users can never construct an equal key). If a record contains at least a single mutable field however, its constructors cannot be folded to a single record instance. Thus, all unique-key instantiations have to be changed to be based on a mutable record. For optimisation, `nongenerative` clauses should also be added to let _Scheme_ compilers inline record-type definitions.
True
portable record-based unique identities - Discussing issue #85 with the _Chez Scheme_ developers (cisco/ChezScheme#226), it turns out uniqueness of immutable records is an editorial mistake of _R6RS_; regarding consistency with other immutable entities and equality tests on such, it should not be part of the standard and be removed to open opportunity to optimise constructions of immutable records with equal constant arguments or without fields to a single instance derived at compile time. The _Chez Scheme_ compiler is implementing such a folding of immutable records. This is relevant for _RACR_, as it uses empty records for unique, internal keys that can be checked without risking that such checks are flawed by users inference (as users can never construct an equal key). If a record contains at least a single mutable field however, its constructors cannot be folded to a single record instance. Thus, all unique-key instantiations have to be changed to be based on a mutable record. For optimisation, `nongenerative` clauses should also be added to let _Scheme_ compilers inline record-type definitions.
main
portable record based unique identities discussing issue with the chez scheme developers cisco chezscheme it turns out uniqueness of immutable records is an editorial mistake of regarding consistency with other immutable entities and equality tests on such it should not be part of the standard and be removed to open opportunity to optimise constructions of immutable records with equal constant arguments or without fields to a single instance derived at compile time the chez scheme compiler is implementing such a folding of immutable records this is relevant for racr as it uses empty records for unique internal keys that can be checked without risking that such checks are flawed by users inference as users can never construct an equal key if a record contains at least a single mutable field however its constructors cannot be folded to a single record instance thus all unique key instantiations have to be changed to be based on a mutable record for optimisation nongenerative clauses should also be added to let scheme compilers inline record type definitions
1
2,822
10,129,450,679
IssuesEvent
2019-08-01 14:49:36
diofant/diofant
https://api.github.com/repos/diofant/diofant
closed
Support len(polys)>2 in dup_isolate_real_roots_list() or fix API
maintainability polys
After diofant/diofant#835 (080dab0) - this method does support only pair of polynomials. XFALed test must be restored or API adapted to work only with two polynomials.
True
Support len(polys)>2 in dup_isolate_real_roots_list() or fix API - After diofant/diofant#835 (080dab0) - this method does support only pair of polynomials. XFALed test must be restored or API adapted to work only with two polynomials.
main
support len polys in dup isolate real roots list or fix api after diofant diofant this method does support only pair of polynomials xfaled test must be restored or api adapted to work only with two polynomials
1
555,457
16,455,030,598
IssuesEvent
2021-05-21 11:19:46
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.2 staging-1852] Face remover troubles with glass and lumber
Category: Tech Priority: Medium Squad: Otter Status: Fixed Type: Bug
Use Lumber Wall and GlassWindow: ![image](https://user-images.githubusercontent.com/45708377/99939996-61817200-2d7c-11eb-8d17-fb1953a1460f.png) ![image](https://user-images.githubusercontent.com/45708377/99940240-d0f76180-2d7c-11eb-81c5-a3ecae5accb7.png)
1.0
[0.9.2 staging-1852] Face remover troubles with glass and lumber - Use Lumber Wall and GlassWindow: ![image](https://user-images.githubusercontent.com/45708377/99939996-61817200-2d7c-11eb-8d17-fb1953a1460f.png) ![image](https://user-images.githubusercontent.com/45708377/99940240-d0f76180-2d7c-11eb-81c5-a3ecae5accb7.png)
non_main
face remover troubles with glass and lumber use lumber wall and glasswindow
0
3,332
12,944,367,196
IssuesEvent
2020-07-18 10:50:50
precice/precice
https://api.github.com/repos/precice/precice
closed
Remove vertexMap in Triangle
maintainability
A [`precice::mesh::Triangle`](http://xgm.de/precice/docs/develop/classprecice_1_1mesh_1_1Triangle.html) holds pointers to its Edges and a map `vertexMap` which maps edges `0,1,2` to their respective unique vertex `0,1`. This is solely used to access the vertices via the edges. We could directly save the Vertices instead. This reduced unnecessary logic in Triangle constructor, reduces the size of this primitive and reduces the logic to access a vector. (Currently `vector(i) = edge(i).vertex(vertexMap[i])`)
True
Remove vertexMap in Triangle - A [`precice::mesh::Triangle`](http://xgm.de/precice/docs/develop/classprecice_1_1mesh_1_1Triangle.html) holds pointers to its Edges and a map `vertexMap` which maps edges `0,1,2` to their respective unique vertex `0,1`. This is solely used to access the vertices via the edges. We could directly save the Vertices instead. This reduced unnecessary logic in Triangle constructor, reduces the size of this primitive and reduces the logic to access a vector. (Currently `vector(i) = edge(i).vertex(vertexMap[i])`)
main
remove vertexmap in triangle a holds pointers to its edges and a map vertexmap which maps edges to their respective unique vertex this is solely used to access the vertices via the edges we could directly save the vertices instead this reduced unnecessary logic in triangle constructor reduces the size of this primitive and reduces the logic to access a vector currently vector i edge i vertex vertexmap
1
241,492
20,144,431,879
IssuesEvent
2022-02-09 05:08:41
chipsalliance/Surelog
https://api.github.com/repos/chipsalliance/Surelog
closed
Surelog doesn't import wires declared in packages
sv-tests
TEST: ``` package my_package1; parameter p1 = 1; localparam p2 = 2; typedef logic [1:0] word; word v; endpackage module test(); import my_package1::*; initial begin v = p1 + p2; end endmodule ``` Surelog fails with message: ``` dut.sv:17:3: Illegal lhs of type wire "v". ``` but it should pass.
1.0
Surelog doesn't import wires declared in packages - TEST: ``` package my_package1; parameter p1 = 1; localparam p2 = 2; typedef logic [1:0] word; word v; endpackage module test(); import my_package1::*; initial begin v = p1 + p2; end endmodule ``` Surelog fails with message: ``` dut.sv:17:3: Illegal lhs of type wire "v". ``` but it should pass.
non_main
surelog doesn t import wires declared in packages test package my parameter localparam typedef logic word word v endpackage module test import my initial begin v end endmodule surelog fails with message dut sv illegal lhs of type wire v but it should pass
0
621,237
19,581,037,456
IssuesEvent
2022-01-04 21:21:36
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
closed
Build a Standard COVID Schema
onboarding-ops High Priority experience Enhancement
## Problem Statement As the Experience Team, we need a standard schema for COVID reporting in ReportStream so that we can test the hypothesis that reporting entities will be able to comply with a standard schema. ## Notes Jim currently taking a first pass at building and will then be handing off to Joel & Rick ## Criteria - [x] Setup a default sender_id named "manualupload" to be used as a test Sender for our validation. - [x] The schema must be developed based on the requirements outlined in the Standard CSV Requirements document named "CSV-StandardFileFormatAndNotes-v2.docx" in the CSV Pilot Sharepoint folder.
1.0
Build a Standard COVID Schema - ## Problem Statement As the Experience Team, we need a standard schema for COVID reporting in ReportStream so that we can test the hypothesis that reporting entities will be able to comply with a standard schema. ## Notes Jim currently taking a first pass at building and will then be handing off to Joel & Rick ## Criteria - [x] Setup a default sender_id named "manualupload" to be used as a test Sender for our validation. - [x] The schema must be developed based on the requirements outlined in the Standard CSV Requirements document named "CSV-StandardFileFormatAndNotes-v2.docx" in the CSV Pilot Sharepoint folder.
non_main
build a standard covid schema problem statement as the experience team we need a standard schema for covid reporting in reportstream so that we can test the hypothesis that reporting entities will be able to comply with a standard schema notes jim currently taking a first pass at building and will then be handing off to joel rick criteria setup a default sender id named manualupload to be used as a test sender for our validation the schema must be developed based on the requirements outlined in the standard csv requirements document named csv standardfileformatandnotes docx in the csv pilot sharepoint folder
0
628,728
20,012,427,633
IssuesEvent
2022-02-01 08:30:43
gardener/gardener
https://api.github.com/repos/gardener/gardener
closed
`shootHasBastions` check is wrong - it checks for a random Bastion in the Project namespace
kind/bug area/ops-productivity area/quality priority/3
**How to categorize this issue?** <!-- Please select area, kind, and priority for this issue. This helps the community categorizing it. Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion. If multiple identifiers make sense you can also state the commands multiple times, e.g. /area control-plane /area auto-scaling ... "/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management "/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test --> /area quality ops-productivity /kind bug **What happened**: Currently `shootHasBastions` func checks whether there is any Bastion in the Project namespace. https://github.com/gardener/gardener/blob/c2276016229f9e49dcc4652e9ad8b652e7d0f64e/pkg/gardenlet/controller/shoot/shoot_control.go#L452-L462 The following check for Bastions associated to Shoots also seems to be wrong (note the `Limit: 1` part). https://github.com/gardener/gardener/blob/c2276016229f9e49dcc4652e9ad8b652e7d0f64e/pkg/controllermanager/controller/bastion/bastion_control.go#L56-L82 **What you expected to happen**: `shootHasBastions` to check whether there is a Bastion with the appropriate `.spec.shootRef.` **How to reproduce it (as minimally and precisely as possible)**: 1. Create a Shoot in a Project that has random Bastion 2. Delete the Shoot 4. Make sure the Shoot deletion does not start/progress. Shoot status is: ```yaml lastOperation: description: Shoot cluster has been successfully reconciled. lastUpdateTime: '2022-01-26T08:38:25Z' progress: 100 state: Succeeded type: Create ``` Shoot metadata and generations: ```yaml metadata: generation: 3 creationTimestamp: '2022-01-26T08:26:48Z' deletionTimestamp: '2022-01-26T08:52:15Z' status: observedGeneration: 2 ``` gardenlet logs: ``` {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:20:50.604Z"} {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:20:50.855Z"} {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:22:32.805Z"} ``` **Anything else we need to know?**: **Environment**: - Gardener version: - Kubernetes version (use `kubectl version`): - Cloud provider or hardware configuration: - Others:
1.0
`shootHasBastions` check is wrong - it checks for a random Bastion in the Project namespace - **How to categorize this issue?** <!-- Please select area, kind, and priority for this issue. This helps the community categorizing it. Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion. If multiple identifiers make sense you can also state the commands multiple times, e.g. /area control-plane /area auto-scaling ... "/area" identifiers: audit-logging|auto-scaling|backup|certification|control-plane-migration|control-plane|cost|delivery|dev-productivity|disaster-recovery|documentation|high-availability|logging|metering|monitoring|networking|open-source|ops-productivity|os|performance|quality|robustness|scalability|security|storage|testing|usability|user-management "/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test --> /area quality ops-productivity /kind bug **What happened**: Currently `shootHasBastions` func checks whether there is any Bastion in the Project namespace. https://github.com/gardener/gardener/blob/c2276016229f9e49dcc4652e9ad8b652e7d0f64e/pkg/gardenlet/controller/shoot/shoot_control.go#L452-L462 The following check for Bastions associated to Shoots also seems to be wrong (note the `Limit: 1` part). https://github.com/gardener/gardener/blob/c2276016229f9e49dcc4652e9ad8b652e7d0f64e/pkg/controllermanager/controller/bastion/bastion_control.go#L56-L82 **What you expected to happen**: `shootHasBastions` to check whether there is a Bastion with the appropriate `.spec.shootRef.` **How to reproduce it (as minimally and precisely as possible)**: 1. Create a Shoot in a Project that has random Bastion 2. Delete the Shoot 4. Make sure the Shoot deletion does not start/progress. Shoot status is: ```yaml lastOperation: description: Shoot cluster has been successfully reconciled. lastUpdateTime: '2022-01-26T08:38:25Z' progress: 100 state: Succeeded type: Create ``` Shoot metadata and generations: ```yaml metadata: generation: 3 creationTimestamp: '2022-01-26T08:26:48Z' deletionTimestamp: '2022-01-26T08:52:15Z' status: observedGeneration: 2 ``` gardenlet logs: ``` {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:20:50.604Z"} {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:20:50.855Z"} {"level":"info","msg":"Error syncing Shoot garden-core/delete-test2: shoot has still Bastions","ts":"2022-01-26T13:22:32.805Z"} ``` **Anything else we need to know?**: **Environment**: - Gardener version: - Kubernetes version (use `kubectl version`): - Cloud provider or hardware configuration: - Others:
non_main
shoothasbastions check is wrong it checks for a random bastion in the project namespace how to categorize this issue please select area kind and priority for this issue this helps the community categorizing it replace below todos or exchange the existing identifiers with those that fit best in your opinion if multiple identifiers make sense you can also state the commands multiple times e g area control plane area auto scaling area identifiers audit logging auto scaling backup certification control plane migration control plane cost delivery dev productivity disaster recovery documentation high availability logging metering monitoring networking open source ops productivity os performance quality robustness scalability security storage testing usability user management kind identifiers api change bug cleanup discussion enhancement epic impediment poc post mortem question regression task technical debt test area quality ops productivity kind bug what happened currently shoothasbastions func checks whether there is any bastion in the project namespace the following check for bastions associated to shoots also seems to be wrong note the limit part what you expected to happen shoothasbastions to check whether there is a bastion with the appropriate spec shootref how to reproduce it as minimally and precisely as possible create a shoot in a project that has random bastion delete the shoot make sure the shoot deletion does not start progress shoot status is yaml lastoperation description shoot cluster has been successfully reconciled lastupdatetime progress state succeeded type create shoot metadata and generations yaml metadata generation creationtimestamp deletiontimestamp status observedgeneration gardenlet logs level info msg error syncing shoot garden core delete shoot has still bastions ts level info msg error syncing shoot garden core delete shoot has still bastions ts level info msg error syncing shoot garden core delete shoot has still bastions ts anything else we need to know environment gardener version kubernetes version use kubectl version cloud provider or hardware configuration others
0
52,593
22,306,759,953
IssuesEvent
2022-06-13 13:40:19
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Re-assign PRs to Budget/Review in Finance and Purchasing
Workgroup: Finance Type: Bug Report Service: Apps Product: Finance & Purchasing
<!-- Email --> <!-- sharon.ballast@austintexas.gov --> > What application are you using? Finance & Purchasing > Describe the problem. Could you please re-assign PRs 2021-11-3466, 2021-11-3467 and 2021-11-3468 to Budget/Review...they are currently in Finance Admin which is no longer active. > How soon do you need this? Flexible — An extended timeline is OK > Requested By Sharon B. Request ID: DTS22-104325
1.0
Re-assign PRs to Budget/Review in Finance and Purchasing - <!-- Email --> <!-- sharon.ballast@austintexas.gov --> > What application are you using? Finance & Purchasing > Describe the problem. Could you please re-assign PRs 2021-11-3466, 2021-11-3467 and 2021-11-3468 to Budget/Review...they are currently in Finance Admin which is no longer active. > How soon do you need this? Flexible — An extended timeline is OK > Requested By Sharon B. Request ID: DTS22-104325
non_main
re assign prs to budget review in finance and purchasing what application are you using finance purchasing describe the problem could you please re assign prs and to budget review they are currently in finance admin which is no longer active how soon do you need this flexible — an extended timeline is ok requested by sharon b request id
0
39,342
8,633,517,933
IssuesEvent
2018-11-22 14:06:09
CollabCodeTech/forum-do-front-ao-end
https://api.github.com/repos/CollabCodeTech/forum-do-front-ao-end
reopened
Minhas configs no VSCode. Manda a sua aí pra gente?
---Aula002--- VSCode [Dica]
``` { "[css]": { "editor.fontLigatures": false }, // Define o tema do VSCode "editor.fontLigatures": true, "workbench.colorTheme": "Dracula", // Configura tamanho e família da fonte "editor.fontSize": 16, "editor.lineHeight": 24, "editor.fontFamily": "Fira Code", // Aplica linhas verticais para lembrar de quebrar linha em códigos muito grandes "editor.rulers": [80, 120], // Aplica um sinal visual na esquerda da linha selecionada "editor.renderLineHighlight": "gutter", // Aumenta a fonte do terminal "terminal.integrated.fontSize": 14, // Define o tema dos ícones na sidebar "workbench.iconTheme": "material-icon-theme", "window.zoomLevel": 3, "editor.formatOnSave": true, "prettier.eslintIntegration": true, "editor.tabSize": 2, "emmet.syntaxProfiles": { "javascript": "jsx" }, "workbench.startupEditor": "newUntitledFile", "explorer.confirmDragAndDrop": false, "javascript.validate.enable": false, "files.associations": { ".babelrc": "json5", "*.html": "html", "*.js": "javascript" }, "editor.wordWrap": "on", "terminal.integrated.shell.windows": "C:\\Windows\\System32\\bash.exe", "editor.minimap.enabled": false, "eslint.autoFixOnSave": true, "eslint.alwaysShowStatus": true, "eslint.": false, "editor.formatOnPaste": true, "javascript.format.enable": false, "explorer.confirmDelete": false, "emmet.includeLanguages": { "javascript": "javascriptreact" }, "breadcrumbs.enabled": true, "workbench.activityBar.visible": false } ```
1.0
Minhas configs no VSCode. Manda a sua aí pra gente? - ``` { "[css]": { "editor.fontLigatures": false }, // Define o tema do VSCode "editor.fontLigatures": true, "workbench.colorTheme": "Dracula", // Configura tamanho e família da fonte "editor.fontSize": 16, "editor.lineHeight": 24, "editor.fontFamily": "Fira Code", // Aplica linhas verticais para lembrar de quebrar linha em códigos muito grandes "editor.rulers": [80, 120], // Aplica um sinal visual na esquerda da linha selecionada "editor.renderLineHighlight": "gutter", // Aumenta a fonte do terminal "terminal.integrated.fontSize": 14, // Define o tema dos ícones na sidebar "workbench.iconTheme": "material-icon-theme", "window.zoomLevel": 3, "editor.formatOnSave": true, "prettier.eslintIntegration": true, "editor.tabSize": 2, "emmet.syntaxProfiles": { "javascript": "jsx" }, "workbench.startupEditor": "newUntitledFile", "explorer.confirmDragAndDrop": false, "javascript.validate.enable": false, "files.associations": { ".babelrc": "json5", "*.html": "html", "*.js": "javascript" }, "editor.wordWrap": "on", "terminal.integrated.shell.windows": "C:\\Windows\\System32\\bash.exe", "editor.minimap.enabled": false, "eslint.autoFixOnSave": true, "eslint.alwaysShowStatus": true, "eslint.": false, "editor.formatOnPaste": true, "javascript.format.enable": false, "explorer.confirmDelete": false, "emmet.includeLanguages": { "javascript": "javascriptreact" }, "breadcrumbs.enabled": true, "workbench.activityBar.visible": false } ```
non_main
minhas configs no vscode manda a sua aí pra gente editor fontligatures false define o tema do vscode editor fontligatures true workbench colortheme dracula configura tamanho e família da fonte editor fontsize editor lineheight editor fontfamily fira code aplica linhas verticais para lembrar de quebrar linha em códigos muito grandes editor rulers aplica um sinal visual na esquerda da linha selecionada editor renderlinehighlight gutter aumenta a fonte do terminal terminal integrated fontsize define o tema dos ícones na sidebar workbench icontheme material icon theme window zoomlevel editor formatonsave true prettier eslintintegration true editor tabsize emmet syntaxprofiles javascript jsx workbench startupeditor newuntitledfile explorer confirmdraganddrop false javascript validate enable false files associations babelrc html html js javascript editor wordwrap on terminal integrated shell windows c windows bash exe editor minimap enabled false eslint autofixonsave true eslint alwaysshowstatus true eslint false editor formatonpaste true javascript format enable false explorer confirmdelete false emmet includelanguages javascript javascriptreact breadcrumbs enabled true workbench activitybar visible false
0
243,976
7,868,916,737
IssuesEvent
2018-06-24 06:43:42
Automattic/amp-wp
https://api.github.com/repos/Automattic/amp-wp
closed
Better Documentation
[Priority] High [Type] Task
Having everything in the readme is a bit clunky. It'd be nice to move to something more structured like Github wiki.
1.0
Better Documentation - Having everything in the readme is a bit clunky. It'd be nice to move to something more structured like Github wiki.
non_main
better documentation having everything in the readme is a bit clunky it d be nice to move to something more structured like github wiki
0
3,915
17,500,657,036
IssuesEvent
2021-08-10 09:01:30
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
MiKo_3205 should be aware of 'yield return'
bug Area: analyzer Area: maintainability
Multiple `yield return` calls without empy lines in between should not trigger a violation.
True
MiKo_3205 should be aware of 'yield return' - Multiple `yield return` calls without empy lines in between should not trigger a violation.
main
miko should be aware of yield return multiple yield return calls without empy lines in between should not trigger a violation
1
54,836
13,930,638,648
IssuesEvent
2020-10-22 03:00:22
jtimberlake/apollo
https://api.github.com/repos/jtimberlake/apollo
opened
CVE-2020-11022 (Medium) detected in jquery-1.8.3.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.3/jquery.min.js</a></p> <p>Path to dependency file: apollo/modules/tools/mobileye_viewer/location_monitor.html</p> <p>Path to vulnerable library: apollo/modules/tools/mobileye_viewer/location_monitor.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/apollo/commit/87080e19ea37501aaa61352468e295bd40aacb20">87080e19ea37501aaa61352468e295bd40aacb20</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.3","isTransitiveDependency":false,"dependencyTree":"jquery:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11022 (Medium) detected in jquery-1.8.3.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.3/jquery.min.js</a></p> <p>Path to dependency file: apollo/modules/tools/mobileye_viewer/location_monitor.html</p> <p>Path to vulnerable library: apollo/modules/tools/mobileye_viewer/location_monitor.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/apollo/commit/87080e19ea37501aaa61352468e295bd40aacb20">87080e19ea37501aaa61352468e295bd40aacb20</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.3","isTransitiveDependency":false,"dependencyTree":"jquery:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file apollo modules tools mobileye viewer location monitor html path to vulnerable library apollo modules tools mobileye viewer location monitor html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl
0
1,711
6,574,449,384
IssuesEvent
2017-09-11 12:56:30
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty"
affects_2.1 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_group / ec2 ##### ANSIBLE VERSION ``` ansible 2.1.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- just added ec2.py and ec2.ini with cache=0 --> ##### OS / ENVIRONMENT ElementaryOS Loki macOS Sierra 10.12.1 ##### SUMMARY While doing some tests suddenly i started receiving this error while creating machines: fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty"} I thought i messed something up so i reverted to my yesterday version of the code but the issue was still there. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: Engine Start hosts: localhost gather_facts: true connection: local tasks: - name: Create VPC ec2_vpc: state: present cidr_block: 172.29.0.0/16 resource_tags: { "env": "{{env}}" , "Name": "{{project_name}}" } region: "{{ aws_region }}" register: ec2_env_vpc - name: Create app private dns zone route53_zone: zone: "{{domain_zone}}" state: present vpc_id: "{{ec2_env_vpc.vpc_id}}" comment: 'Internal Zone for app' - name: Create security group ec2_group: name: "{{ project_name }}_security_group" description: "{{ project_name }} security group" region: "{{ aws_region }}" vpc_id: "{{ec2_env_vpc.vpc_id}}" rules: - proto: tcp from_port: 22 to_port: 22 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 80 to_port: 80 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 443 to_port: 443 cidr_ip: 0.0.0.0/0 rules_egress: - proto: all cidr_ip: 0.0.0.0/0 register: basic_firewall - name: Create an EC2 key ec2_key: name: "{{ project_name }}-{{ env }}-key" region: "{{ aws_region }}" register: ec2_key - name: Save private key copy: content: "{{ ec2_key.key.private_key }}" dest: "./aws-{{ env }}-private.pem" mode: 0600 when: ec2_key.changed - name: Create Redis ec2: key_name: "{{ project_name }}-{{ env }}-key" region: "{{ aws_region }}" group_id: "{{ basic_firewall.group_id }}" instance_type: "{{ instance_type }}" image: "{{ ami }}" wait: yes instance_tags: Name: "redis" env: "{{env}}" exact_count: 1 count_tag: Name: "redis" env: "{{env}}" register: ec2_redis ``` ##### EXPECTED RESULTS create machines ##### ACTUAL RESULTS errors <!--- Paste verbatim command output between quotes below --> ``` PLAY [Engine Start] ************************************************************ TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660 `" && echo ansible-tmp-1478219783.91-199174880918660="` echo $HOME/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpEYWlHH TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Create VPC] ************************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:7 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835 `" && echo ansible-tmp-1478219784.52-97557555770835="` echo $HOME/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpe3WMHj TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "igw_id": null, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "cidr_block": "172.29.0.0/16", "dns_hostnames": true, "dns_support": true, "ec2_url": null, "instance_tenancy": "default", "internet_gateway": false, "profile": null, "region": "eu-west-1", "resource_tags": {"Name": "monx", "env": "prod-ifrastrc"}, "route_tables": null, "security_token": null, "state": "present", "subnets": null, "validate_certs": true, "vpc_id": null, "wait": false, "wait_timeout": "300"}, "module_name": "ec2_vpc"}, "subnets": [], "vpc": {"cidr_block": "172.29.0.0/16", "dhcp_options_id": "dopt-ab0907ce", "id": "vpc-259fea41", "region": "eu-west-1", "state": "available"}, "vpc_id": "vpc-259fea41"} TASK [Create app private dns zone] ********************************************* task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:15 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408 `" && echo ansible-tmp-1478219786.49-20807177095408="` echo $HOME/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpHpqi6a TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "comment": "Internal Zone for app", "ec2_url": null, "profile": null, "region": null, "security_token": null, "state": "present", "validate_certs": true, "vpc_id": "vpc-259fea41", "vpc_region": null, "zone": "twui.gonova.al"}, "module_name": "route53_zone"}, "set": {"comment": "Internal Zone for app", "name": "twui.gonova.al.", "private_zone": false, "vpc_id": "vpc-259fea41", "vpc_region": null, "zone_id": "ZZWP7CFJL8WAC"}} TASK [Create security group] *************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:22 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398 `" && echo ansible-tmp-1478219787.86-187686993911398="` echo $HOME/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpaHUqBj TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "group_id": "sg-2e73a848", "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "description": "monx security group", "ec2_url": null, "name": "monx_security_group", "profile": null, "purge_rules": true, "purge_rules_egress": true, "region": "eu-west-1", "rules": [{"cidr_ip": "0.0.0.0/0", "from_port": 22, "proto": "tcp", "to_port": 22}, {"cidr_ip": "0.0.0.0/0", "from_port": 80, "proto": "tcp", "to_port": 80}, {"cidr_ip": "0.0.0.0/0", "from_port": 443, "proto": "tcp", "to_port": 443}], "rules_egress": [{"cidr_ip": "0.0.0.0/0", "from_port": null, "proto": -1, "to_port": null}], "security_token": null, "state": "present", "validate_certs": true, "vpc_id": "vpc-259fea41"}, "module_name": "ec2_group"}} TASK [Create an EC2 key] ******************************************************* task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:46 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217 `" && echo ansible-tmp-1478219789.04-264318626940217="` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpL0WO8L TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "key_material": null, "name": "monx-prod-ifrastrc-key", "profile": null, "region": "eu-west-1", "security_token": null, "state": "present", "validate_certs": true, "wait": false, "wait_timeout": "300"}, "module_name": "ec2_key"}, "key": {"fingerprint": "59:a6:0d:7d:b1:46:0d:20:e5:37:d3:16:b6:a6:17:b7:6a:03:af:6e", "name": "monx-prod-ifrastrc-key"}} TASK [Save private key] ******************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:58 skipping: [localhost] => {"changed": false, "skip_reason": "Conditional check failed", "skipped": true} TASK [Create Redis] ************************************************************ task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:65 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911 `" && echo ansible-tmp-1478219789.89-32924999230911="` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpcuaZrP TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2 <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2 && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"assign_public_ip": false, "aws_access_key": null, "aws_secret_key": null, "count": 1, "count_tag": "{'Name': 'redis', 'env': 'prod-ifrastrc'}", "ebs_optimized": false, "ec2_url": null, "exact_count": 1, "group": null, "group_id": ["sg-2e73a848"], "id": null, "image": "ami-1c4a046f", "instance_ids": null, "instance_profile_name": null, "instance_tags": {"Name": "redis", "env": "prod-ifrastrc"}, "instance_type": "t2.micro", "kernel": null, "key_name": "monx-prod-ifrastrc-key", "monitoring": false, "network_interfaces": null, "placement_group": null, "private_ip": null, "profile": null, "ramdisk": null, "region": "eu-west-1", "security_token": null, "source_dest_check": true, "spot_launch_group": null, "spot_price": null, "spot_type": "one-time", "spot_wait_timeout": "600", "state": "present", "tenancy": "default", "termination_protection": false, "user_data": null, "validate_certs": true, "volumes": null, "vpc_subnet_id": null, "wait": true, "wait_timeout": "300", "zone": null}, "module_name": "ec2"}, "msg": "Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty"} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @/Users/tarak/Dropbox/IaaS/provision/provision.retry PLAY RECAP ********************************************************************* localhost : ok=5 changed=0 unreachable=0 failed=1 ```
True
Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty" - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_group / ec2 ##### ANSIBLE VERSION ``` ansible 2.1.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- just added ec2.py and ec2.ini with cache=0 --> ##### OS / ENVIRONMENT ElementaryOS Loki macOS Sierra 10.12.1 ##### SUMMARY While doing some tests suddenly i started receiving this error while creating machines: fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty"} I thought i messed something up so i reverted to my yesterday version of the code but the issue was still there. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: Engine Start hosts: localhost gather_facts: true connection: local tasks: - name: Create VPC ec2_vpc: state: present cidr_block: 172.29.0.0/16 resource_tags: { "env": "{{env}}" , "Name": "{{project_name}}" } region: "{{ aws_region }}" register: ec2_env_vpc - name: Create app private dns zone route53_zone: zone: "{{domain_zone}}" state: present vpc_id: "{{ec2_env_vpc.vpc_id}}" comment: 'Internal Zone for app' - name: Create security group ec2_group: name: "{{ project_name }}_security_group" description: "{{ project_name }} security group" region: "{{ aws_region }}" vpc_id: "{{ec2_env_vpc.vpc_id}}" rules: - proto: tcp from_port: 22 to_port: 22 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 80 to_port: 80 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 443 to_port: 443 cidr_ip: 0.0.0.0/0 rules_egress: - proto: all cidr_ip: 0.0.0.0/0 register: basic_firewall - name: Create an EC2 key ec2_key: name: "{{ project_name }}-{{ env }}-key" region: "{{ aws_region }}" register: ec2_key - name: Save private key copy: content: "{{ ec2_key.key.private_key }}" dest: "./aws-{{ env }}-private.pem" mode: 0600 when: ec2_key.changed - name: Create Redis ec2: key_name: "{{ project_name }}-{{ env }}-key" region: "{{ aws_region }}" group_id: "{{ basic_firewall.group_id }}" instance_type: "{{ instance_type }}" image: "{{ ami }}" wait: yes instance_tags: Name: "redis" env: "{{env}}" exact_count: 1 count_tag: Name: "redis" env: "{{env}}" register: ec2_redis ``` ##### EXPECTED RESULTS create machines ##### ACTUAL RESULTS errors <!--- Paste verbatim command output between quotes below --> ``` PLAY [Engine Start] ************************************************************ TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660 `" && echo ansible-tmp-1478219783.91-199174880918660="` echo $HOME/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpEYWlHH TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/setup; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219783.91-199174880918660/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Create VPC] ************************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:7 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835 `" && echo ansible-tmp-1478219784.52-97557555770835="` echo $HOME/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpe3WMHj TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/ec2_vpc; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219784.52-97557555770835/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "igw_id": null, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "cidr_block": "172.29.0.0/16", "dns_hostnames": true, "dns_support": true, "ec2_url": null, "instance_tenancy": "default", "internet_gateway": false, "profile": null, "region": "eu-west-1", "resource_tags": {"Name": "monx", "env": "prod-ifrastrc"}, "route_tables": null, "security_token": null, "state": "present", "subnets": null, "validate_certs": true, "vpc_id": null, "wait": false, "wait_timeout": "300"}, "module_name": "ec2_vpc"}, "subnets": [], "vpc": {"cidr_block": "172.29.0.0/16", "dhcp_options_id": "dopt-ab0907ce", "id": "vpc-259fea41", "region": "eu-west-1", "state": "available"}, "vpc_id": "vpc-259fea41"} TASK [Create app private dns zone] ********************************************* task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:15 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408 `" && echo ansible-tmp-1478219786.49-20807177095408="` echo $HOME/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpHpqi6a TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/route53_zone; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219786.49-20807177095408/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "comment": "Internal Zone for app", "ec2_url": null, "profile": null, "region": null, "security_token": null, "state": "present", "validate_certs": true, "vpc_id": "vpc-259fea41", "vpc_region": null, "zone": "twui.gonova.al"}, "module_name": "route53_zone"}, "set": {"comment": "Internal Zone for app", "name": "twui.gonova.al.", "private_zone": false, "vpc_id": "vpc-259fea41", "vpc_region": null, "zone_id": "ZZWP7CFJL8WAC"}} TASK [Create security group] *************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:22 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398 `" && echo ansible-tmp-1478219787.86-187686993911398="` echo $HOME/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpaHUqBj TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/ec2_group; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219787.86-187686993911398/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "group_id": "sg-2e73a848", "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "description": "monx security group", "ec2_url": null, "name": "monx_security_group", "profile": null, "purge_rules": true, "purge_rules_egress": true, "region": "eu-west-1", "rules": [{"cidr_ip": "0.0.0.0/0", "from_port": 22, "proto": "tcp", "to_port": 22}, {"cidr_ip": "0.0.0.0/0", "from_port": 80, "proto": "tcp", "to_port": 80}, {"cidr_ip": "0.0.0.0/0", "from_port": 443, "proto": "tcp", "to_port": 443}], "rules_egress": [{"cidr_ip": "0.0.0.0/0", "from_port": null, "proto": -1, "to_port": null}], "security_token": null, "state": "present", "validate_certs": true, "vpc_id": "vpc-259fea41"}, "module_name": "ec2_group"}} TASK [Create an EC2 key] ******************************************************* task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:46 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217 `" && echo ansible-tmp-1478219789.04-264318626940217="` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpL0WO8L TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/ec2_key; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219789.04-264318626940217/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "key_material": null, "name": "monx-prod-ifrastrc-key", "profile": null, "region": "eu-west-1", "security_token": null, "state": "present", "validate_certs": true, "wait": false, "wait_timeout": "300"}, "module_name": "ec2_key"}, "key": {"fingerprint": "59:a6:0d:7d:b1:46:0d:20:e5:37:d3:16:b6:a6:17:b7:6a:03:af:6e", "name": "monx-prod-ifrastrc-key"}} TASK [Save private key] ******************************************************** task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:58 skipping: [localhost] => {"changed": false, "skip_reason": "Conditional check failed", "skipped": true} TASK [Create Redis] ************************************************************ task path: /Users/tarak/Dropbox/IaaS/provision/provision.yml:65 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: tarak <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911 `" && echo ansible-tmp-1478219789.89-32924999230911="` echo $HOME/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/md/p4_n32hs4y5_d13b5s5w96dw0000gn/T/tmpcuaZrP TO /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2 <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2 && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/ec2; rm -rf "/Users/tarak/.ansible/tmp/ansible-tmp-1478219789.89-32924999230911/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"assign_public_ip": false, "aws_access_key": null, "aws_secret_key": null, "count": 1, "count_tag": "{'Name': 'redis', 'env': 'prod-ifrastrc'}", "ebs_optimized": false, "ec2_url": null, "exact_count": 1, "group": null, "group_id": ["sg-2e73a848"], "id": null, "image": "ami-1c4a046f", "instance_ids": null, "instance_profile_name": null, "instance_tags": {"Name": "redis", "env": "prod-ifrastrc"}, "instance_type": "t2.micro", "kernel": null, "key_name": "monx-prod-ifrastrc-key", "monitoring": false, "network_interfaces": null, "placement_group": null, "private_ip": null, "profile": null, "ramdisk": null, "region": "eu-west-1", "security_token": null, "source_dest_check": true, "spot_launch_group": null, "spot_price": null, "spot_type": "one-time", "spot_wait_timeout": "600", "state": "present", "tenancy": "default", "termination_protection": false, "user_data": null, "validate_certs": true, "volumes": null, "vpc_subnet_id": null, "wait": true, "wait_timeout": "300", "zone": null}, "module_name": "ec2"}, "msg": "Instance creation failed => InvalidParameterValue: Value () for parameter groupId is invalid. The value cannot be empty"} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @/Users/tarak/Dropbox/IaaS/provision/provision.retry PLAY RECAP ********************************************************************* localhost : ok=5 changed=0 unreachable=0 failed=1 ```
main
instance creation failed invalidparametervalue value for parameter groupid is invalid the value cannot be empty issue type bug report component name group ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration just added py and ini with cache os environment elementaryos loki macos sierra summary while doing some tests suddenly i started receiving this error while creating machines fatal failed changed false failed true msg instance creation failed invalidparametervalue value for parameter groupid is invalid the value cannot be empty i thought i messed something up so i reverted to my yesterday version of the code but the issue was still there steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name engine start hosts localhost gather facts true connection local tasks name create vpc vpc state present cidr block resource tags env env name project name region aws region register env vpc name create app private dns zone zone zone domain zone state present vpc id env vpc vpc id comment internal zone for app name create security group group name project name security group description project name security group region aws region vpc id env vpc vpc id rules proto tcp from port to port cidr ip proto tcp from port to port cidr ip proto tcp from port to port cidr ip rules egress proto all cidr ip register basic firewall name create an key key name project name env key region aws region register key name save private key copy content key key private key dest aws env private pem mode when key changed name create redis key name project name env key region aws region group id basic firewall group id instance type instance type image ami wait yes instance tags name redis env env exact count count tag name redis env env register redis expected results create machines actual results errors play task establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t tmpeywlhh to users tarak ansible tmp ansible tmp setup exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp setup sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp setup rm rf users tarak ansible tmp ansible tmp dev null sleep ok task task path users tarak dropbox iaas provision provision yml establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t to users tarak ansible tmp ansible tmp vpc exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp vpc sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp vpc rm rf users tarak ansible tmp ansible tmp dev null sleep ok changed false igw id null invocation module args aws access key null aws secret key null cidr block dns hostnames true dns support true url null instance tenancy default internet gateway false profile null region eu west resource tags name monx env prod ifrastrc route tables null security token null state present subnets null validate certs true vpc id null wait false wait timeout module name vpc subnets vpc cidr block dhcp options id dopt id vpc region eu west state available vpc id vpc task task path users tarak dropbox iaas provision provision yml establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t to users tarak ansible tmp ansible tmp zone exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp zone sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp zone rm rf users tarak ansible tmp ansible tmp dev null sleep ok changed false invocation module args aws access key null aws secret key null comment internal zone for app url null profile null region null security token null state present validate certs true vpc id vpc vpc region null zone twui gonova al module name zone set comment internal zone for app name twui gonova al private zone false vpc id vpc vpc region null zone id task task path users tarak dropbox iaas provision provision yml establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t tmpahuqbj to users tarak ansible tmp ansible tmp group exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp group sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp group rm rf users tarak ansible tmp ansible tmp dev null sleep ok changed false group id sg invocation module args aws access key null aws secret key null description monx security group url null name monx security group profile null purge rules true purge rules egress true region eu west rules rules egress security token null state present validate certs true vpc id vpc module name group task task path users tarak dropbox iaas provision provision yml establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t to users tarak ansible tmp ansible tmp key exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp key sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp key rm rf users tarak ansible tmp ansible tmp dev null sleep ok changed false invocation module args aws access key null aws secret key null url null key material null name monx prod ifrastrc key profile null region eu west security token null state present validate certs true wait false wait timeout module name key key fingerprint af name monx prod ifrastrc key task task path users tarak dropbox iaas provision provision yml skipping changed false skip reason conditional check failed skipped true task task path users tarak dropbox iaas provision provision yml establish local connection for user tarak exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders md t tmpcuazrp to users tarak ansible tmp ansible tmp exec bin sh c chmod u x users tarak ansible tmp ansible tmp users tarak ansible tmp ansible tmp sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users tarak ansible tmp ansible tmp rm rf users tarak ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args assign public ip false aws access key null aws secret key null count count tag name redis env prod ifrastrc ebs optimized false url null exact count group null group id id null image ami instance ids null instance profile name null instance tags name redis env prod ifrastrc instance type micro kernel null key name monx prod ifrastrc key monitoring false network interfaces null placement group null private ip null profile null ramdisk null region eu west security token null source dest check true spot launch group null spot price null spot type one time spot wait timeout state present tenancy default termination protection false user data null validate certs true volumes null vpc subnet id null wait true wait timeout zone null module name msg instance creation failed invalidparametervalue value for parameter groupid is invalid the value cannot be empty no more hosts left to retry use limit users tarak dropbox iaas provision provision retry play recap localhost ok changed unreachable failed
1
99,119
16,430,774,830
IssuesEvent
2021-05-20 01:02:11
ngocdai94/itc230
https://api.github.com/repos/ngocdai94/itc230
opened
CVE-2020-35149 (Medium) detected in mquery-3.2.2.tgz
security vulnerability
## CVE-2020-35149 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mquery-3.2.2.tgz</b></p></summary> <p>Expressive query building for MongoDB</p> <p>Library home page: <a href="https://registry.npmjs.org/mquery/-/mquery-3.2.2.tgz">https://registry.npmjs.org/mquery/-/mquery-3.2.2.tgz</a></p> <p>Path to dependency file: itc230/package.json</p> <p>Path to vulnerable library: itc230/node_modules/mquery/package.json</p> <p> Dependency Hierarchy: - mongoose-5.7.5.tgz (Root Library) - :x: **mquery-3.2.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149>CVE-2020-35149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aheckmann/mquery/releases/tag/3.2.3">https://github.com/aheckmann/mquery/releases/tag/3.2.3</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 3.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35149 (Medium) detected in mquery-3.2.2.tgz - ## CVE-2020-35149 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mquery-3.2.2.tgz</b></p></summary> <p>Expressive query building for MongoDB</p> <p>Library home page: <a href="https://registry.npmjs.org/mquery/-/mquery-3.2.2.tgz">https://registry.npmjs.org/mquery/-/mquery-3.2.2.tgz</a></p> <p>Path to dependency file: itc230/package.json</p> <p>Path to vulnerable library: itc230/node_modules/mquery/package.json</p> <p> Dependency Hierarchy: - mongoose-5.7.5.tgz (Root Library) - :x: **mquery-3.2.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149>CVE-2020-35149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aheckmann/mquery/releases/tag/3.2.3">https://github.com/aheckmann/mquery/releases/tag/3.2.3</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 3.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in mquery tgz cve medium severity vulnerability vulnerable library mquery tgz expressive query building for mongodb library home page a href path to dependency file package json path to vulnerable library node modules mquery package json dependency hierarchy mongoose tgz root library x mquery tgz vulnerable library found in base branch master vulnerability details lib utils js in mquery before allows a pollution attack because a special property e g proto can be copied during a merge or clone operation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
442,006
30,811,160,947
IssuesEvent
2023-08-01 10:28:06
jqvirtual/jqv-sops
https://api.github.com/repos/jqvirtual/jqv-sops
closed
rank hours
documentation
edit rank hour requirements to be more achievable, maybe so -> fo 15 hours, etc. etc. make gradually harder
1.0
rank hours - edit rank hour requirements to be more achievable, maybe so -> fo 15 hours, etc. etc. make gradually harder
non_main
rank hours edit rank hour requirements to be more achievable maybe so fo hours etc etc make gradually harder
0
114,507
9,740,943,801
IssuesEvent
2019-06-02 02:39:31
fga-eps-mds/2019.1-Aix
https://api.github.com/repos/fga-eps-mds/2019.1-Aix
closed
Enviar o Coverage para o CodeClimate
EPS Teste unitário Testes
Atualizar o script do Travis CI para enviar o Coverage para o Codeclimate. **Tarefas** - [x] Atualizar o script do Travis CI. **Critérios de aceitação** - [x] Mostrar a cobertura de testes no _Badges_.
2.0
Enviar o Coverage para o CodeClimate - Atualizar o script do Travis CI para enviar o Coverage para o Codeclimate. **Tarefas** - [x] Atualizar o script do Travis CI. **Critérios de aceitação** - [x] Mostrar a cobertura de testes no _Badges_.
non_main
enviar o coverage para o codeclimate atualizar o script do travis ci para enviar o coverage para o codeclimate tarefas atualizar o script do travis ci critérios de aceitação mostrar a cobertura de testes no badges
0
150,208
13,326,152,294
IssuesEvent
2020-08-27 11:08:59
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
sys_mutex and futex missing documentation
area: Documentation bug priority: low
Both objects missing documentation and although they are documented in the code, this API documentation is not linked in the main documentation page.
1.0
sys_mutex and futex missing documentation - Both objects missing documentation and although they are documented in the code, this API documentation is not linked in the main documentation page.
non_main
sys mutex and futex missing documentation both objects missing documentation and although they are documented in the code this api documentation is not linked in the main documentation page
0
796,883
28,130,209,058
IssuesEvent
2023-03-31 21:54:42
dtcenter/METcalcpy
https://api.github.com/repos/dtcenter/METcalcpy
closed
replace frame.append method to pandas.concat in agg_stat.py
priority: blocker type: bug alert: NEED ACCOUNT KEY alert: NEED PROJECT ASSIGNMENT METcalcpy: General
## Describe the Problem ## Replace frame.append method to pandas.concat in agg_stat.py line # 1133. 'append' method produces an error: `FutureWarning: The frame.append method is deprecated and will be removed from pandas in a future version. Use pandas.concat instead.` This prevents agg_stat.py to finish successfully. ### Expected Behavior ### *Provide a clear and concise description of what you expected to happen here.* ### Environment ### Describe your runtime environment: *1. Machine: (e.g. HPC name, Linux Workstation, Mac Laptop)* *2. OS: (e.g. RedHat Linux, MacOS)* *3. Software version number(s)* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to '...'* *2. Click on '....'* *3. Scroll down to '....'* *4. See error* *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
1.0
replace frame.append method to pandas.concat in agg_stat.py - ## Describe the Problem ## Replace frame.append method to pandas.concat in agg_stat.py line # 1133. 'append' method produces an error: `FutureWarning: The frame.append method is deprecated and will be removed from pandas in a future version. Use pandas.concat instead.` This prevents agg_stat.py to finish successfully. ### Expected Behavior ### *Provide a clear and concise description of what you expected to happen here.* ### Environment ### Describe your runtime environment: *1. Machine: (e.g. HPC name, Linux Workstation, Mac Laptop)* *2. OS: (e.g. RedHat Linux, MacOS)* *3. Software version number(s)* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to '...'* *2. Click on '....'* *3. Scroll down to '....'* *4. See error* *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
non_main
replace frame append method to pandas concat in agg stat py describe the problem replace frame append method to pandas concat in agg stat py line append method produces an error futurewarning the frame append method is deprecated and will be removed from pandas in a future version use pandas concat instead this prevents agg stat py to finish successfully expected behavior provide a clear and concise description of what you expected to happen here environment describe your runtime environment machine e g hpc name linux workstation mac laptop os e g redhat linux macos software version number s to reproduce describe the steps to reproduce the behavior go to click on scroll down to see error post relevant sample data following these instructions relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and development issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and development issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
0
969
4,708,287,645
IssuesEvent
2016-10-13 22:56:04
Particular/NServiceBus.AzureServiceBus
https://api.github.com/repos/Particular/NServiceBus.AzureServiceBus
closed
V7 RTM
Tag: Maintainer Prio
## Items to complete - ~~Change package author name -> use updated NugetPackager https://github.com/Particular/V6Launch/issues/4~~ not needed - ~~[Performance issues with ConnectionString..ctor(String)](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/332)~~ not critical, can wait - [x] [Delivery count to respect Immediate retries and user settings](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/308) - [x] [Disable prefetching by default with transport transaction None](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/340) - [x] [MessageReceiverNotifier should not invoke recoverability when receiving](https://github.com/Particular/NServiceBus.AzureServiceBus/pull/349) - [x] [Outbox doesn't work with ASB transport](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/352) - [x] Create release notes (general ones, similar to the [Core ones with milestones](https://github.com/Particular/V6Launch/issues/75#issuecomment-251098093)) - [x] Update [V6Launch status list](https://github.com/Particular/V6Launch/issues/4)
True
V7 RTM - ## Items to complete - ~~Change package author name -> use updated NugetPackager https://github.com/Particular/V6Launch/issues/4~~ not needed - ~~[Performance issues with ConnectionString..ctor(String)](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/332)~~ not critical, can wait - [x] [Delivery count to respect Immediate retries and user settings](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/308) - [x] [Disable prefetching by default with transport transaction None](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/340) - [x] [MessageReceiverNotifier should not invoke recoverability when receiving](https://github.com/Particular/NServiceBus.AzureServiceBus/pull/349) - [x] [Outbox doesn't work with ASB transport](https://github.com/Particular/NServiceBus.AzureServiceBus/issues/352) - [x] Create release notes (general ones, similar to the [Core ones with milestones](https://github.com/Particular/V6Launch/issues/75#issuecomment-251098093)) - [x] Update [V6Launch status list](https://github.com/Particular/V6Launch/issues/4)
main
rtm items to complete change package author name use updated nugetpackager not needed not critical can wait create release notes general ones similar to the update
1
3,800
16,370,392,706
IssuesEvent
2021-05-15 01:52:55
Optiboot/optiboot
https://api.github.com/repos/Optiboot/optiboot
closed
Cosmetic reformatting.
Maintainability No-binary-change
The formatting (indentation, use of tabs vs spaces, etc) of optiboot.c has become very inconsistent over the years. Both the SpenceKonde and MCUDude branches are in better shape; the "master" repository ought to get its act together: 1. 2-space indent default 2. no tabs 3. fix trailing whitespace 4. fix spelling mistakes 5. do something to make auto-indent work properly in spite of the conditional compilation Do this separately from any actual code changes!
True
Cosmetic reformatting. - The formatting (indentation, use of tabs vs spaces, etc) of optiboot.c has become very inconsistent over the years. Both the SpenceKonde and MCUDude branches are in better shape; the "master" repository ought to get its act together: 1. 2-space indent default 2. no tabs 3. fix trailing whitespace 4. fix spelling mistakes 5. do something to make auto-indent work properly in spite of the conditional compilation Do this separately from any actual code changes!
main
cosmetic reformatting the formatting indentation use of tabs vs spaces etc of optiboot c has become very inconsistent over the years both the spencekonde and mcudude branches are in better shape the master repository ought to get its act together space indent default no tabs fix trailing whitespace fix spelling mistakes do something to make auto indent work properly in spite of the conditional compilation do this separately from any actual code changes
1
1,633
6,572,657,467
IssuesEvent
2017-09-11 04:08:40
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
stat module doesn't return lnk_source when follow=yes
affects_2.0 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME stat ##### ANSIBLE VERSION ``` ansible 2.0.1.0 (detached HEAD bb6cadefa2) last updated 2016/04/13 12:36:28 (GMT -700) lib/ansible/modules/core: (detached HEAD 262e2a3302) last updated 2016/04/13 12:36:28 (GMT -700) lib/ansible/modules/extras: (detached HEAD e0be11da08) last updated 2016/04/13 12:36:28 (GMT -700) config file = <project repo>/ansible.cfg configured module search path = /usr/share/ansible:playbooks/library ``` ##### CONFIGURATION Irrelevant ##### OS / ENVIRONMENT Host: Fedora 23, kernel 4.4.6-300.fc23.x86_64, python 2.7.11 Target: CentOS 6.7, kernel 2.6.32-573.7.1.el6.x86_64, python 2.6.6 ##### SUMMARY The stat module fails to return the lnk_source attribute as part of its registered output. When trying to retrieve the target path and target attributes of a symlink, one must use stat twice - the first time with follow=no to retrieve lnk_source, and a second time with follow=yes to retrieve the remaining attributes. In both follow settings, 'path' is returned as the symlink path being inspected. ##### STEPS TO REPRODUCE ``` - stat: path="/some/symlink" follow=yes register: stat_result - debug: var=stat_result.stat ``` ##### EXPECTED RESULTS ``` "stat": { "atime": 1460493437.5765483, "ctime": 1459867813.2843106, "dev": 2064, "exists": true, "gid": 1003, "gr_name": "nobody", "inode": 2107891, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/target/file", "mode": "0755", "mtime": 1459867813.2843106, "nlink": 17, "path": "/some/symlink", "pw_name": "nobody", "rgrp": true, "roth": true, "rusr": true, "size": 4096, "uid": 1003, "wgrp": false, "woth": false, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } ``` ##### ACTUAL RESULTS ``` "stat": { "atime": 1460493437.5765483, "ctime": 1459867813.2843106, "dev": 2064, "exists": true, "gid": 1003, "gr_name": "nobody", "inode": 2107891, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mode": "0755", "mtime": 1459867813.2843106, "nlink": 17, "path": "/some/symlink", "pw_name": "nobody", "rgrp": true, "roth": true, "rusr": true, "size": 4096, "uid": 1003, "wgrp": false, "woth": false, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } ```
True
stat module doesn't return lnk_source when follow=yes - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME stat ##### ANSIBLE VERSION ``` ansible 2.0.1.0 (detached HEAD bb6cadefa2) last updated 2016/04/13 12:36:28 (GMT -700) lib/ansible/modules/core: (detached HEAD 262e2a3302) last updated 2016/04/13 12:36:28 (GMT -700) lib/ansible/modules/extras: (detached HEAD e0be11da08) last updated 2016/04/13 12:36:28 (GMT -700) config file = <project repo>/ansible.cfg configured module search path = /usr/share/ansible:playbooks/library ``` ##### CONFIGURATION Irrelevant ##### OS / ENVIRONMENT Host: Fedora 23, kernel 4.4.6-300.fc23.x86_64, python 2.7.11 Target: CentOS 6.7, kernel 2.6.32-573.7.1.el6.x86_64, python 2.6.6 ##### SUMMARY The stat module fails to return the lnk_source attribute as part of its registered output. When trying to retrieve the target path and target attributes of a symlink, one must use stat twice - the first time with follow=no to retrieve lnk_source, and a second time with follow=yes to retrieve the remaining attributes. In both follow settings, 'path' is returned as the symlink path being inspected. ##### STEPS TO REPRODUCE ``` - stat: path="/some/symlink" follow=yes register: stat_result - debug: var=stat_result.stat ``` ##### EXPECTED RESULTS ``` "stat": { "atime": 1460493437.5765483, "ctime": 1459867813.2843106, "dev": 2064, "exists": true, "gid": 1003, "gr_name": "nobody", "inode": 2107891, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/target/file", "mode": "0755", "mtime": 1459867813.2843106, "nlink": 17, "path": "/some/symlink", "pw_name": "nobody", "rgrp": true, "roth": true, "rusr": true, "size": 4096, "uid": 1003, "wgrp": false, "woth": false, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } ``` ##### ACTUAL RESULTS ``` "stat": { "atime": 1460493437.5765483, "ctime": 1459867813.2843106, "dev": 2064, "exists": true, "gid": 1003, "gr_name": "nobody", "inode": 2107891, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mode": "0755", "mtime": 1459867813.2843106, "nlink": 17, "path": "/some/symlink", "pw_name": "nobody", "rgrp": true, "roth": true, "rusr": true, "size": 4096, "uid": 1003, "wgrp": false, "woth": false, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } ```
main
stat module doesn t return lnk source when follow yes issue type bug report component name stat ansible version ansible detached head last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file ansible cfg configured module search path usr share ansible playbooks library configuration irrelevant os environment host fedora kernel python target centos kernel python summary the stat module fails to return the lnk source attribute as part of its registered output when trying to retrieve the target path and target attributes of a symlink one must use stat twice the first time with follow no to retrieve lnk source and a second time with follow yes to retrieve the remaining attributes in both follow settings path is returned as the symlink path being inspected steps to reproduce stat path some symlink follow yes register stat result debug var stat result stat expected results stat atime ctime dev exists true gid gr name nobody inode isblk false ischr false isdir true isfifo false isgid false islnk false isreg false issock false isuid false lnk source target file mode mtime nlink path some symlink pw name nobody rgrp true roth true rusr true size uid wgrp false woth false wusr true xgrp true xoth true xusr true actual results stat atime ctime dev exists true gid gr name nobody inode isblk false ischr false isdir true isfifo false isgid false islnk false isreg false issock false isuid false mode mtime nlink path some symlink pw name nobody rgrp true roth true rusr true size uid wgrp false woth false wusr true xgrp true xoth true xusr true
1
5,350
26,963,534,416
IssuesEvent
2023-02-08 20:12:21
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
overlay2 directory not created, but is referenced
area/docker stage/needs-investigation maintainer/need-followup
ubuntu18-04 h009357:poc$ sam --version SAM CLI, version 0.19.0 h009357:poc$ sam validate 2019-08-06 12:23:19 Found credentials in shared credentials file: ~/.aws/credentials /home/u0138544/aws/samtest/ppr-app/poc/template.yaml is a valid SAM Template h009357:poc$ sam local start-api 2019-08-06 12:08:23 Mounting LoginFunction at http://127.0.0.1:3000/login [GET] 2019-08-06 12:08:23 Mounting FilePairsFunction at http://127.0.0.1:3000/file-pairs [GET] 2019-08-06 12:08:23 You can now browse to the above endpoints to invoke your functions. You do not need to restart/reload SAM CLI while working on your functions, changes will be reflected instantly/automatically. You only need to restart SAM CLI if you update your AWS SAM template 2019-08-06 12:08:23 * Running on http://127.0.0.1:3000/ (Press CTRL+C to quit) 2019-08-06 12:08:38 Invoking app.lambda_handler (python3.7) 2019-08-06 12:08:38 Found credentials in shared credentials file: ~/.aws/credentials 2019-08-06 12:08:38 Exception on /login [GET] Traceback (most recent call last): File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 261, in _raise_for_status response.raise_for_status() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/requests/models.py", line 940, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.35/images/lambci/lambda:python3.7/json During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 2317, in wsgi_app response = self.full_dispatch_request() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1840, in full_dispatch_request rv = self.handle_user_exception(e) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1743, in handle_user_exception reraise(exc_type, exc_value, tb) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/_compat.py", line 36, in reraise raise value File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1838, in full_dispatch_request rv = self.dispatch_request() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1824, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/apigw/local_apigw_service.py", line 156, in _request_handler self.lambda_runner.invoke(route.function_name, event, stdout=stdout_stream_writer, stderr=self.stderr) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/commands/local/lib/local_lambda.py", line 93, in invoke self.local_runtime.invoke(config, event, debug_context=self.debug_context, stdout=stdout, stderr=stderr) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/lambdafn/runtime.py", line 86, in invoke self._container_manager.run(container) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/docker/manager.py", line 75, in run is_image_local = self.has_image(image_name) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/docker/manager.py", line 153, in has_image self.docker_client.images.get(image_name) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/models/images.py", line 316, in get return self.prepare_model(self.client.api.inspect_image(name)) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/utils/decorators.py", line 19, in wrapped return f(self, resource_id, *args, **kwargs) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/image.py", line 245, in inspect_image self._get(self._url("/images/{0}/json", image)), True File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 267, in _result self._raise_for_status(response) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 263, in _raise_for_status raise create_api_error_from_http_exception(e) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/errors.py", line 31, in create_api_error_from_http_exception raise cls(e, response=response, explanation=explanation) docker.errors.APIError: 500 Server Error: Internal Server Error ("stat /var/lib/docker/overlay2/f7b2dd847f0251502948f34f67fbe8c1fc03d8fed5ede4c883982e06cad1477e: no such file or directory") 2019-08-06 12:08:38 127.0.0.1 - - [06/Aug/2019 12:08:38] "GET /login HTTP/1.1" 502 - 2019-08-06 12:08:38 127.0.0.1 - - [06/Aug/2019 12:08:38] "GET /favicon.ico HTTP/1.1" 403 - root@h009357:/var/lib/docker/overlay2# pwd /var/lib/docker/overlay2 root@h009357:/var/lib/docker/overlay2# ls -ltr total 4 drwx------ 2 root root 4096 Aug 6 12:08 l
True
overlay2 directory not created, but is referenced - ubuntu18-04 h009357:poc$ sam --version SAM CLI, version 0.19.0 h009357:poc$ sam validate 2019-08-06 12:23:19 Found credentials in shared credentials file: ~/.aws/credentials /home/u0138544/aws/samtest/ppr-app/poc/template.yaml is a valid SAM Template h009357:poc$ sam local start-api 2019-08-06 12:08:23 Mounting LoginFunction at http://127.0.0.1:3000/login [GET] 2019-08-06 12:08:23 Mounting FilePairsFunction at http://127.0.0.1:3000/file-pairs [GET] 2019-08-06 12:08:23 You can now browse to the above endpoints to invoke your functions. You do not need to restart/reload SAM CLI while working on your functions, changes will be reflected instantly/automatically. You only need to restart SAM CLI if you update your AWS SAM template 2019-08-06 12:08:23 * Running on http://127.0.0.1:3000/ (Press CTRL+C to quit) 2019-08-06 12:08:38 Invoking app.lambda_handler (python3.7) 2019-08-06 12:08:38 Found credentials in shared credentials file: ~/.aws/credentials 2019-08-06 12:08:38 Exception on /login [GET] Traceback (most recent call last): File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 261, in _raise_for_status response.raise_for_status() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/requests/models.py", line 940, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: http+docker://localhost/v1.35/images/lambci/lambda:python3.7/json During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 2317, in wsgi_app response = self.full_dispatch_request() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1840, in full_dispatch_request rv = self.handle_user_exception(e) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1743, in handle_user_exception reraise(exc_type, exc_value, tb) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/_compat.py", line 36, in reraise raise value File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1838, in full_dispatch_request rv = self.dispatch_request() File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/flask/app.py", line 1824, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/apigw/local_apigw_service.py", line 156, in _request_handler self.lambda_runner.invoke(route.function_name, event, stdout=stdout_stream_writer, stderr=self.stderr) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/commands/local/lib/local_lambda.py", line 93, in invoke self.local_runtime.invoke(config, event, debug_context=self.debug_context, stdout=stdout, stderr=stderr) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/lambdafn/runtime.py", line 86, in invoke self._container_manager.run(container) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/docker/manager.py", line 75, in run is_image_local = self.has_image(image_name) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/samcli/local/docker/manager.py", line 153, in has_image self.docker_client.images.get(image_name) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/models/images.py", line 316, in get return self.prepare_model(self.client.api.inspect_image(name)) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/utils/decorators.py", line 19, in wrapped return f(self, resource_id, *args, **kwargs) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/image.py", line 245, in inspect_image self._get(self._url("/images/{0}/json", image)), True File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 267, in _result self._raise_for_status(response) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/api/client.py", line 263, in _raise_for_status raise create_api_error_from_http_exception(e) File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/0.19.0/libexec/lib/python3.7/site-packages/docker/errors.py", line 31, in create_api_error_from_http_exception raise cls(e, response=response, explanation=explanation) docker.errors.APIError: 500 Server Error: Internal Server Error ("stat /var/lib/docker/overlay2/f7b2dd847f0251502948f34f67fbe8c1fc03d8fed5ede4c883982e06cad1477e: no such file or directory") 2019-08-06 12:08:38 127.0.0.1 - - [06/Aug/2019 12:08:38] "GET /login HTTP/1.1" 502 - 2019-08-06 12:08:38 127.0.0.1 - - [06/Aug/2019 12:08:38] "GET /favicon.ico HTTP/1.1" 403 - root@h009357:/var/lib/docker/overlay2# pwd /var/lib/docker/overlay2 root@h009357:/var/lib/docker/overlay2# ls -ltr total 4 drwx------ 2 root root 4096 Aug 6 12:08 l
main
directory not created but is referenced poc sam version sam cli version poc sam validate found credentials in shared credentials file aws credentials home aws samtest ppr app poc template yaml is a valid sam template poc sam local start api mounting loginfunction at mounting filepairsfunction at you can now browse to the above endpoints to invoke your functions you do not need to restart reload sam cli while working on your functions changes will be reflected instantly automatically you only need to restart sam cli if you update your aws sam template running on press ctrl c to quit invoking app lambda handler found credentials in shared credentials file aws credentials exception on login traceback most recent call last file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker api client py line in raise for status response raise for status file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages requests models py line in raise for status raise httperror http error msg response self requests exceptions httperror server error internal server error for url http docker localhost images lambci lambda json during handling of the above exception another exception occurred traceback most recent call last file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask app py line in wsgi app response self full dispatch request file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask app py line in full dispatch request rv self handle user exception e file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask app py line in handle user exception reraise exc type exc value tb file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask compat py line in reraise raise value file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask app py line in full dispatch request rv self dispatch request file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages flask app py line in dispatch request return self view functions req view args file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local apigw local apigw service py line in request handler self lambda runner invoke route function name event stdout stdout stream writer stderr self stderr file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli commands local lib local lambda py line in invoke self local runtime invoke config event debug context self debug context stdout stdout stderr stderr file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local lambdafn runtime py line in invoke self container manager run container file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local docker manager py line in run is image local self has image image name file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local docker manager py line in has image self docker client images get image name file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker models images py line in get return self prepare model self client api inspect image name file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker utils decorators py line in wrapped return f self resource id args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker api image py line in inspect image self get self url images json image true file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker api client py line in result self raise for status response file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker api client py line in raise for status raise create api error from http exception e file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages docker errors py line in create api error from http exception raise cls e response response explanation explanation docker errors apierror server error internal server error stat var lib docker no such file or directory get login http get favicon ico http root var lib docker pwd var lib docker root var lib docker ls ltr total drwx root root aug l
1
3,736
15,649,373,583
IssuesEvent
2021-03-23 07:28:45
skku-npc/skku-coding-platform
https://api.github.com/repos/skku-npc/skku-coding-platform
opened
Replace tar-simditor-markdown
frontend maintain security
## Issue 더 이상 지원되지 않는 패키지인 `tar-simditor-markdown` 대체하기 ## Why? ### Cross-Site Scripting (XSS) Issue 구버전의 jquery를 사용해 XSS Issue 발생 [XSS(Cross-Site Scripting) 이란?](https://noirstar.tistory.com/266) [NPM Advisory Info](https://www.npmjs.com/advisories/328) ### No Longer Maintained 2017년 이후로 업데이트 X https://github.com/itargaryen/tar-simditor ![image](https://user-images.githubusercontent.com/19747913/112109192-98fe0f00-8bf4-11eb-9daf-4ab3b29d2182.png)
True
Replace tar-simditor-markdown - ## Issue 더 이상 지원되지 않는 패키지인 `tar-simditor-markdown` 대체하기 ## Why? ### Cross-Site Scripting (XSS) Issue 구버전의 jquery를 사용해 XSS Issue 발생 [XSS(Cross-Site Scripting) 이란?](https://noirstar.tistory.com/266) [NPM Advisory Info](https://www.npmjs.com/advisories/328) ### No Longer Maintained 2017년 이후로 업데이트 X https://github.com/itargaryen/tar-simditor ![image](https://user-images.githubusercontent.com/19747913/112109192-98fe0f00-8bf4-11eb-9daf-4ab3b29d2182.png)
main
replace tar simditor markdown issue 더 이상 지원되지 않는 패키지인 tar simditor markdown 대체하기 why cross site scripting xss issue 구버전의 jquery를 사용해 xss issue 발생 no longer maintained 이후로 업데이트 x
1