Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
92,538 | 11,671,570,449 | IssuesEvent | 2020-03-04 03:39:11 | microsoft/nni | https://api.github.com/repos/microsoft/nni | closed | Customized trial | design discussion new feature | ## Current Design
### User Side:
* Web UI adds a "clone" button for each trial
* Web UI adds a "tag" column in the trial details table (and maybe "edit tag" buttons)
* When user wants to create a customized trial, they click "clone" and a form pops up
* User edits the parameters and the *tag* (defaults to `CUSTOMIZED` maybe)
* User clicks "submit" button
* Web UI checks whether the parameters are out of search space
- If yes, web UI warns the user that their parameter may break the tuner
* Web UI submits the parameters via REST API
* _Maybe web UI should show a dialog saying the hyper-parameter is successfully submitted but may take some time to get scheduled and to show up_
* If user wants to find the customized trial, they search its tag
### Tuner Side
By default tuner algorithms will not receive metrics from customized trials.
If an algorithm explicitly accepts customized trials (by implementing `receive_customized_trial_result` or by calling `accept_customized_trials` or whatever), results of all customized hyper-parameters will be reported to the tuner.
NNI SDK does NOT guarantee the legality of customized hyper-parameters. Tuners are encouraged to check it themselves.
However, since the user must confirm they know what they are doing before submitting an out-of-range hyper-parameter, tuners can feel free to ignore such case.
### NNI Manager
TODO: check the current logic for appending jobs after experiment is done. | 1.0 | Customized trial - ## Current Design
### User Side:
* Web UI adds a "clone" button for each trial
* Web UI adds a "tag" column in the trial details table (and maybe "edit tag" buttons)
* When user wants to create a customized trial, they click "clone" and a form pops up
* User edits the parameters and the *tag* (defaults to `CUSTOMIZED` maybe)
* User clicks "submit" button
* Web UI checks whether the parameters are out of search space
- If yes, web UI warns the user that their parameter may break the tuner
* Web UI submits the parameters via REST API
* _Maybe web UI should show a dialog saying the hyper-parameter is successfully submitted but may take some time to get scheduled and to show up_
* If user wants to find the customized trial, they search its tag
### Tuner Side
By default tuner algorithms will not receive metrics from customized trials.
If an algorithm explicitly accepts customized trials (by implementing `receive_customized_trial_result` or by calling `accept_customized_trials` or whatever), results of all customized hyper-parameters will be reported to the tuner.
NNI SDK does NOT guarantee the legality of customized hyper-parameters. Tuners are encouraged to check it themselves.
However, since the user must confirm they know what they are doing before submitting an out-of-range hyper-parameter, tuners can feel free to ignore such case.
### NNI Manager
TODO: check the current logic for appending jobs after experiment is done. | non_priority | customized trial current design user side web ui adds a clone button for each trial web ui adds a tag column in the trial details table and maybe edit tag buttons when user wants to create a customized trial they click clone and a form pops up user edits the parameters and the tag defaults to customized maybe user clicks submit button web ui checks whether the parameters are out of search space if yes web ui warns the user that their parameter may break the tuner web ui submits the parameters via rest api maybe web ui should show a dialog saying the hyper parameter is successfully submitted but may take some time to get scheduled and to show up if user wants to find the customized trial they search its tag tuner side by default tuner algorithms will not receive metrics from customized trials if an algorithm explicitly accepts customized trials by implementing receive customized trial result or by calling accept customized trials or whatever results of all customized hyper parameters will be reported to the tuner nni sdk does not guarantee the legality of customized hyper parameters tuners are encouraged to check it themselves however since the user must confirm they know what they are doing before submitting an out of range hyper parameter tuners can feel free to ignore such case nni manager todo check the current logic for appending jobs after experiment is done | 0 |
736,847 | 25,490,263,956 | IssuesEvent | 2022-11-27 00:39:22 | jeremilev/Soen287-project | https://api.github.com/repos/jeremilev/Soen287-project | closed | Database: Use regular queries to update grades for course and student data, since cloud functions are paid | Priority 1 | write db queries that will allow the teacher to insert grades to students, we are not using cloud functions anymore | 1.0 | Database: Use regular queries to update grades for course and student data, since cloud functions are paid - write db queries that will allow the teacher to insert grades to students, we are not using cloud functions anymore | priority | database use regular queries to update grades for course and student data since cloud functions are paid write db queries that will allow the teacher to insert grades to students we are not using cloud functions anymore | 1 |
438,405 | 12,627,410,810 | IssuesEvent | 2020-06-14 21:20:51 | domialex/Sidekick | https://api.github.com/repos/domialex/Sidekick | closed | Make windows/panels resizable | Priority: Medium Status: Review Needed Type: Enhancement | # Resizable windows
- [x] Splash screen
- [x] Price view
- [ ] ~~League view~~
- [x] Settings (vertical only)
- [x] Logs | 1.0 | Make windows/panels resizable - # Resizable windows
- [x] Splash screen
- [x] Price view
- [ ] ~~League view~~
- [x] Settings (vertical only)
- [x] Logs | priority | make windows panels resizable resizable windows splash screen price view league view settings vertical only logs | 1 |
289,845 | 25,017,695,703 | IssuesEvent | 2022-11-03 20:25:19 | cfpb/hmda-frontend | https://api.github.com/repos/cfpb/hmda-frontend | closed | [Cypress] Login bug; Load test too big | Test Automation | - I thought the 500K limit was in the generation script but rather it is a limit of the way Cypress handles file uploads.
- I missed updating the login method for on of the `HMDA Help - Institution` tests | 1.0 | [Cypress] Login bug; Load test too big - - I thought the 500K limit was in the generation script but rather it is a limit of the way Cypress handles file uploads.
- I missed updating the login method for on of the `HMDA Help - Institution` tests | non_priority | login bug load test too big i thought the limit was in the generation script but rather it is a limit of the way cypress handles file uploads i missed updating the login method for on of the hmda help institution tests | 0 |
808,816 | 30,112,334,990 | IssuesEvent | 2023-06-30 08:46:06 | fossasia/open-event-frontend | https://api.github.com/repos/fossasia/open-event-frontend | reopened | Show test/live key of Stripe under Payments section | bug Priority: High Priority: Urgent | **Describe the bug**
Show test/live key of Stripe under Payments section depending on the environment being used, if local/test.eventyay.com is used - show the test keys else show live keys. Currently, the live keys are shown in both environments.

| 2.0 | Show test/live key of Stripe under Payments section - **Describe the bug**
Show test/live key of Stripe under Payments section depending on the environment being used, if local/test.eventyay.com is used - show the test keys else show live keys. Currently, the live keys are shown in both environments.

| priority | show test live key of stripe under payments section describe the bug show test live key of stripe under payments section depending on the environment being used if local test eventyay com is used show the test keys else show live keys currently the live keys are shown in both environments | 1 |
45,421 | 12,797,914,052 | IssuesEvent | 2020-07-02 13:08:03 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | Q1 Critical defects | Defect Epic ⭐️ VFS Product Team support | ## Background
### User Story or Problem Statement
<As a _____, I need _____ so I can _____.>
_or_
<Problem description. How might we _____________ ?>
### Affected users and stakeholders
* CMS_editors_of_X_product
* Developers of Y teams
<link to any existing research or data supporting this>
### Hypothesis
_A hypothesis may depend on a spike ticket to be completed._
We believe that _this_solution_ will achieve _this_outcome_. We'll know that to be true when _this measurable outcome occurs._
### Assumptions
* Tk
* Tk
(How will these assumptions be validated?)
## Acceptance Criteria
* Testable_Outcome_X
* Testable_Outcome_Y
* Testable_Outcome_Z
## Runbook
This is particularly relevant for feature launches, or for making changes to the content model that require front end changes as well.
1. Step_to_get_to_production_X
1. Step_to_get_to_production_Y
1. Step_to_get_to_production_Z
## Possible tickets to create for this epic
- [ ] Spike task, with some of the following attributes:
* Identify the baseline against which we will measure improvements
* Ideation (sketching, cross-displinary meeting of minds around how the problem may be solved)
* Analysis of the drupal landscape
* Possible implementation(s)
* Proof of concept or prototype
* UX writing discovery
- [ ] User research
- [ ] Design
- [ ] UX writing
- [ ] Usability testing (eg with a prototype or other design stimula)
- [ ] Implementation
- [ ] Documentation (What needs changing or adding to the documentation?
- [ ] Product release notes (How will we describe this feature to CMS users?)
- [ ] Training (Will any users require additional training?)
- [ ] Metrics (How will we measure any improvements)
## Labels for this issue
- [/] Issue type (red) (defaults to Epic)
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [ ] CMS objective (orange)
- [ ] CMS-supported product (black)
| 1.0 | Q1 Critical defects - ## Background
### User Story or Problem Statement
<As a _____, I need _____ so I can _____.>
_or_
<Problem description. How might we _____________ ?>
### Affected users and stakeholders
* CMS_editors_of_X_product
* Developers of Y teams
<link to any existing research or data supporting this>
### Hypothesis
_A hypothesis may depend on a spike ticket to be completed._
We believe that _this_solution_ will achieve _this_outcome_. We'll know that to be true when _this measurable outcome occurs._
### Assumptions
* Tk
* Tk
(How will these assumptions be validated?)
## Acceptance Criteria
* Testable_Outcome_X
* Testable_Outcome_Y
* Testable_Outcome_Z
## Runbook
This is particularly relevant for feature launches, or for making changes to the content model that require front end changes as well.
1. Step_to_get_to_production_X
1. Step_to_get_to_production_Y
1. Step_to_get_to_production_Z
## Possible tickets to create for this epic
- [ ] Spike task, with some of the following attributes:
* Identify the baseline against which we will measure improvements
* Ideation (sketching, cross-displinary meeting of minds around how the problem may be solved)
* Analysis of the drupal landscape
* Possible implementation(s)
* Proof of concept or prototype
* UX writing discovery
- [ ] User research
- [ ] Design
- [ ] UX writing
- [ ] Usability testing (eg with a prototype or other design stimula)
- [ ] Implementation
- [ ] Documentation (What needs changing or adding to the documentation?
- [ ] Product release notes (How will we describe this feature to CMS users?)
- [ ] Training (Will any users require additional training?)
- [ ] Metrics (How will we measure any improvements)
## Labels for this issue
- [/] Issue type (red) (defaults to Epic)
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [ ] CMS objective (orange)
- [ ] CMS-supported product (black)
| non_priority | critical defects background user story or problem statement or affected users and stakeholders cms editors of x product developers of y teams hypothesis a hypothesis may depend on a spike ticket to be completed we believe that this solution will achieve this outcome we ll know that to be true when this measurable outcome occurs assumptions tk tk how will these assumptions be validated acceptance criteria testable outcome x testable outcome y testable outcome z runbook this is particularly relevant for feature launches or for making changes to the content model that require front end changes as well step to get to production x step to get to production y step to get to production z possible tickets to create for this epic spike task with some of the following attributes identify the baseline against which we will measure improvements ideation sketching cross displinary meeting of minds around how the problem may be solved analysis of the drupal landscape possible implementation s proof of concept or prototype ux writing discovery user research design ux writing usability testing eg with a prototype or other design stimula implementation documentation what needs changing or adding to the documentation product release notes how will we describe this feature to cms users training will any users require additional training metrics how will we measure any improvements labels for this issue issue type red defaults to epic cms subsystem green cms practice area blue cms objective orange cms supported product black | 0 |
554,245 | 16,415,388,644 | IssuesEvent | 2021-05-19 05:47:36 | UvA-FNWI/GradeTool | https://api.github.com/repos/UvA-FNWI/GradeTool | closed | Do not submit grades that are not shown | bug priority: high | ### Component
Grade submission in main screen.
### Summary
When submitting grades with partial grading enabled, we should always apply the same filter as shown (e.g. when filtering by group or student name). Right now you can submit grades for students that are not shown if you had selected them before filtering.
### Steps to reproduce
Go to a course with partial grading enabled and at least two groups A and B. Select one student from each group, then filter by group B and submit. The grade for the student from group A is also submitted.
### Reported by
FGw | 1.0 | Do not submit grades that are not shown - ### Component
Grade submission in main screen.
### Summary
When submitting grades with partial grading enabled, we should always apply the same filter as shown (e.g. when filtering by group or student name). Right now you can submit grades for students that are not shown if you had selected them before filtering.
### Steps to reproduce
Go to a course with partial grading enabled and at least two groups A and B. Select one student from each group, then filter by group B and submit. The grade for the student from group A is also submitted.
### Reported by
FGw | priority | do not submit grades that are not shown component grade submission in main screen summary when submitting grades with partial grading enabled we should always apply the same filter as shown e g when filtering by group or student name right now you can submit grades for students that are not shown if you had selected them before filtering steps to reproduce go to a course with partial grading enabled and at least two groups a and b select one student from each group then filter by group b and submit the grade for the student from group a is also submitted reported by fgw | 1 |
23,796 | 12,123,741,009 | IssuesEvent | 2020-04-22 13:12:14 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Lazy loading React components | Team:AppArch Team:Geo Team:Platform discuss performance | I created this issue to discuss the lazy loading strategy for react components used by different registries in Kibana.
If you [build Kibana platform plugins locally and check their sizes](https://github.com/elastic/kibana/blob/master/src/core/MIGRATION.md#how-to-understand-how-big-the-bundle-size-of-my-plugin-is). You can notice that `maps` plugin size reaches 4.8Mb.
That's mostly due to new `optimizer` architecture including all the plugin dependencies in the bundle and lack of code tree-shaking (WIP https://github.com/elastic/kibana/pull/62390). Even if we move some deps to shared dependencies and remove unused code, the build size will be quite significant (@elastic/ems-client, @elastic/maki only gives us 1Mb)
Ideally, a plugin code should be as slim as possible and include the thing that necessary for proper **registration**. All other code must be loaded on demand.
I'm wondering if we can improve the situation with loading Maps React components on demand.
To understand how it affects the initial load time let's look at `maps.plugin.js` size (this is the chunk loaded **every time** a user opens the Kibana):
without lazy loading - 4.8Mb
with `MapViewComponent` loaded lazily - 29Kb
There are several approaches to address the problem:
- we can add lazy loading for React components (https://github.com/jamiebuilds/react-loadable, for example)
- we can make registry compatible with async integration points (as [management app does](https://github.com/elastic/kibana/blob/8e17fdabd4954ee49ab6d2e327351cb20abc6e77/src/plugins/management/public/types.ts#L59))
Maybe @elastic/kibana-app-arch has other ideas or opinion of whatever pattern is preferable? | True | Lazy loading React components - I created this issue to discuss the lazy loading strategy for react components used by different registries in Kibana.
If you [build Kibana platform plugins locally and check their sizes](https://github.com/elastic/kibana/blob/master/src/core/MIGRATION.md#how-to-understand-how-big-the-bundle-size-of-my-plugin-is). You can notice that `maps` plugin size reaches 4.8Mb.
That's mostly due to new `optimizer` architecture including all the plugin dependencies in the bundle and lack of code tree-shaking (WIP https://github.com/elastic/kibana/pull/62390). Even if we move some deps to shared dependencies and remove unused code, the build size will be quite significant (@elastic/ems-client, @elastic/maki only gives us 1Mb)
Ideally, a plugin code should be as slim as possible and include the thing that necessary for proper **registration**. All other code must be loaded on demand.
I'm wondering if we can improve the situation with loading Maps React components on demand.
To understand how it affects the initial load time let's look at `maps.plugin.js` size (this is the chunk loaded **every time** a user opens the Kibana):
without lazy loading - 4.8Mb
with `MapViewComponent` loaded lazily - 29Kb
There are several approaches to address the problem:
- we can add lazy loading for React components (https://github.com/jamiebuilds/react-loadable, for example)
- we can make registry compatible with async integration points (as [management app does](https://github.com/elastic/kibana/blob/8e17fdabd4954ee49ab6d2e327351cb20abc6e77/src/plugins/management/public/types.ts#L59))
Maybe @elastic/kibana-app-arch has other ideas or opinion of whatever pattern is preferable? | non_priority | lazy loading react components i created this issue to discuss the lazy loading strategy for react components used by different registries in kibana if you you can notice that maps plugin size reaches that s mostly due to new optimizer architecture including all the plugin dependencies in the bundle and lack of code tree shaking wip even if we move some deps to shared dependencies and remove unused code the build size will be quite significant elastic ems client elastic maki only gives us ideally a plugin code should be as slim as possible and include the thing that necessary for proper registration all other code must be loaded on demand i m wondering if we can improve the situation with loading maps react components on demand to understand how it affects the initial load time let s look at maps plugin js size this is the chunk loaded every time a user opens the kibana without lazy loading with mapviewcomponent loaded lazily there are several approaches to address the problem we can add lazy loading for react components for example we can make registry compatible with async integration points as maybe elastic kibana app arch has other ideas or opinion of whatever pattern is preferable | 0 |
596,289 | 18,102,086,265 | IssuesEvent | 2021-09-22 15:08:41 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | opened | Enhance Graphic SVG Regex | problem: bug bug-type: broken use case priority: low type: corrective | Function `_isUrl` in `graphicsRecord.js` contains some regex that tries to validate if a string is a valid image, either a url to an image or a data url containing encoded treats.
The data url regex will currently not match some valid data urls, including our default north pole flag lol.
This regex covers more cases, and should replace the existing string.
```text
/^\s*data:([a-z]+\/[a-z0-9\-\+]+(;[a-z\-]+\=[a-z0-9\-]+)?)?(;base64)?,[a-z0-9\!\$\&\'\,\(\)\*\+\,\;\=\-\.\_\~\:\@\/\?\%\s]*\s*$/i
``` | 1.0 | Enhance Graphic SVG Regex - Function `_isUrl` in `graphicsRecord.js` contains some regex that tries to validate if a string is a valid image, either a url to an image or a data url containing encoded treats.
The data url regex will currently not match some valid data urls, including our default north pole flag lol.
This regex covers more cases, and should replace the existing string.
```text
/^\s*data:([a-z]+\/[a-z0-9\-\+]+(;[a-z\-]+\=[a-z0-9\-]+)?)?(;base64)?,[a-z0-9\!\$\&\'\,\(\)\*\+\,\;\=\-\.\_\~\:\@\/\?\%\s]*\s*$/i
``` | priority | enhance graphic svg regex function isurl in graphicsrecord js contains some regex that tries to validate if a string is a valid image either a url to an image or a data url containing encoded treats the data url regex will currently not match some valid data urls including our default north pole flag lol this regex covers more cases and should replace the existing string text s data s i | 1 |
300,256 | 25,954,948,198 | IssuesEvent | 2022-12-18 04:32:15 | nuvolaris/nuvolaris | https://api.github.com/repos/nuvolaris/nuvolaris | closed | An script to setup an OKD environment in AWS | nuvolaris-testing | Contribute an ansible script to the [nuvolaris-testing](https://github.com/nuvolaris/nuvolaris-testing) repo that builds in a test server accessible with ssh an OKD Kubernetes cluster to be usable as a test environment. It should export the kubernetes cluster access port and make available the key in /etc/kube/config, so with an `scp server:/etc/kube/config ~/.kube/config` is possible to retrieve it and access the cluster with `kubectl`
┆Issue is synchronized with this [Jira Task](https://nuvolaris.atlassian.net/browse/MVP-18) by [Unito](https://www.unito.io)
| 1.0 | An script to setup an OKD environment in AWS - Contribute an ansible script to the [nuvolaris-testing](https://github.com/nuvolaris/nuvolaris-testing) repo that builds in a test server accessible with ssh an OKD Kubernetes cluster to be usable as a test environment. It should export the kubernetes cluster access port and make available the key in /etc/kube/config, so with an `scp server:/etc/kube/config ~/.kube/config` is possible to retrieve it and access the cluster with `kubectl`
┆Issue is synchronized with this [Jira Task](https://nuvolaris.atlassian.net/browse/MVP-18) by [Unito](https://www.unito.io)
| non_priority | an script to setup an okd environment in aws contribute an ansible script to the repo that builds in a test server accessible with ssh an okd kubernetes cluster to be usable as a test environment it should export the kubernetes cluster access port and make available the key in etc kube config so with an scp server etc kube config kube config is possible to retrieve it and access the cluster with kubectl ┆issue is synchronized with this by | 0 |
767,093 | 26,910,601,122 | IssuesEvent | 2023-02-06 23:15:01 | projectLEMDO/lemdoIssues | https://api.github.com/repos/projectLEMDO/lemdoIssues | closed | Need to constrain image height in PDF rendering | bug Priority: urgent fix committed PDF rendering | Building the latest PDF has resulted in a lot of warnings like this:
`LaTeX Warning: Float too large for page`
and it's mostly because while I'm setting the width for images to 0.9 of textwidth, I'm not setting a maximum height. This provides a useful guide:
https://infolatex.com/set-a-maximum-width-and-height-for-an-image-in-latex/
but of course I'll have to implement it carefully and look at the results for other images. Ideally we want a maximum height that allows for the caption, which is a bit tricky and might require doing some measurement of the length of captions at build time. For the moment, I'm OKing the warning in our build process so that the unstable build doesn't confuse the team, but this must be fixed. | 1.0 | Need to constrain image height in PDF rendering - Building the latest PDF has resulted in a lot of warnings like this:
`LaTeX Warning: Float too large for page`
and it's mostly because while I'm setting the width for images to 0.9 of textwidth, I'm not setting a maximum height. This provides a useful guide:
https://infolatex.com/set-a-maximum-width-and-height-for-an-image-in-latex/
but of course I'll have to implement it carefully and look at the results for other images. Ideally we want a maximum height that allows for the caption, which is a bit tricky and might require doing some measurement of the length of captions at build time. For the moment, I'm OKing the warning in our build process so that the unstable build doesn't confuse the team, but this must be fixed. | priority | need to constrain image height in pdf rendering building the latest pdf has resulted in a lot of warnings like this latex warning float too large for page and it s mostly because while i m setting the width for images to of textwidth i m not setting a maximum height this provides a useful guide but of course i ll have to implement it carefully and look at the results for other images ideally we want a maximum height that allows for the caption which is a bit tricky and might require doing some measurement of the length of captions at build time for the moment i m oking the warning in our build process so that the unstable build doesn t confuse the team but this must be fixed | 1 |
641,988 | 20,864,176,743 | IssuesEvent | 2022-03-22 04:18:26 | angelewilliams/flashcards-starter | https://api.github.com/repos/angelewilliams/flashcards-starter | opened | Turn Class and Turn Test | iteration2 high priority | Should be:
Instantiated with two arguments - a string (that represents a user’s guess to the question), and a Card object for the current card in play.
returnGuess: method that returns the guess
returnCard: method that returns the Card
evaluateGuess: method that returns a boolean indicating if the user’s guess matches the correct answer on the card
giveFeedback - method that returns either ‘incorrect!’ or ‘correct!’ based on whether the guess is correct or not. | 1.0 | Turn Class and Turn Test - Should be:
Instantiated with two arguments - a string (that represents a user’s guess to the question), and a Card object for the current card in play.
returnGuess: method that returns the guess
returnCard: method that returns the Card
evaluateGuess: method that returns a boolean indicating if the user’s guess matches the correct answer on the card
giveFeedback - method that returns either ‘incorrect!’ or ‘correct!’ based on whether the guess is correct or not. | priority | turn class and turn test should be instantiated with two arguments a string that represents a user’s guess to the question and a card object for the current card in play returnguess method that returns the guess returncard method that returns the card evaluateguess method that returns a boolean indicating if the user’s guess matches the correct answer on the card givefeedback method that returns either ‘incorrect ’ or ‘correct ’ based on whether the guess is correct or not | 1 |
631,098 | 20,143,986,936 | IssuesEvent | 2022-02-09 04:20:12 | brave/brave-ios | https://api.github.com/repos/brave/brave-ios | closed | [feature/webauthn] Touch key modal does not show up if the key is already plugged in | security priority/P4 sec-low QA/No release-notes/exclude Epic: WebAuthn | <!-- Have you searched for similar issues on the repository?
Before submitting this issue, please visit our wiki for common ones: https://github.com/brave/browser-ios/wiki
For more, check out our community site: https://community.brave.com/ -->
### Description:
Touch key modal should show up if the lightning key is inserted.
### Steps to Reproduce
1. Navigate to https://mdp.github.io/u2fdemo
2. Insert the lightning key
3. Register and Sign
**Actual result:** <!-- Add screenshots if needed -->
On sign the modal is not presented if the lightning key is plugged in
**Expected result:**
Touch key modal should be displayed
**Reproduces how often:** [Easily reproduced, Intermittent Issue]
Easily
**Brave Version:** <!-- Provide full details Eg: v1.4.2(17.09.08.16) -->
1.20 (20.09.11.20)
**Device details:** <!-- Model type and iOS version Eg: iPhone 6s+ (iOS 10.3.3)-->
iPhone 11 + iOS 14 | 1.0 | [feature/webauthn] Touch key modal does not show up if the key is already plugged in - <!-- Have you searched for similar issues on the repository?
Before submitting this issue, please visit our wiki for common ones: https://github.com/brave/browser-ios/wiki
For more, check out our community site: https://community.brave.com/ -->
### Description:
Touch key modal should show up if the lightning key is inserted.
### Steps to Reproduce
1. Navigate to https://mdp.github.io/u2fdemo
2. Insert the lightning key
3. Register and Sign
**Actual result:** <!-- Add screenshots if needed -->
On sign the modal is not presented if the lightning key is plugged in
**Expected result:**
Touch key modal should be displayed
**Reproduces how often:** [Easily reproduced, Intermittent Issue]
Easily
**Brave Version:** <!-- Provide full details Eg: v1.4.2(17.09.08.16) -->
1.20 (20.09.11.20)
**Device details:** <!-- Model type and iOS version Eg: iPhone 6s+ (iOS 10.3.3)-->
iPhone 11 + iOS 14 | priority | touch key modal does not show up if the key is already plugged in have you searched for similar issues on the repository before submitting this issue please visit our wiki for common ones for more check out our community site description touch key modal should show up if the lightning key is inserted steps to reproduce navigate to insert the lightning key register and sign actual result on sign the modal is not presented if the lightning key is plugged in expected result touch key modal should be displayed reproduces how often easily brave version device details iphone ios | 1 |
530,786 | 15,436,692,368 | IssuesEvent | 2021-03-07 14:01:34 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.4 Develop-203]Missing marker on map when editing Deed or marking district on Zoning office | Category: Gameplay Priority: High Regression Type: Bug | Build: 0.9.4 Develop-203
### Issues
1. When editing Deed or marking district area via Zoning Office, the map UI opens but the marker is missing. The market where you either mark an area for the selected district and increase or decrease the area of a deed.
2. When trying to claim via Deed edit, pressing and holding the RMB while moving the mouse pointer gives the player a corresponding number of Claim Papers after clicking submit for free.
Repro:
1. Claim a property.
2. Edit the Deed.
3. On the map, press and hold RMB then move the mouse pointer.
4. Click submit.
5. Observe Claim Paper count.
Video Repro:
https://images.zenhubusercontent.com/42199084/209fc4b8-5f2c-450d-8085-e51f9e6ee2ac/2021_03_07_21_55_trim.mp4
Zoning Office
https://images.zenhubusercontent.com/42199084/0a4ed674-db3d-40a2-82dd-456e86282cf2/2021_03_07_21_37_trim__4_.mp4 | 1.0 | [0.9.4 Develop-203]Missing marker on map when editing Deed or marking district on Zoning office - Build: 0.9.4 Develop-203
### Issues
1. When editing Deed or marking district area via Zoning Office, the map UI opens but the marker is missing. The market where you either mark an area for the selected district and increase or decrease the area of a deed.
2. When trying to claim via Deed edit, pressing and holding the RMB while moving the mouse pointer gives the player a corresponding number of Claim Papers after clicking submit for free.
Repro:
1. Claim a property.
2. Edit the Deed.
3. On the map, press and hold RMB then move the mouse pointer.
4. Click submit.
5. Observe Claim Paper count.
Video Repro:
https://images.zenhubusercontent.com/42199084/209fc4b8-5f2c-450d-8085-e51f9e6ee2ac/2021_03_07_21_55_trim.mp4
Zoning Office
https://images.zenhubusercontent.com/42199084/0a4ed674-db3d-40a2-82dd-456e86282cf2/2021_03_07_21_37_trim__4_.mp4 | priority | missing marker on map when editing deed or marking district on zoning office build develop issues when editing deed or marking district area via zoning office the map ui opens but the marker is missing the market where you either mark an area for the selected district and increase or decrease the area of a deed when trying to claim via deed edit pressing and holding the rmb while moving the mouse pointer gives the player a corresponding number of claim papers after clicking submit for free repro claim a property edit the deed on the map press and hold rmb then move the mouse pointer click submit observe claim paper count video repro zoning office | 1 |
214,205 | 7,267,780,871 | IssuesEvent | 2018-02-20 07:19:06 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Need a meaningful error message when the user denies to share data | Affected/5.5.0-Alpha Priority/High Severity/Major Type/Bug | Need a meaningful error message when the user denies to share data. The message should mention the reason for the failure; that you cannot proceed as you cannot share the information.

| 1.0 | Need a meaningful error message when the user denies to share data - Need a meaningful error message when the user denies to share data. The message should mention the reason for the failure; that you cannot proceed as you cannot share the information.

| priority | need a meaningful error message when the user denies to share data need a meaningful error message when the user denies to share data the message should mention the reason for the failure that you cannot proceed as you cannot share the information | 1 |
25,659 | 12,702,761,909 | IssuesEvent | 2020-06-22 20:49:38 | openstreetmap/iD | https://api.github.com/repos/openstreetmap/iD | closed | iD performance issue with long lines | performance | Hello,
after a short discussion in the OSM Discord group I was referred here to address my problem.
I am experiencing an issue that affects the performance of iD (Windows 10, Google Chrome). When I want to continute drawing a line that already has ~200+ nodes in it, each new node I draw takes about 2-3 seconds to appear.
This issue only occurs in the specific region I am mainly mapping. Other OSM users confirmed this slow behaviour when they tried to continue drawing lines in this region.
One example would be this: https://www.openstreetmap.org/way/799622518
Attempting to continue drawing this exact path eastwards will result in a performance drop, taking each newly placed node a few seconds to appear.
I already tried different computers (including high-end) and browsers. However, the fact that also other users experience the performance issue when trying to edit this line rather points at an iD related issue. When I last mapped long lines (some months back), this has never been an issue. Hence I suspect that it might have come with one of the recent updates.
I would appreciate this being looked into!
Thanks,
ian | True | iD performance issue with long lines - Hello,
after a short discussion in the OSM Discord group I was referred here to address my problem.
I am experiencing an issue that affects the performance of iD (Windows 10, Google Chrome). When I want to continute drawing a line that already has ~200+ nodes in it, each new node I draw takes about 2-3 seconds to appear.
This issue only occurs in the specific region I am mainly mapping. Other OSM users confirmed this slow behaviour when they tried to continue drawing lines in this region.
One example would be this: https://www.openstreetmap.org/way/799622518
Attempting to continue drawing this exact path eastwards will result in a performance drop, taking each newly placed node a few seconds to appear.
I already tried different computers (including high-end) and browsers. However, the fact that also other users experience the performance issue when trying to edit this line rather points at an iD related issue. When I last mapped long lines (some months back), this has never been an issue. Hence I suspect that it might have come with one of the recent updates.
I would appreciate this being looked into!
Thanks,
ian | non_priority | id performance issue with long lines hello after a short discussion in the osm discord group i was referred here to address my problem i am experiencing an issue that affects the performance of id windows google chrome when i want to continute drawing a line that already has nodes in it each new node i draw takes about seconds to appear this issue only occurs in the specific region i am mainly mapping other osm users confirmed this slow behaviour when they tried to continue drawing lines in this region one example would be this attempting to continue drawing this exact path eastwards will result in a performance drop taking each newly placed node a few seconds to appear i already tried different computers including high end and browsers however the fact that also other users experience the performance issue when trying to edit this line rather points at an id related issue when i last mapped long lines some months back this has never been an issue hence i suspect that it might have come with one of the recent updates i would appreciate this being looked into thanks ian | 0 |
130,023 | 5,108,190,749 | IssuesEvent | 2017-01-05 16:58:42 | semihalf-berestovskyy-andriy/test2 | https://api.github.com/repos/semihalf-berestovskyy-andriy/test2 | closed | can not show fastpath and get some error | docs & examples low priority | Note: the issue was imported automatically from Bugzilla with bugzilla2issues.py tool
# Bugzilla Bug ID: 95
Date: 2016-09-20 08:32:08 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
To: Sorin Vultureanu <sorin.vultureanu@enea.com>
CC: xiaowen.1.yang@gmail.com
Last updated: 2016-09-30 11:17:38 +0200
## Bugzilla Comment ID: 179
Date: 2016-09-20 08:32:08 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
I builded ODP and OFP on my ubuntu VM , then running ofp example fpm -i ens33 output as below
17 udp RW Node (UDP)
1 checksum RW int (compute udp checksum)
3 maxdgram RW u_long (Maximum outgoing UDP datagram size)
4 recvspace RW u_long (Maximum space for incoming UDP datagrams)
256 blackhole RW int (Do not send port unreachables for refused connects)
257 log_in_vain RW int (Log all incoming UDP packets)
255 raw RW Node (RAW)
I 0 0:407821056 ofp_uma.c:44] Creating pool 'udp_inpcb', nitems=1024 size=904 total=925696
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcp_inpcb', nitems=2048 size=904 total=1851392
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcpcb', nitems=2048 size=784 total=1605632
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcptw', nitems=409 size=80 total=32720
I 0 0:407821056 ofp_uma.c:44] Creating pool 'syncache', nitems=30720 size=168 total=5160960
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcpreass', nitems=320 size=48 total=15360
I 0 0:407821056 ofp_uma.c:44] Creating pool 'sackhole', nitems=65536 size=40 total=2621440
I 0 0:407821056 ofp_init.c:191] Slow path threads on core 0
I 0 0:407821056 ofp_ifnet.c:143] Device 'ens33' addr 00:0c:29:2a:f5:fd
I 0 0:407821056 ofp_ifnet.c:152] Device 'ens33' MTU=1500
I 0 0:238683904 ofp_cli.c:1599] CLI server started on core 0
but can not see the fastpath when telnet CLI
> show ifconfig
fp0 (7) (ens33) slowpath: on
Link encap:Ethernet HWaddr: 00:0c:29:2a:f5:fd
inet6 addr: fe80:0000:0000:0000:020c:29ff:fe2a:f5fd Scope:Link
MTU: 1500
fp1 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:01:73:e6:8f:57
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp2 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:02:b3:ad:06:d7
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp3 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:03:45:24:e9:49
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp4 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:04:f6:14:c2:13
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp5 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:05:ec:e1:dd:14
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp6 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:06:90:0f:f1:70
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp7 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:07:1a:2a:a5:c1
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
gre
Link not configured
vxlan
Link not configured
lo
Link not configured
root@ubuntu:~# ifconfig
ens33 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:fd
inet addr:192.168.75.128 Bcast:192.168.75.255 Mask:255.255.255.0
inet6 addr: fe80::7e20:f368:9e00:2399/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:5918 errors:0 dropped:0 overruns:0 frame:0
TX packets:7768 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:492520 (492.5 KB) TX bytes:1260645 (1.2 MB)
ens37 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:07
inet addr:192.168.75.130 Bcast:192.168.75.255 Mask:255.255.255.0
inet6 addr: fe80::ccf4:29ec:680:e471/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:208 errors:0 dropped:0 overruns:0 frame:0
TX packets:60 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:26339 (26.3 KB) TX bytes:7523 (7.5 KB)
fp0 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:fd
inet6 addr: fe80::20c:29ff:fe2a:f5fd/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:68 errors:0 dropped:0 overruns:0 frame:0
TX packets:24 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:5819 (5.8 KB) TX bytes:2940 (2.9 KB)
lo Link encap:Local Loopback
inet addr:127.0.0.1 Mask:255.0.0.0
inet6 addr: ::1/128 Scope:Host
UP LOOPBACK RUNNING MTU:65536 Metric:1
RX packets:982 errors:0 dropped:0 overruns:0 frame:0
TX packets:982 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1
RX bytes:67774 (67.7 KB) TX bytes:67774 (67.7 KB)
## Bugzilla Comment ID: 180
Date: 2016-09-20 08:41:51 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
there are some error information as below:
E 237 0:335931136 ofp_portconf.c:1126] port:65535 is outside the valid interval
E 237 0:335931136 ofp_netlink.c:426] ! Interface index 2 not found
## Bugzilla Comment ID: 181
Date: 2016-09-21 08:08:18 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
--==--==--==--==--==--==--
-- WELCOME to OFP CLI --
--==--==--==--==--==--==--
>
>
> show ifconfig
fp0 (3) (ens33) slowpath: on
Link encap:Ethernet HWaddr: 00:0c:29:44:55:6b
inet6 addr: fe80:0000:0000:0000:020c:29ff:fe44:556b Scope:Link
MTU: 1500
fp1 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:01:2a:9d:c9:50
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp2 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:02:4e:d6:1b:c8
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp3 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:03:28:b6:39:de
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp4 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:04:10:79:30:81
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp5 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:05:36:85:35:96
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp6 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:06:28:a4:59:03
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp7 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:07:3a:59:00:bf
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
gre
Link not configured
vxlan
Link not configured
lo
Link not configured
> show stat
Settings:
compute latency - no
compute performance - no
Packets:
Core: 0 1
ODP to FP: 0 56
FP to ODP: 0 0
FP to SP: 56 0
SP to ODP: 22 0
Tx frag: 0 0
Rx IP frag: 0 0
RX IP reass: 0 0
Allocated memory:
avl tree alloc now=12 max=12 total=16384
rt tree alloc now=0 max=0 total=128
rt6 tree alloc now=1 max=1 total=65536
## Bugzilla Comment ID: 184
Date: 2016-09-30 11:17:38 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
it is a configuration issue,fixed.
| 1.0 | can not show fastpath and get some error - Note: the issue was imported automatically from Bugzilla with bugzilla2issues.py tool
# Bugzilla Bug ID: 95
Date: 2016-09-20 08:32:08 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
To: Sorin Vultureanu <sorin.vultureanu@enea.com>
CC: xiaowen.1.yang@gmail.com
Last updated: 2016-09-30 11:17:38 +0200
## Bugzilla Comment ID: 179
Date: 2016-09-20 08:32:08 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
I builded ODP and OFP on my ubuntu VM , then running ofp example fpm -i ens33 output as below
17 udp RW Node (UDP)
1 checksum RW int (compute udp checksum)
3 maxdgram RW u_long (Maximum outgoing UDP datagram size)
4 recvspace RW u_long (Maximum space for incoming UDP datagrams)
256 blackhole RW int (Do not send port unreachables for refused connects)
257 log_in_vain RW int (Log all incoming UDP packets)
255 raw RW Node (RAW)
I 0 0:407821056 ofp_uma.c:44] Creating pool 'udp_inpcb', nitems=1024 size=904 total=925696
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcp_inpcb', nitems=2048 size=904 total=1851392
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcpcb', nitems=2048 size=784 total=1605632
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcptw', nitems=409 size=80 total=32720
I 0 0:407821056 ofp_uma.c:44] Creating pool 'syncache', nitems=30720 size=168 total=5160960
I 0 0:407821056 ofp_uma.c:44] Creating pool 'tcpreass', nitems=320 size=48 total=15360
I 0 0:407821056 ofp_uma.c:44] Creating pool 'sackhole', nitems=65536 size=40 total=2621440
I 0 0:407821056 ofp_init.c:191] Slow path threads on core 0
I 0 0:407821056 ofp_ifnet.c:143] Device 'ens33' addr 00:0c:29:2a:f5:fd
I 0 0:407821056 ofp_ifnet.c:152] Device 'ens33' MTU=1500
I 0 0:238683904 ofp_cli.c:1599] CLI server started on core 0
but can not see the fastpath when telnet CLI
> show ifconfig
fp0 (7) (ens33) slowpath: on
Link encap:Ethernet HWaddr: 00:0c:29:2a:f5:fd
inet6 addr: fe80:0000:0000:0000:020c:29ff:fe2a:f5fd Scope:Link
MTU: 1500
fp1 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:01:73:e6:8f:57
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp2 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:02:b3:ad:06:d7
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp3 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:03:45:24:e9:49
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp4 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:04:f6:14:c2:13
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp5 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:05:ec:e1:dd:14
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp6 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:06:90:0f:f1:70
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp7 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:07:1a:2a:a5:c1
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
gre
Link not configured
vxlan
Link not configured
lo
Link not configured
root@ubuntu:~# ifconfig
ens33 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:fd
inet addr:192.168.75.128 Bcast:192.168.75.255 Mask:255.255.255.0
inet6 addr: fe80::7e20:f368:9e00:2399/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:5918 errors:0 dropped:0 overruns:0 frame:0
TX packets:7768 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:492520 (492.5 KB) TX bytes:1260645 (1.2 MB)
ens37 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:07
inet addr:192.168.75.130 Bcast:192.168.75.255 Mask:255.255.255.0
inet6 addr: fe80::ccf4:29ec:680:e471/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:208 errors:0 dropped:0 overruns:0 frame:0
TX packets:60 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:26339 (26.3 KB) TX bytes:7523 (7.5 KB)
fp0 Link encap:Ethernet HWaddr 00:0c:29:2a:f5:fd
inet6 addr: fe80::20c:29ff:fe2a:f5fd/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:68 errors:0 dropped:0 overruns:0 frame:0
TX packets:24 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:5819 (5.8 KB) TX bytes:2940 (2.9 KB)
lo Link encap:Local Loopback
inet addr:127.0.0.1 Mask:255.0.0.0
inet6 addr: ::1/128 Scope:Host
UP LOOPBACK RUNNING MTU:65536 Metric:1
RX packets:982 errors:0 dropped:0 overruns:0 frame:0
TX packets:982 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1
RX bytes:67774 (67.7 KB) TX bytes:67774 (67.7 KB)
## Bugzilla Comment ID: 180
Date: 2016-09-20 08:41:51 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
there are some error information as below:
E 237 0:335931136 ofp_portconf.c:1126] port:65535 is outside the valid interval
E 237 0:335931136 ofp_netlink.c:426] ! Interface index 2 not found
## Bugzilla Comment ID: 181
Date: 2016-09-21 08:08:18 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
--==--==--==--==--==--==--
-- WELCOME to OFP CLI --
--==--==--==--==--==--==--
>
>
> show ifconfig
fp0 (3) (ens33) slowpath: on
Link encap:Ethernet HWaddr: 00:0c:29:44:55:6b
inet6 addr: fe80:0000:0000:0000:020c:29ff:fe44:556b Scope:Link
MTU: 1500
fp1 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:01:2a:9d:c9:50
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp2 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:02:4e:d6:1b:c8
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp3 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:03:28:b6:39:de
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp4 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:04:10:79:30:81
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp5 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:05:36:85:35:96
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp6 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:06:28:a4:59:03
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
fp7 (0) () slowpath: off
Link encap:Ethernet HWaddr: 02:07:3a:59:00:bf
inet6 addr: 0000:0000:0000:0000:0000:0000:0000:0000 Scope:Link
MTU: 1500
gre
Link not configured
vxlan
Link not configured
lo
Link not configured
> show stat
Settings:
compute latency - no
compute performance - no
Packets:
Core: 0 1
ODP to FP: 0 56
FP to ODP: 0 0
FP to SP: 56 0
SP to ODP: 22 0
Tx frag: 0 0
Rx IP frag: 0 0
RX IP reass: 0 0
Allocated memory:
avl tree alloc now=12 max=12 total=16384
rt tree alloc now=0 max=0 total=128
rt6 tree alloc now=1 max=1 total=65536
## Bugzilla Comment ID: 184
Date: 2016-09-30 11:17:38 +0200
From: xiaowen yang <xiaowen.1.yang@gmail.com>
it is a configuration issue,fixed.
| priority | can not show fastpath and get some error note the issue was imported automatically from bugzilla with py tool bugzilla bug id date from xiaowen yang to sorin vultureanu cc xiaowen yang gmail com last updated bugzilla comment id date from xiaowen yang i builded odp and ofp on my ubuntu vm then running ofp example fpm i output as below udp rw node udp checksum rw int compute udp checksum maxdgram rw u long maximum outgoing udp datagram size recvspace rw u long maximum space for incoming udp datagrams blackhole rw int do not send port unreachables for refused connects log in vain rw int log all incoming udp packets raw rw node raw i ofp uma c creating pool udp inpcb nitems size total i ofp uma c creating pool tcp inpcb nitems size total i ofp uma c creating pool tcpcb nitems size total i ofp uma c creating pool tcptw nitems size total i ofp uma c creating pool syncache nitems size total i ofp uma c creating pool tcpreass nitems size total i ofp uma c creating pool sackhole nitems size total i ofp init c slow path threads on core i ofp ifnet c device addr fd i ofp ifnet c device mtu i ofp cli c cli server started on core but can not see the fastpath when telnet cli show ifconfig slowpath on link encap ethernet hwaddr fd addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr ad addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr ec dd addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu gre link not configured vxlan link not configured lo link not configured root ubuntu ifconfig link encap ethernet hwaddr fd inet addr bcast mask addr scope link up broadcast running multicast mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes kb tx bytes mb link encap ethernet hwaddr inet addr bcast mask addr scope link up broadcast running multicast mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes kb tx bytes kb link encap ethernet hwaddr fd addr scope link up broadcast running multicast mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes kb tx bytes kb lo link encap local loopback inet addr mask addr scope host up loopback running mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes kb tx bytes kb bugzilla comment id date from xiaowen yang there are some error information as below e ofp portconf c port is outside the valid interval e ofp netlink c interface index not found bugzilla comment id date from xiaowen yang welcome to ofp cli show ifconfig slowpath on link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr de addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr addr scope link mtu slowpath off link encap ethernet hwaddr bf addr scope link mtu gre link not configured vxlan link not configured lo link not configured show stat settings compute latency no compute performance no packets core odp to fp fp to odp fp to sp sp to odp tx frag rx ip frag rx ip reass allocated memory avl tree alloc now max total rt tree alloc now max total tree alloc now max total bugzilla comment id date from xiaowen yang it is a configuration issue fixed | 1 |
791,506 | 27,865,782,880 | IssuesEvent | 2023-03-21 10:09:19 | Snapmaker/Luban | https://api.github.com/repos/Snapmaker/Luban | closed | Bug: Show estimated remaining time when running Gcode directly from Luban | Type: Bug/Bug Fix Priority: Medium | At the moment this is what you see when starting the Gcode while connected to the machine directly from Luban:

It would be nice to get an estimated time for finish time and remaining time, this could be a simple calculation based on received lines vs. total lines and elapsed time. | 1.0 | Bug: Show estimated remaining time when running Gcode directly from Luban - At the moment this is what you see when starting the Gcode while connected to the machine directly from Luban:

It would be nice to get an estimated time for finish time and remaining time, this could be a simple calculation based on received lines vs. total lines and elapsed time. | priority | bug show estimated remaining time when running gcode directly from luban at the moment this is what you see when starting the gcode while connected to the machine directly from luban it would be nice to get an estimated time for finish time and remaining time this could be a simple calculation based on received lines vs total lines and elapsed time | 1 |
153,254 | 5,887,988,248 | IssuesEvent | 2017-05-17 09:00:12 | MatthijsKok/TI2806-Contextproject | https://api.github.com/repos/MatthijsKok/TI2806-Contextproject | closed | API key Research | 5h priority:high ready story | Find out if 1 API key can be used for multiple IP addresses. As we cannot estimate the implementation there is no way of telling if it can be implemented immediately
| 1.0 | API key Research - Find out if 1 API key can be used for multiple IP addresses. As we cannot estimate the implementation there is no way of telling if it can be implemented immediately
| priority | api key research find out if api key can be used for multiple ip addresses as we cannot estimate the implementation there is no way of telling if it can be implemented immediately | 1 |
30,603 | 7,236,338,346 | IssuesEvent | 2018-02-13 06:25:43 | Microsoft/vscode-python | https://api.github.com/repos/Microsoft/vscode-python | closed | Add an option to exclude certain kinds of `ctags` | awaiting 2-PR feature-code navigation good first issue type-enhancement | This is a feature request.
Currently, `ctags` generates all kinds of tags for Python code: class, function, class member, variable, import, file, etc.
When doing "Go To Symbol in Workspace" I'm only interested in definitions, such as class and function definitions and class members.
Allowing to specify what kinds of tags should be generated by `ctags` would increase the relevancy of the "Go To Symbol in Workspace" output, and also reduce the size of `tags` file drasticly.
The proposed solution is to add an option allowing to specify the included/excluded kinds of tags. This option will be passed to the `--Python-kinds` argument of the `ctags` executable. | 1.0 | Add an option to exclude certain kinds of `ctags` - This is a feature request.
Currently, `ctags` generates all kinds of tags for Python code: class, function, class member, variable, import, file, etc.
When doing "Go To Symbol in Workspace" I'm only interested in definitions, such as class and function definitions and class members.
Allowing to specify what kinds of tags should be generated by `ctags` would increase the relevancy of the "Go To Symbol in Workspace" output, and also reduce the size of `tags` file drasticly.
The proposed solution is to add an option allowing to specify the included/excluded kinds of tags. This option will be passed to the `--Python-kinds` argument of the `ctags` executable. | non_priority | add an option to exclude certain kinds of ctags this is a feature request currently ctags generates all kinds of tags for python code class function class member variable import file etc when doing go to symbol in workspace i m only interested in definitions such as class and function definitions and class members allowing to specify what kinds of tags should be generated by ctags would increase the relevancy of the go to symbol in workspace output and also reduce the size of tags file drasticly the proposed solution is to add an option allowing to specify the included excluded kinds of tags this option will be passed to the python kinds argument of the ctags executable | 0 |
225,655 | 24,881,071,391 | IssuesEvent | 2022-10-28 01:10:24 | Nivaskumark/kernel_v4.1.15 | https://api.github.com/repos/Nivaskumark/kernel_v4.1.15 | opened | CVE-2022-3646 (Medium) detected in linuxlinux-4.6 | security vulnerability | ## CVE-2022-3646 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nilfs2/segment.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nilfs2/segment.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability, which was classified as problematic, has been found in Linux Kernel. This issue affects the function nilfs_attach_log_writer of the file fs/nilfs2/segment.c of the component BPF. The manipulation leads to memory leak. The attack may be initiated remotely. It is recommended to apply a patch to fix this issue. The identifier VDB-211961 was assigned to this vulnerability.
<p>Publish Date: 2022-10-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3646>CVE-2022-3646</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3646">https://www.linuxkernelcves.com/cves/CVE-2022-3646</a></p>
<p>Release Date: 2022-10-21</p>
<p>Fix Resolution: v5.4.218,v5.10.148,v5.15.74,v6.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-3646 (Medium) detected in linuxlinux-4.6 - ## CVE-2022-3646 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nilfs2/segment.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nilfs2/segment.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability, which was classified as problematic, has been found in Linux Kernel. This issue affects the function nilfs_attach_log_writer of the file fs/nilfs2/segment.c of the component BPF. The manipulation leads to memory leak. The attack may be initiated remotely. It is recommended to apply a patch to fix this issue. The identifier VDB-211961 was assigned to this vulnerability.
<p>Publish Date: 2022-10-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3646>CVE-2022-3646</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3646">https://www.linuxkernelcves.com/cves/CVE-2022-3646</a></p>
<p>Release Date: 2022-10-21</p>
<p>Fix Resolution: v5.4.218,v5.10.148,v5.15.74,v6.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files fs segment c fs segment c vulnerability details a vulnerability which was classified as problematic has been found in linux kernel this issue affects the function nilfs attach log writer of the file fs segment c of the component bpf the manipulation leads to memory leak the attack may be initiated remotely it is recommended to apply a patch to fix this issue the identifier vdb was assigned to this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
135,529 | 5,253,992,919 | IssuesEvent | 2017-02-02 11:23:40 | hpi-swt2/workshop-portal | https://api.github.com/repos/hpi-swt2/workshop-portal | closed | Fix Fonts | High Priority needs acceptance team-hendrik | **Acceptance Criteria**
- [ ] make sure that all fonts are loaded correctly, especially "Roboto Slab" and "Open Sans"
- [ ] fix font loading via Google Fonts
- [ ] if not possible: download fonts and serve them from our own server | 1.0 | Fix Fonts - **Acceptance Criteria**
- [ ] make sure that all fonts are loaded correctly, especially "Roboto Slab" and "Open Sans"
- [ ] fix font loading via Google Fonts
- [ ] if not possible: download fonts and serve them from our own server | priority | fix fonts acceptance criteria make sure that all fonts are loaded correctly especially roboto slab and open sans fix font loading via google fonts if not possible download fonts and serve them from our own server | 1 |
386,390 | 11,437,931,354 | IssuesEvent | 2020-02-05 01:36:38 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | Dispensary state not cleared when navigating away | Bug: development Docs: not needed Effort: small Module: dispensary Priority: high | ## Describe the bug
When navigating away from the dispensary page, the redux state is not reset
### To reproduce
Dispensary development bug
### Expected behaviour
Dispensary development bug
### Proposed Solution
Dispensary development bug
### Version and device info
Dispensary development bug
### Additional context
Dispensary development bug
| 1.0 | Dispensary state not cleared when navigating away - ## Describe the bug
When navigating away from the dispensary page, the redux state is not reset
### To reproduce
Dispensary development bug
### Expected behaviour
Dispensary development bug
### Proposed Solution
Dispensary development bug
### Version and device info
Dispensary development bug
### Additional context
Dispensary development bug
| priority | dispensary state not cleared when navigating away describe the bug when navigating away from the dispensary page the redux state is not reset to reproduce dispensary development bug expected behaviour dispensary development bug proposed solution dispensary development bug version and device info dispensary development bug additional context dispensary development bug | 1 |
16,528 | 6,220,578,719 | IssuesEvent | 2017-07-10 00:10:55 | atom/atom | https://api.github.com/repos/atom/atom | closed | chmod build for deb or build in place instead of /tmp | build-error linux stale | The assumption of your build scripts is that all people use the default insecure permissions provided by Ubuntu but this is not true, quite a few of us default all user `umask`'s to `0077` and use `getfacl` and `setfacl` to adjust per directory permissions inside of the home directory, this results in dpkg going nuts unless you adjust the `umask` ahead of time which is fair enough for programmers and systems experts but if you plan to build a one-stop shop for building atom please consider taking into consideration consistency enforcement to prevent silly `dpkg-deb` errors.
There are several options here in my opinion:
- Set the `umask` before you even get started.
- Build in place and use `setfacl` on a custom tmp folder inside of the build directory.
- `chmod` everything.
If you guys provide me with your preference I'd be more than happy to send a pull addressing this issue and providing quite a bit of system consistency enforcement for your Debian build scripts.
| 1.0 | chmod build for deb or build in place instead of /tmp - The assumption of your build scripts is that all people use the default insecure permissions provided by Ubuntu but this is not true, quite a few of us default all user `umask`'s to `0077` and use `getfacl` and `setfacl` to adjust per directory permissions inside of the home directory, this results in dpkg going nuts unless you adjust the `umask` ahead of time which is fair enough for programmers and systems experts but if you plan to build a one-stop shop for building atom please consider taking into consideration consistency enforcement to prevent silly `dpkg-deb` errors.
There are several options here in my opinion:
- Set the `umask` before you even get started.
- Build in place and use `setfacl` on a custom tmp folder inside of the build directory.
- `chmod` everything.
If you guys provide me with your preference I'd be more than happy to send a pull addressing this issue and providing quite a bit of system consistency enforcement for your Debian build scripts.
| non_priority | chmod build for deb or build in place instead of tmp the assumption of your build scripts is that all people use the default insecure permissions provided by ubuntu but this is not true quite a few of us default all user umask s to and use getfacl and setfacl to adjust per directory permissions inside of the home directory this results in dpkg going nuts unless you adjust the umask ahead of time which is fair enough for programmers and systems experts but if you plan to build a one stop shop for building atom please consider taking into consideration consistency enforcement to prevent silly dpkg deb errors there are several options here in my opinion set the umask before you even get started build in place and use setfacl on a custom tmp folder inside of the build directory chmod everything if you guys provide me with your preference i d be more than happy to send a pull addressing this issue and providing quite a bit of system consistency enforcement for your debian build scripts | 0 |
673,447 | 22,969,886,105 | IssuesEvent | 2022-07-20 01:21:55 | stackcollision/Nebulous-BugReporting | https://api.github.com/repos/stackcollision/Nebulous-BugReporting | reopened | Missiles with Semi-active RADAR primary seekers and other different forms of RADAR backup seekers still displays errors. | bug branch-modmis incorrect behavior priority high | **Describe the bug**
Missiles with Semi-active RADAR primary seekers and other different forms of RADAR backup seekers still displays errors.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Missile Designer
2. Create new hybrid missile (Cyclone, Atlatl) and put primary seeker as Semi-active RADAR and backup seeker as Active radar (FARS, SARS, SEARS)
3. Game still displays "WARN: Primary seeker provides direction only, staging will not trigger."
**Expected behavior**
Error does not pop up with backup active RADAR seeker.
**Additional context**
None.
**Attachments**


 | 1.0 | Missiles with Semi-active RADAR primary seekers and other different forms of RADAR backup seekers still displays errors. - **Describe the bug**
Missiles with Semi-active RADAR primary seekers and other different forms of RADAR backup seekers still displays errors.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Missile Designer
2. Create new hybrid missile (Cyclone, Atlatl) and put primary seeker as Semi-active RADAR and backup seeker as Active radar (FARS, SARS, SEARS)
3. Game still displays "WARN: Primary seeker provides direction only, staging will not trigger."
**Expected behavior**
Error does not pop up with backup active RADAR seeker.
**Additional context**
None.
**Attachments**


 | priority | missiles with semi active radar primary seekers and other different forms of radar backup seekers still displays errors describe the bug missiles with semi active radar primary seekers and other different forms of radar backup seekers still displays errors to reproduce steps to reproduce the behavior go to missile designer create new hybrid missile cyclone atlatl and put primary seeker as semi active radar and backup seeker as active radar fars sars sears game still displays warn primary seeker provides direction only staging will not trigger expected behavior error does not pop up with backup active radar seeker additional context none attachments | 1 |
125,634 | 16,822,421,870 | IssuesEvent | 2021-06-17 14:31:02 | ProjectMirador/mirador | https://api.github.com/repos/ProjectMirador/mirador | closed | Research and design a new two-up view that does not do book view calculations. | Mirador3 design needed user experience | It may be useful for Mirador users to have a two-up view within an object where the second canvas is not the back of the first canvas.
An example here is the scanning and representation of a booklike thing was scanned published in a non-IIIF `paged` compatible way. Or maybe there are two images next two each other that would be useful to shown together?
This ticket indicates that we need better use cases to support what this should be. Also design will need to go in supporting these cases. | 1.0 | Research and design a new two-up view that does not do book view calculations. - It may be useful for Mirador users to have a two-up view within an object where the second canvas is not the back of the first canvas.
An example here is the scanning and representation of a booklike thing was scanned published in a non-IIIF `paged` compatible way. Or maybe there are two images next two each other that would be useful to shown together?
This ticket indicates that we need better use cases to support what this should be. Also design will need to go in supporting these cases. | non_priority | research and design a new two up view that does not do book view calculations it may be useful for mirador users to have a two up view within an object where the second canvas is not the back of the first canvas an example here is the scanning and representation of a booklike thing was scanned published in a non iiif paged compatible way or maybe there are two images next two each other that would be useful to shown together this ticket indicates that we need better use cases to support what this should be also design will need to go in supporting these cases | 0 |
154,132 | 5,911,830,367 | IssuesEvent | 2017-05-21 01:16:37 | HaxeFoundation/intellij-haxe | https://api.github.com/repos/HaxeFoundation/intellij-haxe | closed | For in loop type inference only works for arrays | bug Priority 3 | Type inference works fine for the following because it's an array. Item will be recognized as a String and will give code completion for the "indexOf" method.
```
var array : Array<String> = ["item"];
for (item in array) {
var index = item.indexOf("e"); // works fine
}
```
However, on a map, the same thing does not work. The type of "key" is unknown, so there is no code completion, and calling "indexOf" gives an unresolved symbol warning.
```
var map : Map<String,String> = ["key" => "value"];
for (key in map.keys()) {
var index = key.indexOf("e"); // unresolved symbol warning
}
```
I am not sure if it works on other types of iterators, but it doesn't work on Map.keys() or Map.iterator(), so I assume it's not working for any others.
TIR: STB-7289
| 1.0 | For in loop type inference only works for arrays - Type inference works fine for the following because it's an array. Item will be recognized as a String and will give code completion for the "indexOf" method.
```
var array : Array<String> = ["item"];
for (item in array) {
var index = item.indexOf("e"); // works fine
}
```
However, on a map, the same thing does not work. The type of "key" is unknown, so there is no code completion, and calling "indexOf" gives an unresolved symbol warning.
```
var map : Map<String,String> = ["key" => "value"];
for (key in map.keys()) {
var index = key.indexOf("e"); // unresolved symbol warning
}
```
I am not sure if it works on other types of iterators, but it doesn't work on Map.keys() or Map.iterator(), so I assume it's not working for any others.
TIR: STB-7289
| priority | for in loop type inference only works for arrays type inference works fine for the following because it s an array item will be recognized as a string and will give code completion for the indexof method var array array for item in array var index item indexof e works fine however on a map the same thing does not work the type of key is unknown so there is no code completion and calling indexof gives an unresolved symbol warning var map map for key in map keys var index key indexof e unresolved symbol warning i am not sure if it works on other types of iterators but it doesn t work on map keys or map iterator so i assume it s not working for any others tir stb | 1 |
772,644 | 27,130,408,412 | IssuesEvent | 2023-02-16 09:22:48 | thebaselab/codeapp | https://api.github.com/repos/thebaselab/codeapp | closed | Feature Request: SSH Improvements | enhancement priority | This app would be AMAZING if it had better SSH/Remote connections support. If you could mimic VS Code's remote connections plugin this would be great.
Beyond that minor SSH improvements would be awesome, for example pressing tab to auto complete and adding arrow keys on the keyboard to go up/down. | 1.0 | Feature Request: SSH Improvements - This app would be AMAZING if it had better SSH/Remote connections support. If you could mimic VS Code's remote connections plugin this would be great.
Beyond that minor SSH improvements would be awesome, for example pressing tab to auto complete and adding arrow keys on the keyboard to go up/down. | priority | feature request ssh improvements this app would be amazing if it had better ssh remote connections support if you could mimic vs code s remote connections plugin this would be great beyond that minor ssh improvements would be awesome for example pressing tab to auto complete and adding arrow keys on the keyboard to go up down | 1 |
245,339 | 20,763,371,290 | IssuesEvent | 2022-03-15 18:12:18 | MRC-CSO-SPHSU/jamjam | https://api.github.com/repos/MRC-CSO-SPHSU/jamjam | opened | Add a test for absolute errors | component:tests | This should be a separate utility class with auxiliary functions only. | 1.0 | Add a test for absolute errors - This should be a separate utility class with auxiliary functions only. | non_priority | add a test for absolute errors this should be a separate utility class with auxiliary functions only | 0 |
262,768 | 19,830,882,299 | IssuesEvent | 2022-01-20 11:51:42 | ngneat/falso | https://api.github.com/repos/ngneat/falso | closed | Interactive Playground within the Docs | documentation enhancement | ### Description
https://dev.to/mrmuhammadali/live-code-editing-in-docusaurus-ux-at-its-best-2hj1
### Proposed solution
_No response_
### Alternatives considered
_No response_
### Do you want to create a pull request?
No | 1.0 | Interactive Playground within the Docs - ### Description
https://dev.to/mrmuhammadali/live-code-editing-in-docusaurus-ux-at-its-best-2hj1
### Proposed solution
_No response_
### Alternatives considered
_No response_
### Do you want to create a pull request?
No | non_priority | interactive playground within the docs description proposed solution no response alternatives considered no response do you want to create a pull request no | 0 |
534,356 | 15,615,084,045 | IssuesEvent | 2021-03-19 18:38:31 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | Using pre-existing deployments | area/deploy area/modules kind/feature-request priority/p1 | Hello!
Let's say I already have a shared Dev remote cluster running some deployments (say App1, App2 and App3) that all communicate with each other.
Is it possible to temporary "replace" the cluster's App2 by a local App2 using Skaffold?
I may have misunderstood the documentation but I'm afraid I'll have to re-deploy an entire, personal new set of deployments each time I need to do some development to one component of my cluster.
It seems like `remoteManifests`, as mentioned in #2160, could help me. But I'd like some confirmation and guidance as of how to achieve this.
Thanks in advance. | 1.0 | Using pre-existing deployments - Hello!
Let's say I already have a shared Dev remote cluster running some deployments (say App1, App2 and App3) that all communicate with each other.
Is it possible to temporary "replace" the cluster's App2 by a local App2 using Skaffold?
I may have misunderstood the documentation but I'm afraid I'll have to re-deploy an entire, personal new set of deployments each time I need to do some development to one component of my cluster.
It seems like `remoteManifests`, as mentioned in #2160, could help me. But I'd like some confirmation and guidance as of how to achieve this.
Thanks in advance. | priority | using pre existing deployments hello let s say i already have a shared dev remote cluster running some deployments say and that all communicate with each other is it possible to temporary replace the cluster s by a local using skaffold i may have misunderstood the documentation but i m afraid i ll have to re deploy an entire personal new set of deployments each time i need to do some development to one component of my cluster it seems like remotemanifests as mentioned in could help me but i d like some confirmation and guidance as of how to achieve this thanks in advance | 1 |
171,069 | 27,056,940,347 | IssuesEvent | 2023-02-13 16:47:30 | microsoft/pyright | https://api.github.com/repos/microsoft/pyright | closed | Cannot use type variable for the init argument type and method type | as designed | **Describe the bug**
I created a class that receives any subclass instance of `Exception` class as an argument when initializing the class.
I defined a property that returns the exception and I want to use a type variable because those two types should be the same, but the pyright error occurs.
```
Type "T@__init__" cannot be assigned to type "T@orig_exp" (reportGeneralTypeIssues)
```
**To Reproduce**
```sh
$ pyright sample.py
```
sample.py
```python
from typing import TypeVar
T = TypeVar("T", bound=Exception)
class SampleException(Exception):
def __init__(self, msg: str, orig_exp: T):
self.__msg = msg
self.__orig_exp = orig_exp
@property
def msg(self) -> str:
return self.__msg
@property
def orig_exp(self) -> T:
return self.__orig_exp
```
**Expected behavior**
There should be no lint error.
**VS Code extension or command-line**
command-line with the pyright version 1.1.293 | 1.0 | Cannot use type variable for the init argument type and method type - **Describe the bug**
I created a class that receives any subclass instance of `Exception` class as an argument when initializing the class.
I defined a property that returns the exception and I want to use a type variable because those two types should be the same, but the pyright error occurs.
```
Type "T@__init__" cannot be assigned to type "T@orig_exp" (reportGeneralTypeIssues)
```
**To Reproduce**
```sh
$ pyright sample.py
```
sample.py
```python
from typing import TypeVar
T = TypeVar("T", bound=Exception)
class SampleException(Exception):
def __init__(self, msg: str, orig_exp: T):
self.__msg = msg
self.__orig_exp = orig_exp
@property
def msg(self) -> str:
return self.__msg
@property
def orig_exp(self) -> T:
return self.__orig_exp
```
**Expected behavior**
There should be no lint error.
**VS Code extension or command-line**
command-line with the pyright version 1.1.293 | non_priority | cannot use type variable for the init argument type and method type describe the bug i created a class that receives any subclass instance of exception class as an argument when initializing the class i defined a property that returns the exception and i want to use a type variable because those two types should be the same but the pyright error occurs type t init cannot be assigned to type t orig exp reportgeneraltypeissues to reproduce sh pyright sample py sample py python from typing import typevar t typevar t bound exception class sampleexception exception def init self msg str orig exp t self msg msg self orig exp orig exp property def msg self str return self msg property def orig exp self t return self orig exp expected behavior there should be no lint error vs code extension or command line command line with the pyright version | 0 |
195,071 | 14,701,879,417 | IssuesEvent | 2021-01-04 12:40:16 | microcks/microcks | https://api.github.com/repos/microcks/microcks | closed | Reported test duration for asynchronous tests is not correct | component/tests component/ux kind/bug | When dealing with asynchronous tests, the total test duration that is reported is not correct.
Each and every message as a duration assigned to the maximum timeout > that's not correct
and the total test duration is a sum of message tests duration > that's not correct too.
<img width="1046" alt="Capture d’écran 2020-12-16 à 17 22 21" src="https://user-images.githubusercontent.com/1538635/102375896-56101800-3fc3-11eb-8bc1-4abe7aa5e908.png">
Fix this. | 1.0 | Reported test duration for asynchronous tests is not correct - When dealing with asynchronous tests, the total test duration that is reported is not correct.
Each and every message as a duration assigned to the maximum timeout > that's not correct
and the total test duration is a sum of message tests duration > that's not correct too.
<img width="1046" alt="Capture d’écran 2020-12-16 à 17 22 21" src="https://user-images.githubusercontent.com/1538635/102375896-56101800-3fc3-11eb-8bc1-4abe7aa5e908.png">
Fix this. | non_priority | reported test duration for asynchronous tests is not correct when dealing with asynchronous tests the total test duration that is reported is not correct each and every message as a duration assigned to the maximum timeout that s not correct and the total test duration is a sum of message tests duration that s not correct too img width alt capture d’écran à src fix this | 0 |
659,724 | 21,939,289,953 | IssuesEvent | 2022-05-23 16:22:45 | hashicorp/terraform-provider-google | https://api.github.com/repos/hashicorp/terraform-provider-google | opened | Add support for `tpu_config` to `google_container_cluster` | enhancement priority/0 | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment. If the issue is assigned to the "modular-magician" user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If the issue is assigned to a user, that user is claiming responsibility for the issue. If the issue is assigned to "hashibot", a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. Including use cases and why it would help you is a great way to convince maintainers to spend time on it. --->
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* google_container_cluster
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
# Propose what you think the configuration to take advantage of this feature should look like.
# We may not use it verbatim, but it's helpful in understanding your intent.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* b/230658609
<!---
Note Google Cloud customers who are working with a dedicated Technical Account Manager / Customer Engineer: to expedite the investigation and resolution of this issue, please refer to these instructions: https://github.com/hashicorp/terraform-provider-google/wiki/Customer-Contact#raising-gcp-internal-issues-with-the-provider-development-team
--->
| 1.0 | Add support for `tpu_config` to `google_container_cluster` - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment. If the issue is assigned to the "modular-magician" user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If the issue is assigned to a user, that user is claiming responsibility for the issue. If the issue is assigned to "hashibot", a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. Including use cases and why it would help you is a great way to convince maintainers to spend time on it. --->
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* google_container_cluster
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
# Propose what you think the configuration to take advantage of this feature should look like.
# We may not use it verbatim, but it's helpful in understanding your intent.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* b/230658609
<!---
Note Google Cloud customers who are working with a dedicated Technical Account Manager / Customer Engineer: to expedite the investigation and resolution of this issue, please refer to these instructions: https://github.com/hashicorp/terraform-provider-google/wiki/Customer-Contact#raising-gcp-internal-issues-with-the-provider-development-team
--->
| priority | add support for tpu config to google container cluster community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment if the issue is assigned to the modular magician user it is either in the process of being autogenerated or is planned to be autogenerated soon if the issue is assigned to a user that user is claiming responsibility for the issue if the issue is assigned to hashibot a community member has claimed the issue already description new or affected resource s google container cluster potential terraform configuration tf propose what you think the configuration to take advantage of this feature should look like we may not use it verbatim but it s helpful in understanding your intent references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation b note google cloud customers who are working with a dedicated technical account manager customer engineer to expedite the investigation and resolution of this issue please refer to these instructions | 1 |
255,893 | 8,126,570,711 | IssuesEvent | 2018-08-17 03:05:11 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Modify build_visit to use vtkjpeg with NetCDF on all platforms | Bug Likelihood: 3 - Occasional Priority: Normal Severity: 2 - Minor Irritation | There are some systems other than Macs that don't have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build_visit.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 576
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Modify build_visit to use vtkjpeg with NetCDF on all platforms
Assigned to: Eric Brugger
Category:
Target version: 2.2.1
Author: Eric Brugger
Start: 01/31/2011
Due date:
% Done: 100
Estimated time: 2.0
Created: 01/31/2011 03:15 pm
Updated: 02/04/2011 05:53 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.0.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
There are some systems other than Macs that don't have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build_visit.
Comments:
I modified build_visit to build HDF4 with vtkjpeg on all platforms,not just the Mac. A side effect of this was changing the version ofHDF4 to 4.2.5 on all platforms as well.M svn_bin/build_visit
| 1.0 | Modify build_visit to use vtkjpeg with NetCDF on all platforms - There are some systems other than Macs that don't have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build_visit.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 576
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Modify build_visit to use vtkjpeg with NetCDF on all platforms
Assigned to: Eric Brugger
Category:
Target version: 2.2.1
Author: Eric Brugger
Start: 01/31/2011
Due date:
% Done: 100
Estimated time: 2.0
Created: 01/31/2011 03:15 pm
Updated: 02/04/2011 05:53 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.0.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
There are some systems other than Macs that don't have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build_visit.
Comments:
I modified build_visit to build HDF4 with vtkjpeg on all platforms,not just the Mac. A side effect of this was changing the version ofHDF4 to 4.2.5 on all platforms as well.M svn_bin/build_visit
| priority | modify build visit to use vtkjpeg with netcdf on all platforms there are some systems other than macs that don t have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build visit redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject modify build visit to use vtkjpeg with netcdf on all platforms assigned to eric brugger category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description there are some systems other than macs that don t have libjpeg installed on them so we should switch to using libvtkjpeg on all platforms in build visit comments i modified build visit to build with vtkjpeg on all platforms not just the mac a side effect of this was changing the version to on all platforms as well m svn bin build visit | 1 |
578,751 | 17,152,983,562 | IssuesEvent | 2021-07-14 00:23:25 | hep-gc/glint | https://api.github.com/repos/hep-gc/glint | closed | No error message when failing to create repository | Priority-Low enhancement | In e.g. #56 , there should be some kind of error message displayed instead of just nothing happening at all.
| 1.0 | No error message when failing to create repository - In e.g. #56 , there should be some kind of error message displayed instead of just nothing happening at all.
| priority | no error message when failing to create repository in e g there should be some kind of error message displayed instead of just nothing happening at all | 1 |
263,486 | 8,290,227,207 | IssuesEvent | 2018-09-19 16:43:27 | CyberReboot/poseidon | https://api.github.com/repos/CyberReboot/poseidon | closed | poseidon throws a logging error when the log file hits 10M | bug high-priority | poseidon[31429]: --- Logging error ---
poseidon[31429]: Traceback (most recent call last):
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 72, in emit
poseidon[31429]: self.doRollover()
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 173, in doRollover
poseidon[31429]: self.rotate(self.baseFilename, dfn)
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 113, in rotate
poseidon[31429]: os.rename(source, dest)
poseidon[31429]: OSError: [Errno 16] Resource busy: '/var/log/poseidon.log' -> '/var/log/poseidon.log.1' | 1.0 | poseidon throws a logging error when the log file hits 10M - poseidon[31429]: --- Logging error ---
poseidon[31429]: Traceback (most recent call last):
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 72, in emit
poseidon[31429]: self.doRollover()
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 173, in doRollover
poseidon[31429]: self.rotate(self.baseFilename, dfn)
poseidon[31429]: File "/usr/lib/python3.6/logging/handlers.py", line 113, in rotate
poseidon[31429]: os.rename(source, dest)
poseidon[31429]: OSError: [Errno 16] Resource busy: '/var/log/poseidon.log' -> '/var/log/poseidon.log.1' | priority | poseidon throws a logging error when the log file hits poseidon logging error poseidon traceback most recent call last poseidon file usr lib logging handlers py line in emit poseidon self dorollover poseidon file usr lib logging handlers py line in dorollover poseidon self rotate self basefilename dfn poseidon file usr lib logging handlers py line in rotate poseidon os rename source dest poseidon oserror resource busy var log poseidon log var log poseidon log | 1 |
566,955 | 16,835,047,482 | IssuesEvent | 2021-06-18 10:50:40 | Beep6581/RawTherapee | https://api.github.com/repos/Beep6581/RawTherapee | opened | Nightly builds on Windows create 5-dev-dev cache folder | priority: medium scope: distribution | As reported here https://discuss.pixls.us/t/cant-start-the-latest-dev-version-of-rawtherapee-for-win64/24746/30, using our nightly builds creates a `RawTherapee5-dev-dev` folder in `%LOCALAPPDATA%`. The expected suffix is `5-dev` as per http://rawpedia.rawtherapee.com/File_Paths.
Ping @aferrero2707 and @Beep6581, can you investigate/comment?
@gaaned92 I haven't tested your builds, how do they behave?
Contrary to what is reported on Pixls, I don't see any reason why this would prevent RawTherapee from starting. It does potentially make people lose their preferences unexpectedly and require them to rebuild their cache (which can be slow). | 1.0 | Nightly builds on Windows create 5-dev-dev cache folder - As reported here https://discuss.pixls.us/t/cant-start-the-latest-dev-version-of-rawtherapee-for-win64/24746/30, using our nightly builds creates a `RawTherapee5-dev-dev` folder in `%LOCALAPPDATA%`. The expected suffix is `5-dev` as per http://rawpedia.rawtherapee.com/File_Paths.
Ping @aferrero2707 and @Beep6581, can you investigate/comment?
@gaaned92 I haven't tested your builds, how do they behave?
Contrary to what is reported on Pixls, I don't see any reason why this would prevent RawTherapee from starting. It does potentially make people lose their preferences unexpectedly and require them to rebuild their cache (which can be slow). | priority | nightly builds on windows create dev dev cache folder as reported here using our nightly builds creates a dev dev folder in localappdata the expected suffix is dev as per ping and can you investigate comment i haven t tested your builds how do they behave contrary to what is reported on pixls i don t see any reason why this would prevent rawtherapee from starting it does potentially make people lose their preferences unexpectedly and require them to rebuild their cache which can be slow | 1 |
258,745 | 8,179,392,792 | IssuesEvent | 2018-08-28 16:18:04 | LBHackney-IT/lbh-pattern-library | https://api.github.com/repos/LBHackney-IT/lbh-pattern-library | closed | Configure theme | Priority: High Status: Available Type: Enhancement | Check https://github.com/AcasDigital/acas-component-library
- [x] Hackney colour
- [x] Hackney logo
- [x] How to add/import image?
- [x] Favicon | 1.0 | Configure theme - Check https://github.com/AcasDigital/acas-component-library
- [x] Hackney colour
- [x] Hackney logo
- [x] How to add/import image?
- [x] Favicon | priority | configure theme check hackney colour hackney logo how to add import image favicon | 1 |
702,360 | 24,120,934,505 | IssuesEvent | 2022-09-20 18:38:30 | googleapis/nodejs-billing-budgets | https://api.github.com/repos/googleapis/nodejs-billing-budgets | closed | Integration Tests: should try to list billing account failed | type: bug priority: p1 api: cloudbilling flakybot: issue flakybot: flaky | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 70cac3b65e02641f0e573ed3bbf32be84729da00
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e6f099ee-6f60-4d93-bda6-338241c48e17), [Sponge](http://sponge2/e6f099ee-6f60-4d93-bda6-338241c48e17)
status: failed
<details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to include 'PERMISSION_DENIED'
AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to include 'PERMISSION_DENIED'
at Context.<anonymous> (system-test/system.js:29:14)
at processTicksAndRejections (internal/process/task_queues.js:97:5)</pre></details> | 1.0 | Integration Tests: should try to list billing account failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 70cac3b65e02641f0e573ed3bbf32be84729da00
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e6f099ee-6f60-4d93-bda6-338241c48e17), [Sponge](http://sponge2/e6f099ee-6f60-4d93-bda6-338241c48e17)
status: failed
<details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to include 'PERMISSION_DENIED'
AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to include 'PERMISSION_DENIED'
at Context.<anonymous> (system-test/system.js:29:14)
at processTicksAndRejections (internal/process/task_queues.js:97:5)</pre></details> | priority | integration tests should try to list billing account failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output expected unauthenticated request had inval… to include permission denied assertionerror expected unauthenticated request had inval… to include permission denied at context system test system js at processticksandrejections internal process task queues js | 1 |
28,082 | 5,184,886,053 | IssuesEvent | 2017-01-20 08:25:34 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | CachedExecutorServiceDelegateTest shutdownNow() | Team: Core Type: Defect | from https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/6849/testReport/junit/com.hazelcast.util.executor/CachedExecutorServiceDelegateTest/shutdownNow/
```
com.hazelcast.util.executor.CachedExecutorServiceDelegateTest.shutdownNow
Failing for the past 1 build (Since Failed#6849 )
Took 0.11 sec.
Error Message
expected:<1> but was:<0>
Stacktrace
java.lang.AssertionError: expected:<1> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at org.junit.Assert.assertEquals(Assert.java:631)
at com.hazelcast.util.executor.CachedExecutorServiceDelegateTest.shutdownNow(CachedExecutorServiceDelegateTest.java:294)
```
It looks like a bug in a test - the "indefinite" task is parked, but it can be woke-up spuriously. | 1.0 | CachedExecutorServiceDelegateTest shutdownNow() - from https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/6849/testReport/junit/com.hazelcast.util.executor/CachedExecutorServiceDelegateTest/shutdownNow/
```
com.hazelcast.util.executor.CachedExecutorServiceDelegateTest.shutdownNow
Failing for the past 1 build (Since Failed#6849 )
Took 0.11 sec.
Error Message
expected:<1> but was:<0>
Stacktrace
java.lang.AssertionError: expected:<1> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at org.junit.Assert.assertEquals(Assert.java:631)
at com.hazelcast.util.executor.CachedExecutorServiceDelegateTest.shutdownNow(CachedExecutorServiceDelegateTest.java:294)
```
It looks like a bug in a test - the "indefinite" task is parked, but it can be woke-up spuriously. | non_priority | cachedexecutorservicedelegatetest shutdownnow from com hazelcast util executor cachedexecutorservicedelegatetest shutdownnow failing for the past build since failed took sec error message expected but was stacktrace java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at com hazelcast util executor cachedexecutorservicedelegatetest shutdownnow cachedexecutorservicedelegatetest java it looks like a bug in a test the indefinite task is parked but it can be woke up spuriously | 0 |
80,494 | 30,307,583,429 | IssuesEvent | 2023-07-10 10:32:11 | vector-im/element-call | https://api.github.com/repos/vector-im/element-call | opened | Glitch in tiling, in spotlight mode, when speaker has video muted | T-Defect | ### Steps to reproduce
View in spotlight mode.
The tile for speakers with video muted tends to jump position, somewhat erratically. Glitch confirmed by other call participants. Recorded below:
https://drive.google.com/file/d/17ZWPIcBQBdhdgfEXC4qUWIvkzwmrFSJ9/view?usp=sharing
Recorded at https://element-call-livekit.netlify.app/room/#design-warm-ups:call.ems.host
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
macOS
### Browser information
Chrome Version 114.0.5735.198 (Official Build) (arm64)
### URL for webapp
https://element-call-livekit.netlify.app/room/#design-warm-ups:call.ems.host
### Will you send logs?
No | 1.0 | Glitch in tiling, in spotlight mode, when speaker has video muted - ### Steps to reproduce
View in spotlight mode.
The tile for speakers with video muted tends to jump position, somewhat erratically. Glitch confirmed by other call participants. Recorded below:
https://drive.google.com/file/d/17ZWPIcBQBdhdgfEXC4qUWIvkzwmrFSJ9/view?usp=sharing
Recorded at https://element-call-livekit.netlify.app/room/#design-warm-ups:call.ems.host
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
macOS
### Browser information
Chrome Version 114.0.5735.198 (Official Build) (arm64)
### URL for webapp
https://element-call-livekit.netlify.app/room/#design-warm-ups:call.ems.host
### Will you send logs?
No | non_priority | glitch in tiling in spotlight mode when speaker has video muted steps to reproduce view in spotlight mode the tile for speakers with video muted tends to jump position somewhat erratically glitch confirmed by other call participants recorded below recorded at outcome what did you expect what happened instead operating system macos browser information chrome version official build url for webapp will you send logs no | 0 |
566,271 | 16,817,165,078 | IssuesEvent | 2021-06-17 08:45:45 | IgniteUI/igniteui-angular | https://api.github.com/repos/IgniteUI/igniteui-angular | closed | IgxSelect dropdown is not positioned correctly in IE11 | browser: IE-11 bug priority: low select status: resolved version: 11.1.x version: 12.0.x version: 12.1.x | ## Description
IgxSelect dropdown is not positioned correctly in IE11.
* igniteui-angular version: 11.1.3
* browser: IE11
## Steps to reproduce
1. Start the attached reproducible sample and open http://localhost:4200/ in IE11.
2. Open the dropdown.
3. Select an item around at the end of the list. For example, select "Item 87".
4. Open the dropdown.
## Result
The dropdown is displayed at the bottom of the window.

## Expected result
The dropdown is displayed at the position where IgxSelect is located as it is when "Item0" is selected.
## Attachments
[cas34222-app1.zip](https://github.com/IgniteUI/igniteui-angular/files/6161981/cas34222-app1.zip)
| 1.0 | IgxSelect dropdown is not positioned correctly in IE11 - ## Description
IgxSelect dropdown is not positioned correctly in IE11.
* igniteui-angular version: 11.1.3
* browser: IE11
## Steps to reproduce
1. Start the attached reproducible sample and open http://localhost:4200/ in IE11.
2. Open the dropdown.
3. Select an item around at the end of the list. For example, select "Item 87".
4. Open the dropdown.
## Result
The dropdown is displayed at the bottom of the window.

## Expected result
The dropdown is displayed at the position where IgxSelect is located as it is when "Item0" is selected.
## Attachments
[cas34222-app1.zip](https://github.com/IgniteUI/igniteui-angular/files/6161981/cas34222-app1.zip)
| priority | igxselect dropdown is not positioned correctly in description igxselect dropdown is not positioned correctly in igniteui angular version browser steps to reproduce start the attached reproducible sample and open in open the dropdown select an item around at the end of the list for example select item open the dropdown result the dropdown is displayed at the bottom of the window expected result the dropdown is displayed at the position where igxselect is located as it is when is selected attachments | 1 |
567,547 | 16,870,468,255 | IssuesEvent | 2021-06-22 03:23:53 | VEuPathDB/plot.data | https://api.github.com/repos/VEuPathDB/plot.data | closed | improve unit tests | good first issue priority 1 | check for returning num bins as well as bin width, null viewports, dates as input in addition to numbers etc | 1.0 | improve unit tests - check for returning num bins as well as bin width, null viewports, dates as input in addition to numbers etc | priority | improve unit tests check for returning num bins as well as bin width null viewports dates as input in addition to numbers etc | 1 |
1,883 | 3,025,181,446 | IssuesEvent | 2015-08-03 06:15:30 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 21523690: Safari should display UIActivityViewController when inline links are long-pressed | classification:ui/usability reproducible:always status:open | #### Description
Summary:
When an inline link is long-pressed in Safari, an action sheet with only a few built-in options is displayed. It would be much better if a UIActivityViewController was presented instead; this would allow all of the share and action extensions that the user has enabled to act on the URL being interacted with.
Steps to Reproduce:
Long-press an inline link in Safari.
Expected Results:
A UIActivityViewController should be displayed with system actions as well as any action and share extensions that the user has enabled.
Actual Results:
An action sheet with only a few select built-in options is displayed.
Regression:
N/A
Notes:
N/A
-
Product Version: iOS 8-9
Created: 2015-06-24T14:26:12.741030
Originated: 2015-06-24T10:26:00
Open Radar Link: http://www.openradar.me/21523690 | True | 21523690: Safari should display UIActivityViewController when inline links are long-pressed - #### Description
Summary:
When an inline link is long-pressed in Safari, an action sheet with only a few built-in options is displayed. It would be much better if a UIActivityViewController was presented instead; this would allow all of the share and action extensions that the user has enabled to act on the URL being interacted with.
Steps to Reproduce:
Long-press an inline link in Safari.
Expected Results:
A UIActivityViewController should be displayed with system actions as well as any action and share extensions that the user has enabled.
Actual Results:
An action sheet with only a few select built-in options is displayed.
Regression:
N/A
Notes:
N/A
-
Product Version: iOS 8-9
Created: 2015-06-24T14:26:12.741030
Originated: 2015-06-24T10:26:00
Open Radar Link: http://www.openradar.me/21523690 | non_priority | safari should display uiactivityviewcontroller when inline links are long pressed description summary when an inline link is long pressed in safari an action sheet with only a few built in options is displayed it would be much better if a uiactivityviewcontroller was presented instead this would allow all of the share and action extensions that the user has enabled to act on the url being interacted with steps to reproduce long press an inline link in safari expected results a uiactivityviewcontroller should be displayed with system actions as well as any action and share extensions that the user has enabled actual results an action sheet with only a few select built in options is displayed regression n a notes n a product version ios created originated open radar link | 0 |
79,706 | 7,723,869,910 | IssuesEvent | 2018-05-24 13:40:50 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | Change Initiatives: add logical flows and indicator sections | fixed (test & close) noteworthy waiting on client | `dynamicSections.logicalFlowsTabgroupSection`
| 1.0 | Change Initiatives: add logical flows and indicator sections - `dynamicSections.logicalFlowsTabgroupSection`
| non_priority | change initiatives add logical flows and indicator sections dynamicsections logicalflowstabgroupsection | 0 |
258,749 | 8,179,401,721 | IssuesEvent | 2018-08-28 16:20:27 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Theme name form Logs-Reviews redirects to a 404 page after deletion | component: reviewer tools priority: p4 state: wontfix triaged type: bug | Steps to reproduce:
1. Submit a theme and review it
2. Delete it
3. Go to Reviewer Tools - Logs - Reviews https://addons.allizom.org/en-US/editors/themes/logs
4. Click on the deleted theme name
Expected results:
**Theme name is no longer a link or it is a link but clicking on it reloads the same page (as it happens for extensions).**
Actual results:
Theme name is a link and redirects to a 404 page.
Notes/issues:
The 404 page that loads has the URL: https://addons.allizom.org/en-US/editors/themes/queue/single/None
Verified on AMO-stage FF56(Win 7). Issue is reproducing on -dev too.
Video for this issue:

| 1.0 | Theme name form Logs-Reviews redirects to a 404 page after deletion - Steps to reproduce:
1. Submit a theme and review it
2. Delete it
3. Go to Reviewer Tools - Logs - Reviews https://addons.allizom.org/en-US/editors/themes/logs
4. Click on the deleted theme name
Expected results:
**Theme name is no longer a link or it is a link but clicking on it reloads the same page (as it happens for extensions).**
Actual results:
Theme name is a link and redirects to a 404 page.
Notes/issues:
The 404 page that loads has the URL: https://addons.allizom.org/en-US/editors/themes/queue/single/None
Verified on AMO-stage FF56(Win 7). Issue is reproducing on -dev too.
Video for this issue:

| priority | theme name form logs reviews redirects to a page after deletion steps to reproduce submit a theme and review it delete it go to reviewer tools logs reviews click on the deleted theme name expected results theme name is no longer a link or it is a link but clicking on it reloads the same page as it happens for extensions actual results theme name is a link and redirects to a page notes issues the page that loads has the url verified on amo stage win issue is reproducing on dev too video for this issue | 1 |
149,090 | 11,882,171,636 | IssuesEvent | 2020-03-27 13:59:06 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/advanced_job·ts - machine learning anomaly detection advanced job with categorization detector and default datafeed settings job creation displays the pick fields step | :ml blocker failed-test skipped-test | A test failed on a tracked branch
```
Error: expected testSubject(mlJobWizardStepTitlePickFields) to exist
at TestSubjects.existOrFail (/dev/shm/workspace/kibana/test/functional/services/test_subjects.ts:62:15)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/3932/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/advanced_job·ts","test.name":"machine learning anomaly detection advanced job with categorization detector and default datafeed settings job creation displays the pick fields step","test.failCount":3}} --> | 2.0 | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/advanced_job·ts - machine learning anomaly detection advanced job with categorization detector and default datafeed settings job creation displays the pick fields step - A test failed on a tracked branch
```
Error: expected testSubject(mlJobWizardStepTitlePickFields) to exist
at TestSubjects.existOrFail (/dev/shm/workspace/kibana/test/functional/services/test_subjects.ts:62:15)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/3932/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/advanced_job·ts","test.name":"machine learning anomaly detection advanced job with categorization detector and default datafeed settings job creation displays the pick fields step","test.failCount":3}} --> | non_priority | failing test chrome x pack ui functional tests x pack test functional apps machine learning anomaly detection advanced job·ts machine learning anomaly detection advanced job with categorization detector and default datafeed settings job creation displays the pick fields step a test failed on a tracked branch error expected testsubject mljobwizardsteptitlepickfields to exist at testsubjects existorfail dev shm workspace kibana test functional services test subjects ts first failure | 0 |
315,647 | 9,630,241,096 | IssuesEvent | 2019-05-15 11:34:46 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | No logs and containers not being updated | Priority: Low Status: Awaiting user Type: Question | Hi, I use a ansible role to deploy watchtower to my server, however I notice there is no logs with `docker logs watchtower`, zero output.
And my containers are not also being updated... I know this because one of the dockers I use changed their dockerfile and it did not update.
```
- name: Create container
docker_container:
name: watchtower
image: "containrrr/watchtower"
pull: yes
command: “--cleanup --stop-timeout 90s --debug”
volumes:
- "/etc/localtime:/etc/localtime:ro"
- "/var/run/docker.sock:/var/run/docker.sock"
labels:
"com.github.cloudbox.cloudbox_managed": "true"
networks:
- name: cloudbox
aliases:
- watchtower
purge_networks: yes
restart_policy: always
state: started
```
What could be causing this? | 1.0 | No logs and containers not being updated - Hi, I use a ansible role to deploy watchtower to my server, however I notice there is no logs with `docker logs watchtower`, zero output.
And my containers are not also being updated... I know this because one of the dockers I use changed their dockerfile and it did not update.
```
- name: Create container
docker_container:
name: watchtower
image: "containrrr/watchtower"
pull: yes
command: “--cleanup --stop-timeout 90s --debug”
volumes:
- "/etc/localtime:/etc/localtime:ro"
- "/var/run/docker.sock:/var/run/docker.sock"
labels:
"com.github.cloudbox.cloudbox_managed": "true"
networks:
- name: cloudbox
aliases:
- watchtower
purge_networks: yes
restart_policy: always
state: started
```
What could be causing this? | priority | no logs and containers not being updated hi i use a ansible role to deploy watchtower to my server however i notice there is no logs with docker logs watchtower zero output and my containers are not also being updated i know this because one of the dockers i use changed their dockerfile and it did not update name create container docker container name watchtower image containrrr watchtower pull yes command “ cleanup stop timeout debug” volumes etc localtime etc localtime ro var run docker sock var run docker sock labels com github cloudbox cloudbox managed true networks name cloudbox aliases watchtower purge networks yes restart policy always state started what could be causing this | 1 |
60,642 | 17,023,480,417 | IssuesEvent | 2021-07-03 02:14:55 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [amenity-points] Some names wrap, some don't. | Component: mapnik Priority: major Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 1.20am, Friday, 18th September 2009]**
I've noticed that some names wrap vertically, almost with a line break on every word, and some don't wrap, making a long horizontal label.
For example:
http://osm.org/go/54z6C6pPp-
You can see Hyeseong Girls' Middle School, Gaenjiseu and Yeongnak Church render nice and compactly, but Eokjo Fish Restaurant, Arim Mutual Savings and Sangdong Gas Station render horizontally only.
I think that names should render in the same way for all POIs, in the vertical, wrapped style. Certainly, if there are two pieces of code doing the same thing, but they happen to do it slightly differently, they should be rationalised to avoid code duplication and the likelihood of bugs.
Oh, and my library here renders the name:
http://osm.org/go/54z6I5hHS--
But this one doesn't:
http://osm.org/go/54z6JAiF4--
Is it because it's a horizontal label and too long?
Thank you. | 1.0 | [amenity-points] Some names wrap, some don't. - **[Submitted to the original trac issue database at 1.20am, Friday, 18th September 2009]**
I've noticed that some names wrap vertically, almost with a line break on every word, and some don't wrap, making a long horizontal label.
For example:
http://osm.org/go/54z6C6pPp-
You can see Hyeseong Girls' Middle School, Gaenjiseu and Yeongnak Church render nice and compactly, but Eokjo Fish Restaurant, Arim Mutual Savings and Sangdong Gas Station render horizontally only.
I think that names should render in the same way for all POIs, in the vertical, wrapped style. Certainly, if there are two pieces of code doing the same thing, but they happen to do it slightly differently, they should be rationalised to avoid code duplication and the likelihood of bugs.
Oh, and my library here renders the name:
http://osm.org/go/54z6I5hHS--
But this one doesn't:
http://osm.org/go/54z6JAiF4--
Is it because it's a horizontal label and too long?
Thank you. | non_priority | some names wrap some don t i ve noticed that some names wrap vertically almost with a line break on every word and some don t wrap making a long horizontal label for example you can see hyeseong girls middle school gaenjiseu and yeongnak church render nice and compactly but eokjo fish restaurant arim mutual savings and sangdong gas station render horizontally only i think that names should render in the same way for all pois in the vertical wrapped style certainly if there are two pieces of code doing the same thing but they happen to do it slightly differently they should be rationalised to avoid code duplication and the likelihood of bugs oh and my library here renders the name but this one doesn t is it because it s a horizontal label and too long thank you | 0 |
72,928 | 3,393,034,524 | IssuesEvent | 2015-11-30 22:13:29 | ualbertalib/HydraNorth | https://api.github.com/repos/ualbertalib/HydraNorth | closed | add status facet to results page | priority:high size:small | (from #841) add status facet to results page (for public etc. - not sure if this is possible, since read_access_group_ssim has values for public, university_of_alberta and registered, but not for private - would have to look at how the search results display is being populated)
| 1.0 | add status facet to results page - (from #841) add status facet to results page (for public etc. - not sure if this is possible, since read_access_group_ssim has values for public, university_of_alberta and registered, but not for private - would have to look at how the search results display is being populated)
| priority | add status facet to results page from add status facet to results page for public etc not sure if this is possible since read access group ssim has values for public university of alberta and registered but not for private would have to look at how the search results display is being populated | 1 |
33,192 | 4,817,908,807 | IssuesEvent | 2016-11-04 15:00:52 | uProxy/uproxy | https://api.github.com/repos/uProxy/uproxy | closed | Reverse backpressure | C:SOCKS C:Testing P3 T:Obsolete | Right now, we have a backpressure implementation that takes effect when a TCP socket is feeding into a DataChannel, and the DataChannel is slower. This is sufficient for most cases. However, it is not sufficient if (1) the getter is uploading a large file, (2) the link between the getter and sharer is fast, and (3) the link from the sharer to the server is slow. In that case, the upload will pile up in memory on the sharer side.
Implementing this correctly will require a break in our DataChannel protocol, so it must be a negotiated feature. One possible implementation pattern is to use a pull-based protocol over the DataChannel.
| 1.0 | Reverse backpressure - Right now, we have a backpressure implementation that takes effect when a TCP socket is feeding into a DataChannel, and the DataChannel is slower. This is sufficient for most cases. However, it is not sufficient if (1) the getter is uploading a large file, (2) the link between the getter and sharer is fast, and (3) the link from the sharer to the server is slow. In that case, the upload will pile up in memory on the sharer side.
Implementing this correctly will require a break in our DataChannel protocol, so it must be a negotiated feature. One possible implementation pattern is to use a pull-based protocol over the DataChannel.
| non_priority | reverse backpressure right now we have a backpressure implementation that takes effect when a tcp socket is feeding into a datachannel and the datachannel is slower this is sufficient for most cases however it is not sufficient if the getter is uploading a large file the link between the getter and sharer is fast and the link from the sharer to the server is slow in that case the upload will pile up in memory on the sharer side implementing this correctly will require a break in our datachannel protocol so it must be a negotiated feature one possible implementation pattern is to use a pull based protocol over the datachannel | 0 |
289,140 | 8,855,196,522 | IssuesEvent | 2019-01-09 05:14:58 | visit-dav/issues-test | https://api.github.com/repos/visit-dav/issues-test | closed | Windows installer should have wbronze as the default bank for the LLNL host profiles. | bug likelihood medium priority reviewed severity low | I just installed VisIt 2.9.2 on my 64 bit Windows system and bdivp still appears as the default bank when installing the LLNL open host profiles. It should be wbronze.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2305
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Windows installer should have wbronze as the default bank for the LLNL host profiles.
Assigned to: Kathleen Biagas
Category: -
Target version: 2.10
Author: Eric Brugger
Start: 06/22/2015
Due date:
% Done: 100%
Estimated time: 0.10 hour
Created: 06/22/2015 10:53 am
Updated: 07/01/2015 04:24 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.9.2
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
I just installed VisIt 2.9.2 on my 64 bit Windows system and bdivp still appears as the default bank when installing the LLNL open host profiles. It should be wbronze.
Comments:
Made the mod to the installer.M /windowsbuild/distribution/installation/binaryinstallation.nsi
| 1.0 | Windows installer should have wbronze as the default bank for the LLNL host profiles. - I just installed VisIt 2.9.2 on my 64 bit Windows system and bdivp still appears as the default bank when installing the LLNL open host profiles. It should be wbronze.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2305
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Windows installer should have wbronze as the default bank for the LLNL host profiles.
Assigned to: Kathleen Biagas
Category: -
Target version: 2.10
Author: Eric Brugger
Start: 06/22/2015
Due date:
% Done: 100%
Estimated time: 0.10 hour
Created: 06/22/2015 10:53 am
Updated: 07/01/2015 04:24 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.9.2
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
I just installed VisIt 2.9.2 on my 64 bit Windows system and bdivp still appears as the default bank when installing the LLNL open host profiles. It should be wbronze.
Comments:
Made the mod to the installer.M /windowsbuild/distribution/installation/binaryinstallation.nsi
| priority | windows installer should have wbronze as the default bank for the llnl host profiles i just installed visit on my bit windows system and bdivp still appears as the default bank when installing the llnl open host profiles it should be wbronze redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject windows installer should have wbronze as the default bank for the llnl host profiles assigned to kathleen biagas category target version author eric brugger start due date done estimated time hour created am updated pm likelihood occasional severity minor irritation found in version impact expected use os windows support group any description i just installed visit on my bit windows system and bdivp still appears as the default bank when installing the llnl open host profiles it should be wbronze comments made the mod to the installer m windowsbuild distribution installation binaryinstallation nsi | 1 |
37,037 | 2,814,446,981 | IssuesEvent | 2015-05-18 20:05:45 | geoffhumphrey/brewcompetitiononlineentry | https://api.github.com/repos/geoffhumphrey/brewcompetitiononlineentry | closed | The assigned Stewards list pulls through everyone who volunteered to be stewards.Judges etc, not just assigned stewards | auto-migrated Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `lee.Imm...@btinternet.com` on 23 Apr 2015 at 10:46 | 1.0 | The assigned Stewards list pulls through everyone who volunteered to be stewards.Judges etc, not just assigned stewards - ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `lee.Imm...@btinternet.com` on 23 Apr 2015 at 10:46 | priority | the assigned stewards list pulls through everyone who volunteered to be stewards judges etc not just assigned stewards what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by lee imm btinternet com on apr at | 1 |
35,719 | 5,004,945,662 | IssuesEvent | 2016-12-12 08:58:39 | Starcounter/Starcounter | https://api.github.com/repos/Starcounter/Starcounter | closed | Create regression tests based on Client Node and PMail | test | Open a session using Client side Node and force fixed session id. Run a series of deterministic http request and check all JSON (and JSON-Patch) results using exact string match.
Use PMail on the server side.
| 1.0 | Create regression tests based on Client Node and PMail - Open a session using Client side Node and force fixed session id. Run a series of deterministic http request and check all JSON (and JSON-Patch) results using exact string match.
Use PMail on the server side.
| non_priority | create regression tests based on client node and pmail open a session using client side node and force fixed session id run a series of deterministic http request and check all json and json patch results using exact string match use pmail on the server side | 0 |
64,312 | 6,899,294,069 | IssuesEvent | 2017-11-24 13:11:16 | geosolutions-it/GeoCollect | https://api.github.com/repos/geosolutions-it/GeoCollect | reopened | Add an administrative tool to create a survey record on demand | enhancement In Test Project: C040 | The administrator should be able to create a survey record with the same info of an item.
This is to have an history record without waiting for a survey to come back from the field. | 1.0 | Add an administrative tool to create a survey record on demand - The administrator should be able to create a survey record with the same info of an item.
This is to have an history record without waiting for a survey to come back from the field. | non_priority | add an administrative tool to create a survey record on demand the administrator should be able to create a survey record with the same info of an item this is to have an history record without waiting for a survey to come back from the field | 0 |
46,179 | 13,150,873,289 | IssuesEvent | 2020-08-09 13:57:52 | shaundmorris/ddf | https://api.github.com/repos/shaundmorris/ddf | closed | CVE-2018-9159 Medium Severity Vulnerability detected by WhiteSource | security vulnerability wontfix | ## CVE-2018-9159 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spark-core-2.5.5.jar</b></p></summary>
<p></p>
<p>path: /root/.m2/repository/com/sparkjava/spark-core/2.5.5/spark-core-2.5.5.jar,2/repository/com/sparkjava/spark-core/2.5.5/spark-core-2.5.5.jar</p>
<p>
Dependency Hierarchy:
- apps-2.14.0-SNAPSHOT.xml (Root Library)
- search-ui-app-2.14.0-SNAPSHOT.xml
- catalog-ui-search-2.14.0-SNAPSHOT.jar
- :x: **spark-core-2.5.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaundmorris/ddf/commit/ea35fc52f05b85ef437e9a9e39a887ad51692ff0">ea35fc52f05b85ef437e9a9e39a887ad51692ff0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spark before 2.7.2, a remote attacker can read unintended static files via various representations of absolute or relative pathnames, as demonstrated by file: URLs and directory traversal sequences. NOTE: this product is unrelated to Ignite Realtime Spark.
<p>Publish Date: 2018-03-31
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-9159>CVE-2018-9159</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/perwendel/spark/commit/a221a864db28eb736d36041df2fa6eb8839fc5cd">https://github.com/perwendel/spark/commit/a221a864db28eb736d36041df2fa6eb8839fc5cd</a></p>
<p>Release Date: 2018-03-07</p>
<p>Fix Resolution: Replace or update the following files: EmbeddedJettyFactoryTest.java, StaticResources.java, ClassPathResource.java</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-9159 Medium Severity Vulnerability detected by WhiteSource - ## CVE-2018-9159 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spark-core-2.5.5.jar</b></p></summary>
<p></p>
<p>path: /root/.m2/repository/com/sparkjava/spark-core/2.5.5/spark-core-2.5.5.jar,2/repository/com/sparkjava/spark-core/2.5.5/spark-core-2.5.5.jar</p>
<p>
Dependency Hierarchy:
- apps-2.14.0-SNAPSHOT.xml (Root Library)
- search-ui-app-2.14.0-SNAPSHOT.xml
- catalog-ui-search-2.14.0-SNAPSHOT.jar
- :x: **spark-core-2.5.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaundmorris/ddf/commit/ea35fc52f05b85ef437e9a9e39a887ad51692ff0">ea35fc52f05b85ef437e9a9e39a887ad51692ff0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spark before 2.7.2, a remote attacker can read unintended static files via various representations of absolute or relative pathnames, as demonstrated by file: URLs and directory traversal sequences. NOTE: this product is unrelated to Ignite Realtime Spark.
<p>Publish Date: 2018-03-31
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-9159>CVE-2018-9159</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/perwendel/spark/commit/a221a864db28eb736d36041df2fa6eb8839fc5cd">https://github.com/perwendel/spark/commit/a221a864db28eb736d36041df2fa6eb8839fc5cd</a></p>
<p>Release Date: 2018-03-07</p>
<p>Fix Resolution: Replace or update the following files: EmbeddedJettyFactoryTest.java, StaticResources.java, ClassPathResource.java</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium severity vulnerability detected by whitesource cve medium severity vulnerability vulnerable library spark core jar path root repository com sparkjava spark core spark core jar repository com sparkjava spark core spark core jar dependency hierarchy apps snapshot xml root library search ui app snapshot xml catalog ui search snapshot jar x spark core jar vulnerable library found in head commit a href vulnerability details in spark before a remote attacker can read unintended static files via various representations of absolute or relative pathnames as demonstrated by file urls and directory traversal sequences note this product is unrelated to ignite realtime spark publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following files embeddedjettyfactorytest java staticresources java classpathresource java step up your open source security game with whitesource | 0 |
361,492 | 10,709,466,218 | IssuesEvent | 2019-10-24 22:14:33 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.downtowndoglounge.com - Streaming video appears to buffer, but never plays | browser-fenix engine-gecko priority-normal severity-important | <!-- @browser: Firefox Preview 1.3.1 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 -->
<!-- @reported_with: -->
**URL**: http://www.downtowndoglounge.com/webcam-slu_westpark.html
**Browser / Version**: Firefox Preview 1.3.1
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: Streaming video appears to buffer, but never plays
**Steps to Reproduce**:
Load page, tap on video.
Says buffering, then goes back to static preview.
Should start streaming video.
Firefox Preview
1.3.1 (Build #12341949)
📦: 8.0.0, 78354bc12
🦎: 69.0-20190812090043
Same problem on Firefox Mobile 68.1.
Works on Chrome mobile 76.
Works on Safari on an iPhone.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Submitted in the name of `@jacktose`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.downtowndoglounge.com - Streaming video appears to buffer, but never plays - <!-- @browser: Firefox Preview 1.3.1 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 -->
<!-- @reported_with: -->
**URL**: http://www.downtowndoglounge.com/webcam-slu_westpark.html
**Browser / Version**: Firefox Preview 1.3.1
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: Streaming video appears to buffer, but never plays
**Steps to Reproduce**:
Load page, tap on video.
Says buffering, then goes back to static preview.
Should start streaming video.
Firefox Preview
1.3.1 (Build #12341949)
📦: 8.0.0, 78354bc12
🦎: 69.0-20190812090043
Same problem on Firefox Mobile 68.1.
Works on Chrome mobile 76.
Works on Safari on an iPhone.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Submitted in the name of `@jacktose`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | streaming video appears to buffer but never plays url browser version firefox preview operating system android tested another browser yes problem type video or audio doesn t play description streaming video appears to buffer but never plays steps to reproduce load page tap on video says buffering then goes back to static preview should start streaming video firefox preview build 📦 🦎 same problem on firefox mobile works on chrome mobile works on safari on an iphone browser configuration none submitted in the name of jacktose from with ❤️ | 1 |
376,550 | 26,205,737,566 | IssuesEvent | 2023-01-03 22:23:05 | rstudio/rstudio | https://api.github.com/repos/rstudio/rstudio | opened | Consider incorporating software vs. hardware rendering into RStudio User Guide | enhancement documentation | Issues resolved by switching to software rendering pop up every once in a while. It might be worth documenting the rendering engine by incorporating the ["Troubleshooting RStudio Rendering Errors"](https://support.posit.co/hc/en-us/articles/360017886674-Troubleshooting-RStudio-Rendering-Errors) article in the RStudio User Guide, with the perspective of a feature that might also resolve some issues, rather than as a problem that might be resolved by changing a setting.
This is by all means not urgent since this doesn't come up as often as it used to, but we've linked from here to this article probably more than to all of the other support articles combined. | 1.0 | Consider incorporating software vs. hardware rendering into RStudio User Guide - Issues resolved by switching to software rendering pop up every once in a while. It might be worth documenting the rendering engine by incorporating the ["Troubleshooting RStudio Rendering Errors"](https://support.posit.co/hc/en-us/articles/360017886674-Troubleshooting-RStudio-Rendering-Errors) article in the RStudio User Guide, with the perspective of a feature that might also resolve some issues, rather than as a problem that might be resolved by changing a setting.
This is by all means not urgent since this doesn't come up as often as it used to, but we've linked from here to this article probably more than to all of the other support articles combined. | non_priority | consider incorporating software vs hardware rendering into rstudio user guide issues resolved by switching to software rendering pop up every once in a while it might be worth documenting the rendering engine by incorporating the article in the rstudio user guide with the perspective of a feature that might also resolve some issues rather than as a problem that might be resolved by changing a setting this is by all means not urgent since this doesn t come up as often as it used to but we ve linked from here to this article probably more than to all of the other support articles combined | 0 |
282,404 | 8,706,266,187 | IssuesEvent | 2018-12-06 01:57:56 | zulip/zulip | https://api.github.com/repos/zulip/zulip | opened | Add support for denying guest users access to all other users in the organization | area: settings (admin/org) difficult priority: high | Zulip's current "guest users" feature is designed to handle well the case of contractors, where you want someone to be part of your organization and able to send PMs to anyone, but limited to only a few streams.
For some use cases (e.g. customer support), folks often want to have a more limited guest user feature, where the guests are unable to send PMs to or otherwise interact with users not subscribed to streams they are a member of. This issue is to track for what would be required to support that.
* [ ] The set of "users" data delivered by the `/register` endpoint (and also any `GET /users` type endpoints) will need to be restricted to the set of users who have ever been subscribed to a stream the guest user has ever been on (i.e. we'll be using a `Subscription()` model lookup, but without limiting for `active=True`, both on the guest side and the other user side. We need to do this because the guest user's clients need to be able to understand any messages that they might have access to.
* [ ] We need to restrict which users the guests can send PMs to, and possibly also which users can PM guest users in a corresponding fashion.
* [ ] The places where we have `realm.presence_disabled` checks need to be updated to support sending limited presence data to guest users. Probably for this, we should just pull the general presence cache and then strip data from it, because that's what makes sense from a cache performance perspective (each guest will have access to a different set of users)
* [ ] (Probably a task for me) We need to look at the denormalized `message` objects that clients send to ensure that they have all the data a guest user client might need to render them. I think this is OK already.
| 1.0 | Add support for denying guest users access to all other users in the organization - Zulip's current "guest users" feature is designed to handle well the case of contractors, where you want someone to be part of your organization and able to send PMs to anyone, but limited to only a few streams.
For some use cases (e.g. customer support), folks often want to have a more limited guest user feature, where the guests are unable to send PMs to or otherwise interact with users not subscribed to streams they are a member of. This issue is to track for what would be required to support that.
* [ ] The set of "users" data delivered by the `/register` endpoint (and also any `GET /users` type endpoints) will need to be restricted to the set of users who have ever been subscribed to a stream the guest user has ever been on (i.e. we'll be using a `Subscription()` model lookup, but without limiting for `active=True`, both on the guest side and the other user side. We need to do this because the guest user's clients need to be able to understand any messages that they might have access to.
* [ ] We need to restrict which users the guests can send PMs to, and possibly also which users can PM guest users in a corresponding fashion.
* [ ] The places where we have `realm.presence_disabled` checks need to be updated to support sending limited presence data to guest users. Probably for this, we should just pull the general presence cache and then strip data from it, because that's what makes sense from a cache performance perspective (each guest will have access to a different set of users)
* [ ] (Probably a task for me) We need to look at the denormalized `message` objects that clients send to ensure that they have all the data a guest user client might need to render them. I think this is OK already.
| priority | add support for denying guest users access to all other users in the organization zulip s current guest users feature is designed to handle well the case of contractors where you want someone to be part of your organization and able to send pms to anyone but limited to only a few streams for some use cases e g customer support folks often want to have a more limited guest user feature where the guests are unable to send pms to or otherwise interact with users not subscribed to streams they are a member of this issue is to track for what would be required to support that the set of users data delivered by the register endpoint and also any get users type endpoints will need to be restricted to the set of users who have ever been subscribed to a stream the guest user has ever been on i e we ll be using a subscription model lookup but without limiting for active true both on the guest side and the other user side we need to do this because the guest user s clients need to be able to understand any messages that they might have access to we need to restrict which users the guests can send pms to and possibly also which users can pm guest users in a corresponding fashion the places where we have realm presence disabled checks need to be updated to support sending limited presence data to guest users probably for this we should just pull the general presence cache and then strip data from it because that s what makes sense from a cache performance perspective each guest will have access to a different set of users probably a task for me we need to look at the denormalized message objects that clients send to ensure that they have all the data a guest user client might need to render them i think this is ok already | 1 |
25,113 | 12,495,545,436 | IssuesEvent | 2020-06-01 13:23:29 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | storage: large value performance degradation since switching to pebble | A-storage C-performance | **What is your situation?**
Starting May 11th, after the merging of https://github.com/cockroachdb/cockroach/pull/48145, we've begun to observe performance regressions in our large-value KV workloads (4kb and 64kb values). See

https://roachperf.crdb.dev/?filter=&view=kv0%2Fenc%3Dfalse%2Fnodes%3D3%2Fsize%3D64kb&tab=aws
| True | storage: large value performance degradation since switching to pebble - **What is your situation?**
Starting May 11th, after the merging of https://github.com/cockroachdb/cockroach/pull/48145, we've begun to observe performance regressions in our large-value KV workloads (4kb and 64kb values). See

https://roachperf.crdb.dev/?filter=&view=kv0%2Fenc%3Dfalse%2Fnodes%3D3%2Fsize%3D64kb&tab=aws
| non_priority | storage large value performance degradation since switching to pebble what is your situation starting may after the merging of we ve begun to observe performance regressions in our large value kv workloads and values see | 0 |
714,047 | 24,548,839,234 | IssuesEvent | 2022-10-12 10:56:04 | alan-turing-institute/AutisticaCitizenScience | https://api.github.com/repos/alan-turing-institute/AutisticaCitizenScience | closed | Plans for August | community events priority-high fujitsu | ## Summary
Kirstie and Georgia won't be able to make agile sprint meetings throughout August, so this issue is to plan out that time. It would be great to continue to do the user testing as Katharina is only here during August. A presentation gathering and showcasing all the work would be wonderful at the end of August.
We were thinking August 26th, with a three-hour slot? We can use the backlog refinement session on the 4th to plan this out!
* [x] Confirm date, time, and attendees.
* [x] Invite community members - possibly insight group members, previous participants, and current volunteers.
* [x] Create agenda
- [ ] Run session and put notes from the session onto GitHub
Thanks everyone!
🌻
| 1.0 | Plans for August - ## Summary
Kirstie and Georgia won't be able to make agile sprint meetings throughout August, so this issue is to plan out that time. It would be great to continue to do the user testing as Katharina is only here during August. A presentation gathering and showcasing all the work would be wonderful at the end of August.
We were thinking August 26th, with a three-hour slot? We can use the backlog refinement session on the 4th to plan this out!
* [x] Confirm date, time, and attendees.
* [x] Invite community members - possibly insight group members, previous participants, and current volunteers.
* [x] Create agenda
- [ ] Run session and put notes from the session onto GitHub
Thanks everyone!
🌻
| priority | plans for august summary kirstie and georgia won t be able to make agile sprint meetings throughout august so this issue is to plan out that time it would be great to continue to do the user testing as katharina is only here during august a presentation gathering and showcasing all the work would be wonderful at the end of august we were thinking august with a three hour slot we can use the backlog refinement session on the to plan this out confirm date time and attendees invite community members possibly insight group members previous participants and current volunteers create agenda run session and put notes from the session onto github thanks everyone 🌻 | 1 |
61,509 | 17,023,711,401 | IssuesEvent | 2021-07-03 03:26:07 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Clos Belyn address incorrect | Component: nominatim Priority: minor Resolution: wontfix Type: defect | **[Submitted to the original trac issue database at 12.56am, Thursday, 12th May 2011]**
The address for Clos Belyn is incorrect. It should read
Clos Belyn, Llandudno Junction, Conwy, Conwy, Wales, United Kingdom
however it currently reads
Clos Belyn, Craig-y-don, Conwy, Clwyd, Wales, United Kingdom
| 1.0 | Clos Belyn address incorrect - **[Submitted to the original trac issue database at 12.56am, Thursday, 12th May 2011]**
The address for Clos Belyn is incorrect. It should read
Clos Belyn, Llandudno Junction, Conwy, Conwy, Wales, United Kingdom
however it currently reads
Clos Belyn, Craig-y-don, Conwy, Clwyd, Wales, United Kingdom
| non_priority | clos belyn address incorrect the address for clos belyn is incorrect it should read clos belyn llandudno junction conwy conwy wales united kingdom however it currently reads clos belyn craig y don conwy clwyd wales united kingdom | 0 |
10,248 | 12,238,241,704 | IssuesEvent | 2020-05-04 19:27:26 | oracle/truffleruby | https://api.github.com/repos/oracle/truffleruby | closed | Default gems can't be upgraded without booting RubyGems | compatibility | TruffleRuby currently lazy-loads RubyGems to improve startup. Unfortunately this means that unless RubyGems is forced to boot, default gems will not honor `gem install`ed upgrades.
Default gems are installed in the standard library (so they are trivially loadable by `require`) but also have a specially-treated set of gemspecs. Those gemspecs are parsed by RubyGems at boot time so that upgraded default gems can be activated when the related standard libraries get required.
The full set of default gems in TruffleRuby is visible here: https://github.com/oracle/truffleruby/tree/master/lib/gems/specifications/default
RubyGems can be forced to boot by running `bundle exec` or via a gem-based executable like `rails`, but a simple Ruby script that requires a default gem-based standard library will not see upgraded versions of that library.
This is important for a couple reasons:
* Running bundler-based apps by requiring `bundler/setup`. By the time you require anything from the `bundler/` subdir, it's too late to activate an upgraded bundler.
* Security and functionality updates to standard libraries like `json` or `psych` won't be honored even if the gems are installed.
* Different default gems may be activated when running a script directly or running it with RubyGems booted, which will be unexpected for users coming from CRuby.
I do not have a fix to suggest, because it may not be possible (currently) to lazy load RubyGems *and* support always-upgradeable default gems.
Full disclosure: I implemented the default gem feature many years ago, and it is necessarily limited by the way RubyGems and `require` work.
For a related example see https://github.com/jruby/jruby/pull/6109 which adapts the same lazy logic to JRuby, and has the same impact on default gems. | True | Default gems can't be upgraded without booting RubyGems - TruffleRuby currently lazy-loads RubyGems to improve startup. Unfortunately this means that unless RubyGems is forced to boot, default gems will not honor `gem install`ed upgrades.
Default gems are installed in the standard library (so they are trivially loadable by `require`) but also have a specially-treated set of gemspecs. Those gemspecs are parsed by RubyGems at boot time so that upgraded default gems can be activated when the related standard libraries get required.
The full set of default gems in TruffleRuby is visible here: https://github.com/oracle/truffleruby/tree/master/lib/gems/specifications/default
RubyGems can be forced to boot by running `bundle exec` or via a gem-based executable like `rails`, but a simple Ruby script that requires a default gem-based standard library will not see upgraded versions of that library.
This is important for a couple reasons:
* Running bundler-based apps by requiring `bundler/setup`. By the time you require anything from the `bundler/` subdir, it's too late to activate an upgraded bundler.
* Security and functionality updates to standard libraries like `json` or `psych` won't be honored even if the gems are installed.
* Different default gems may be activated when running a script directly or running it with RubyGems booted, which will be unexpected for users coming from CRuby.
I do not have a fix to suggest, because it may not be possible (currently) to lazy load RubyGems *and* support always-upgradeable default gems.
Full disclosure: I implemented the default gem feature many years ago, and it is necessarily limited by the way RubyGems and `require` work.
For a related example see https://github.com/jruby/jruby/pull/6109 which adapts the same lazy logic to JRuby, and has the same impact on default gems. | non_priority | default gems can t be upgraded without booting rubygems truffleruby currently lazy loads rubygems to improve startup unfortunately this means that unless rubygems is forced to boot default gems will not honor gem install ed upgrades default gems are installed in the standard library so they are trivially loadable by require but also have a specially treated set of gemspecs those gemspecs are parsed by rubygems at boot time so that upgraded default gems can be activated when the related standard libraries get required the full set of default gems in truffleruby is visible here rubygems can be forced to boot by running bundle exec or via a gem based executable like rails but a simple ruby script that requires a default gem based standard library will not see upgraded versions of that library this is important for a couple reasons running bundler based apps by requiring bundler setup by the time you require anything from the bundler subdir it s too late to activate an upgraded bundler security and functionality updates to standard libraries like json or psych won t be honored even if the gems are installed different default gems may be activated when running a script directly or running it with rubygems booted which will be unexpected for users coming from cruby i do not have a fix to suggest because it may not be possible currently to lazy load rubygems and support always upgradeable default gems full disclosure i implemented the default gem feature many years ago and it is necessarily limited by the way rubygems and require work for a related example see which adapts the same lazy logic to jruby and has the same impact on default gems | 0 |
151,479 | 5,821,020,426 | IssuesEvent | 2017-05-06 01:06:10 | paceuniversity/CS3892017team2 | https://api.github.com/repos/paceuniversity/CS3892017team2 | closed | US10 - Battle System - Implement algorithm - 8hrs | High Priority Sprint 2 Sprint 3 Task3 | Implement the algorithm/equations into code for the battle system that determines if the user wins based on the power of their army | 1.0 | US10 - Battle System - Implement algorithm - 8hrs - Implement the algorithm/equations into code for the battle system that determines if the user wins based on the power of their army | priority | battle system implement algorithm implement the algorithm equations into code for the battle system that determines if the user wins based on the power of their army | 1 |
646,518 | 21,051,067,801 | IssuesEvent | 2022-03-31 20:36:03 | model-bakers/model_bakery | https://api.github.com/repos/model-bakers/model_bakery | closed | AttributeError: 'ManyToManyRel' object has no attribute 'has_default' | bug high priority | After update from 1.3.2 to 1.3.3 started getting exception from title.
Sorry I didn't debug properly this issue and can't say why this is happening but my best guess would be because of this change https://github.com/model-bakers/model_bakery/compare/1.3.2...1.3.3#diff-e5857deb915e241f429a0c118e89e06a3388d3ce1466e3aa4b960b7055172b6dL322
## Expected behavior
```
Baker.get_fields() 1.3.2 version
(
<django.db.models.fields.AutoField: id>,
<django.db.models.fields.related.ForeignKey: group>,
<django.db.models.fields.related.ManyToManyField: service_lines>,
)
```
## Actual behavior
```
Baker.get_fields() 1.3.3 version
{
<django.db.models.fields.AutoField: id>,
<django.db.models.fields.related.ForeignKey: group>,
<django.db.models.fields.related.ManyToManyField: service_lines>,
<ManyToManyRel: myapp.foo1>, # I guess it not suppose to be here
}
```
And as a result of new element from Baker.get_fields()
```
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/python3.8/site-packages/model_bakery/baker.py:89: in make
return [
/python3.8/site-packages/model_bakery/baker.py:90: in <listcomp>
baker.make(
/model_bakery/baker.py:324: in make
return self._make(**params)
/model_bakery/baker.py:371: in _make
self.model_attrs[field.name] = self.generate_value(
> if field.has_default() and field.name not in self.rel_fields:
E AttributeError: 'ManyToManyRel' object has no attribute 'has_default'
/model_bakery/baker.py:566: AttributeError
```
## Reproduction Steps
I don't think that model that I use has any custom behavior and it's just because of how Baker.get_fields() works in new 1.3.3 version
Models that I use anyway
```python
class Foo(models.Model):
slug = models.SlugField("Service line slug", unique=True, max_length=150)
name = models.CharField("Service line name", max_length=150, null=True)
class Foo1(models.Model):
bars = models.ManyToManyField("myapp.Bar")
class Bar(models.Model):
foo = models.ManyToManyField("myapp.Foo", related_name="foos")
baker.make("core.Bar", _quantity=3, slug=cycle(["1", "2", "3"]), _fill_optional=True)
```
### Versions
Python: 3.8.10
Django: 2.2.24
Model Bakery: 1.3.3 | 1.0 | AttributeError: 'ManyToManyRel' object has no attribute 'has_default' - After update from 1.3.2 to 1.3.3 started getting exception from title.
Sorry I didn't debug properly this issue and can't say why this is happening but my best guess would be because of this change https://github.com/model-bakers/model_bakery/compare/1.3.2...1.3.3#diff-e5857deb915e241f429a0c118e89e06a3388d3ce1466e3aa4b960b7055172b6dL322
## Expected behavior
```
Baker.get_fields() 1.3.2 version
(
<django.db.models.fields.AutoField: id>,
<django.db.models.fields.related.ForeignKey: group>,
<django.db.models.fields.related.ManyToManyField: service_lines>,
)
```
## Actual behavior
```
Baker.get_fields() 1.3.3 version
{
<django.db.models.fields.AutoField: id>,
<django.db.models.fields.related.ForeignKey: group>,
<django.db.models.fields.related.ManyToManyField: service_lines>,
<ManyToManyRel: myapp.foo1>, # I guess it not suppose to be here
}
```
And as a result of new element from Baker.get_fields()
```
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/python3.8/site-packages/model_bakery/baker.py:89: in make
return [
/python3.8/site-packages/model_bakery/baker.py:90: in <listcomp>
baker.make(
/model_bakery/baker.py:324: in make
return self._make(**params)
/model_bakery/baker.py:371: in _make
self.model_attrs[field.name] = self.generate_value(
> if field.has_default() and field.name not in self.rel_fields:
E AttributeError: 'ManyToManyRel' object has no attribute 'has_default'
/model_bakery/baker.py:566: AttributeError
```
## Reproduction Steps
I don't think that model that I use has any custom behavior and it's just because of how Baker.get_fields() works in new 1.3.3 version
Models that I use anyway
```python
class Foo(models.Model):
slug = models.SlugField("Service line slug", unique=True, max_length=150)
name = models.CharField("Service line name", max_length=150, null=True)
class Foo1(models.Model):
bars = models.ManyToManyField("myapp.Bar")
class Bar(models.Model):
foo = models.ManyToManyField("myapp.Foo", related_name="foos")
baker.make("core.Bar", _quantity=3, slug=cycle(["1", "2", "3"]), _fill_optional=True)
```
### Versions
Python: 3.8.10
Django: 2.2.24
Model Bakery: 1.3.3 | priority | attributeerror manytomanyrel object has no attribute has default after update from to started getting exception from title sorry i didn t debug properly this issue and can t say why this is happening but my best guess would be because of this change expected behavior baker get fields version actual behavior baker get fields version i guess it not suppose to be here and as a result of new element from baker get fields site packages model bakery baker py in make return site packages model bakery baker py in baker make model bakery baker py in make return self make params model bakery baker py in make self model attrs self generate value if field has default and field name not in self rel fields e attributeerror manytomanyrel object has no attribute has default model bakery baker py attributeerror reproduction steps i don t think that model that i use has any custom behavior and it s just because of how baker get fields works in new version models that i use anyway python class foo models model slug models slugfield service line slug unique true max length name models charfield service line name max length null true class models model bars models manytomanyfield myapp bar class bar models model foo models manytomanyfield myapp foo related name foos baker make core bar quantity slug cycle fill optional true versions python django model bakery | 1 |
209,346 | 16,190,953,596 | IssuesEvent | 2021-05-04 08:25:07 | nilearn/nilearn | https://api.github.com/repos/nilearn/nilearn | opened | Use bibtex in nilearn/decoding | Documentation Good first issue | Convert references in `nilearn/decoding/decoder.py` to use footcite / footbibliography
See #2759 | 1.0 | Use bibtex in nilearn/decoding - Convert references in `nilearn/decoding/decoder.py` to use footcite / footbibliography
See #2759 | non_priority | use bibtex in nilearn decoding convert references in nilearn decoding decoder py to use footcite footbibliography see | 0 |
627,742 | 19,913,578,207 | IssuesEvent | 2022-01-25 19:49:54 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | [base_bc] can't leave public chat | bug priority 3: low branch: base_bc | ### Description
1. join public chat
2. right click -> leave chat
As result, nothing happens | 1.0 | [base_bc] can't leave public chat - ### Description
1. join public chat
2. right click -> leave chat
As result, nothing happens | priority | can t leave public chat description join public chat right click leave chat as result nothing happens | 1 |
340,070 | 30,492,780,919 | IssuesEvent | 2023-07-18 08:50:08 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: disk-stalled/cgroup/read-write/logs-too=false failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-storage | roachtest.disk-stalled/cgroup/read-write/logs-too=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/disk-stalled/cgroup/read-write/logs-too=false) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-88-n4cpu4/1689663751954223324?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-88-n4cpu4/1689663751954223324?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/disk-stalled/cgroup/read-write/logs-too=false/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/storage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*disk-stalled/cgroup/read-write/logs-too=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29844 | 2.0 | roachtest: disk-stalled/cgroup/read-write/logs-too=false failed - roachtest.disk-stalled/cgroup/read-write/logs-too=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/disk-stalled/cgroup/read-write/logs-too=false) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-88-n4cpu4/1689663751954223324?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-88-n4cpu4/1689663751954223324?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/disk-stalled/cgroup/read-write/logs-too=false/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/storage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*disk-stalled/cgroup/read-write/logs-too=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29844 | non_priority | roachtest disk stalled cgroup read write logs too false failed roachtest disk stalled cgroup read write logs too false with on master cluster go start cockroach connect timeout cockroach sql url postgres root localhost sslmode disable e create schedule if not exists test only backup for backup into gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit recurring full backup hourly with schedule options first run now error unexpected error occurred when checking for existing backups in gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit unable to list files in gcs bucket googleapi error compute developer gserviceaccount com does not have storage objects list access to the google cloud storage bucket permission storage objects list denied on resource or it may not exist sqlstate failed running sql command problem exit status test artifacts and logs in artifacts disk stalled cgroup read write logs too false run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted true roachtest fs roachtest localssd false roachtest ssd help see see cc cockroachdb storage jira issue crdb | 0 |
89,451 | 17,912,903,636 | IssuesEvent | 2021-09-09 08:04:10 | andriy-baran/mother_ship | https://api.github.com/repos/andriy-baran/mother_ship | closed | Fix "method_complexity" issue in lib/mother_ship/builder/strategies/init.rb | codestyle | Method `call` has a Cognitive Complexity of 9 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/andriy-baran/mother_ship/lib/mother_ship/builder/strategies/init.rb#issue_612f7562a428680001000033 | 1.0 | Fix "method_complexity" issue in lib/mother_ship/builder/strategies/init.rb - Method `call` has a Cognitive Complexity of 9 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/andriy-baran/mother_ship/lib/mother_ship/builder/strategies/init.rb#issue_612f7562a428680001000033 | non_priority | fix method complexity issue in lib mother ship builder strategies init rb method call has a cognitive complexity of exceeds allowed consider refactoring | 0 |
675,099 | 23,078,793,350 | IssuesEvent | 2022-07-26 04:15:33 | ballerina-platform/ballerina-dev-website | https://api.github.com/repos/ballerina-platform/ballerina-dev-website | closed | `Report Issues` link in community page `get involved` section should open in a new tab | Priority/Highest Area/UIUX Type/Bug WebsiteRewrite | ## Description
$subject
| 1.0 | `Report Issues` link in community page `get involved` section should open in a new tab - ## Description
$subject
| priority | report issues link in community page get involved section should open in a new tab description subject | 1 |
336,339 | 10,186,081,934 | IssuesEvent | 2019-08-10 09:59:40 | ilakeful/LakeBot | https://api.github.com/repos/ilakeful/LakeBot | opened | Fix of the awaiting user to rejoin the VC | bug changes: patch priority: high type: issue | The feature implies fixing the issue when LakeBot leaves the VC (after the last member in the VC does) and clears the queue if it consists of only one single track. | 1.0 | Fix of the awaiting user to rejoin the VC - The feature implies fixing the issue when LakeBot leaves the VC (after the last member in the VC does) and clears the queue if it consists of only one single track. | priority | fix of the awaiting user to rejoin the vc the feature implies fixing the issue when lakebot leaves the vc after the last member in the vc does and clears the queue if it consists of only one single track | 1 |
75,191 | 9,828,339,844 | IssuesEvent | 2019-06-15 10:37:01 | opensds/documentation | https://api.github.com/repos/opensds/documentation | opened | getting started document | documentation | /kind new-documentation
**What happened**:
There is no proper getting started document available where all the other docs and information can be fetched.
Document Name: New document needed.
Document Path / Location: https://opensds.readthedocs.io/en/latest/
(Like from website/github etc - give the complete link)
Issue Faced / Error in the documentation: Not easy to get the project information, install guides etc
**What you expected to happen**:
Easy access to basic documentation. Usage, Installation and integration information can be fetched easily.
**How to reproduce it (as minimally and precisely as possible)**:
NA
**Anything else we need to know?**:
NA
**Environment**:
- Hotpot(release/branch) version: 0.5.4
- OS (e.g. from /etc/os-release):
- locale of the OS:
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| 1.0 | getting started document - /kind new-documentation
**What happened**:
There is no proper getting started document available where all the other docs and information can be fetched.
Document Name: New document needed.
Document Path / Location: https://opensds.readthedocs.io/en/latest/
(Like from website/github etc - give the complete link)
Issue Faced / Error in the documentation: Not easy to get the project information, install guides etc
**What you expected to happen**:
Easy access to basic documentation. Usage, Installation and integration information can be fetched easily.
**How to reproduce it (as minimally and precisely as possible)**:
NA
**Anything else we need to know?**:
NA
**Environment**:
- Hotpot(release/branch) version: 0.5.4
- OS (e.g. from /etc/os-release):
- locale of the OS:
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| non_priority | getting started document kind new documentation what happened there is no proper getting started document available where all the other docs and information can be fetched document name new document needed document path location like from website github etc give the complete link issue faced error in the documentation not easy to get the project information install guides etc what you expected to happen easy access to basic documentation usage installation and integration information can be fetched easily how to reproduce it as minimally and precisely as possible na anything else we need to know na environment hotpot release branch version os e g from etc os release locale of the os kernel e g uname a install tools others | 0 |
245,280 | 7,884,102,916 | IssuesEvent | 2018-06-27 08:06:51 | LaCoolCo/LaCOOLBoard | https://api.github.com/repos/LaCoolCo/LaCOOLBoard | closed | Prevent SPIFFS corruption on low battery | bug hard high priority | **estimate:** 3d
**Steps to reproduce**
* Make a device/sketch that drains the battery real, real fast
* Use 2 kinds of battery: 2000 mAh, 6600 mAh
* Plug those batteries to unplugged COOL Boards
* Let the batteries drain and observe
**Expected**
- If voltage is 1.8V < v < 3.55V shut down immediately and do not attempt to log anything, read anything on SPIFFS.
- If voltage is > 3.55V or < 1.8V (we have no battery and a jetPack), continue as normal
**Actual**
- SPIFFS ends up corrupted and board is bricked | 1.0 | Prevent SPIFFS corruption on low battery - **estimate:** 3d
**Steps to reproduce**
* Make a device/sketch that drains the battery real, real fast
* Use 2 kinds of battery: 2000 mAh, 6600 mAh
* Plug those batteries to unplugged COOL Boards
* Let the batteries drain and observe
**Expected**
- If voltage is 1.8V < v < 3.55V shut down immediately and do not attempt to log anything, read anything on SPIFFS.
- If voltage is > 3.55V or < 1.8V (we have no battery and a jetPack), continue as normal
**Actual**
- SPIFFS ends up corrupted and board is bricked | priority | prevent spiffs corruption on low battery estimate steps to reproduce make a device sketch that drains the battery real real fast use kinds of battery mah mah plug those batteries to unplugged cool boards let the batteries drain and observe expected if voltage is v shut down immediately and do not attempt to log anything read anything on spiffs if voltage is or we have no battery and a jetpack continue as normal actual spiffs ends up corrupted and board is bricked | 1 |
264,787 | 8,319,444,890 | IssuesEvent | 2018-09-25 17:14:37 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Product-name and summary for localized add-ons are not correctly submitted to Akismet | component: admin tools component: devhub priority: p2 | STR:
1. Log in to Developer Hub
2. Submit an add-on that supports various localisations
3. With an Admin account, check the Akismet reports for the submitted add-on
Actual result:
`Product-summary` and `product-name` seem incorrectly extracted (i.e displayed as `__MSG_name__`)
Expected result:
Localized strings are correctly displayed by Akismet reports
Notes:
- this is an example submitted on the site: https://addons-dev.allizom.org/en-US/firefox/addon/privacy_badger/
- I've submitted the same example a few days ago and this issue did not occur
- could it be a regression from #9461 ?
- reproduced on -dev and stage with FF62, Win10x64
[1] Akismet report for a localized add-on

[2] same example from a few days ago

| 1.0 | Product-name and summary for localized add-ons are not correctly submitted to Akismet - STR:
1. Log in to Developer Hub
2. Submit an add-on that supports various localisations
3. With an Admin account, check the Akismet reports for the submitted add-on
Actual result:
`Product-summary` and `product-name` seem incorrectly extracted (i.e displayed as `__MSG_name__`)
Expected result:
Localized strings are correctly displayed by Akismet reports
Notes:
- this is an example submitted on the site: https://addons-dev.allizom.org/en-US/firefox/addon/privacy_badger/
- I've submitted the same example a few days ago and this issue did not occur
- could it be a regression from #9461 ?
- reproduced on -dev and stage with FF62, Win10x64
[1] Akismet report for a localized add-on

[2] same example from a few days ago

| priority | product name and summary for localized add ons are not correctly submitted to akismet str log in to developer hub submit an add on that supports various localisations with an admin account check the akismet reports for the submitted add on actual result product summary and product name seem incorrectly extracted i e displayed as msg name expected result localized strings are correctly displayed by akismet reports notes this is an example submitted on the site i ve submitted the same example a few days ago and this issue did not occur could it be a regression from reproduced on dev and stage with akismet report for a localized add on same example from a few days ago | 1 |
127,434 | 5,030,408,738 | IssuesEvent | 2016-12-16 00:39:29 | MinetestForFun/server-minetestforfun-skyblock | https://api.github.com/repos/MinetestForFun/server-minetestforfun-skyblock | closed | Worldmap POI request | Priority: Low Website | 13211,x, 11301
Name: Ikoma Japanese Gardens
by: tjnenrtn
:large_orange_diamond: | 1.0 | Worldmap POI request - 13211,x, 11301
Name: Ikoma Japanese Gardens
by: tjnenrtn
:large_orange_diamond: | priority | worldmap poi request x name ikoma japanese gardens by tjnenrtn large orange diamond | 1 |
21,347 | 10,605,471,835 | IssuesEvent | 2019-10-10 20:33:13 | ros-security/aws-roadmap | https://api.github.com/repos/ros-security/aws-roadmap | closed | Run Sandbox integration tests as a batch | theme: security user-story | ## Description
"As a ROS2 Developer, I want to run all sandbox integration tests to ensure any changes are OK"
Instead of running all integration tests separately, run them all in a single bash script or pytest module. This allows easier development workflow for developers.
## Test Plan
* Test the tests, ensure all are still passing.
* Tests must run as part of `colcon test`
## Documentation Plan
Update README. Add a section that tests should pass before opening a PR.
## Release Plan
Push to repo directly, bloom if time allows.
## Acceptance Criteria
* [ ] Code has been implemented, reviewed and merged.
* [ ] Test plan has been completed
* [ ] Release plan has been completed
Once all above items are checked, this story can be moved to done.
## Resources
| True | Run Sandbox integration tests as a batch - ## Description
"As a ROS2 Developer, I want to run all sandbox integration tests to ensure any changes are OK"
Instead of running all integration tests separately, run them all in a single bash script or pytest module. This allows easier development workflow for developers.
## Test Plan
* Test the tests, ensure all are still passing.
* Tests must run as part of `colcon test`
## Documentation Plan
Update README. Add a section that tests should pass before opening a PR.
## Release Plan
Push to repo directly, bloom if time allows.
## Acceptance Criteria
* [ ] Code has been implemented, reviewed and merged.
* [ ] Test plan has been completed
* [ ] Release plan has been completed
Once all above items are checked, this story can be moved to done.
## Resources
| non_priority | run sandbox integration tests as a batch description as a developer i want to run all sandbox integration tests to ensure any changes are ok instead of running all integration tests separately run them all in a single bash script or pytest module this allows easier development workflow for developers test plan test the tests ensure all are still passing tests must run as part of colcon test documentation plan update readme add a section that tests should pass before opening a pr release plan push to repo directly bloom if time allows acceptance criteria code has been implemented reviewed and merged test plan has been completed release plan has been completed once all above items are checked this story can be moved to done resources | 0 |
506,837 | 14,674,073,927 | IssuesEvent | 2020-12-30 14:33:13 | enso-org/ide | https://api.github.com/repos/enso-org/ide | closed | Improper behavior while trying to reconnect node with arguments | Category: GUI Priority: Highest Type: Bug | <!--
Please ensure that you are using the latest version of Enso IDE before reporting
the bug! It may have been fixed since.
-->
### General Summary
<!--
- Please include a high-level description of your bug here.
-->
Reconnecting a node with an argument and a method (`foo.bar baz`) leads to improper behavior. For nodes just with methods (`foo.bar`) everything is ok.
### Steps to Reproduce
<!--
Please list the reproduction steps for your bug.
-->
Open IDE
add node `"foo bar baz"`
add second node `text1.split_at " "`
add new node `"a b c"`
reconnect tex2 instead of text1
### Expected Result
<!--
- A description of the results you expected from the reproduction steps.
-->
smooth reconnection (possibility to connect just to text1 or argument) resulting with `text2.split_at " "` instead of `text1.split_at " "`
### Actual Result
<!--
- A description of what actually happens when you perform these steps.
- Please include any error output if relevant.
-->


### Enso Version
<!--
- Please include the version of Enso IDE you are using here.
-->
core : 2.0.0-alpha.0
build : ec5b0e0
electron : 8.1.1
chrome : 80.0.3987.141
| 1.0 | Improper behavior while trying to reconnect node with arguments - <!--
Please ensure that you are using the latest version of Enso IDE before reporting
the bug! It may have been fixed since.
-->
### General Summary
<!--
- Please include a high-level description of your bug here.
-->
Reconnecting a node with an argument and a method (`foo.bar baz`) leads to improper behavior. For nodes just with methods (`foo.bar`) everything is ok.
### Steps to Reproduce
<!--
Please list the reproduction steps for your bug.
-->
Open IDE
add node `"foo bar baz"`
add second node `text1.split_at " "`
add new node `"a b c"`
reconnect tex2 instead of text1
### Expected Result
<!--
- A description of the results you expected from the reproduction steps.
-->
smooth reconnection (possibility to connect just to text1 or argument) resulting with `text2.split_at " "` instead of `text1.split_at " "`
### Actual Result
<!--
- A description of what actually happens when you perform these steps.
- Please include any error output if relevant.
-->


### Enso Version
<!--
- Please include the version of Enso IDE you are using here.
-->
core : 2.0.0-alpha.0
build : ec5b0e0
electron : 8.1.1
chrome : 80.0.3987.141
| priority | improper behavior while trying to reconnect node with arguments please ensure that you are using the latest version of enso ide before reporting the bug it may have been fixed since general summary please include a high level description of your bug here reconnecting a node with an argument and a method foo bar baz leads to improper behavior for nodes just with methods foo bar everything is ok steps to reproduce please list the reproduction steps for your bug open ide add node foo bar baz add second node split at add new node a b c reconnect instead of expected result a description of the results you expected from the reproduction steps smooth reconnection possibility to connect just to or argument resulting with split at instead of split at actual result a description of what actually happens when you perform these steps please include any error output if relevant enso version please include the version of enso ide you are using here core alpha build electron chrome | 1 |
328,345 | 9,993,524,328 | IssuesEvent | 2019-07-11 15:31:24 | prysmaticlabs/prysm | https://api.github.com/repos/prysmaticlabs/prysm | closed | Set bitfield replaces existing bitfield | Priority: Low | This [line](https://github.com/prysmaticlabs/prysm/blob/master/shared/bitutil/bit.go#L22) from set bit assumes there's no aggregation and the bitfield is only represented by *one* validator since it overrides. We'll need to fix this when we support aggregation
Example:
1.) Before `01001111`
2.) After setting bit for validator at index 0: `1000000`
3.) But it should be `11001111` | 1.0 | Set bitfield replaces existing bitfield - This [line](https://github.com/prysmaticlabs/prysm/blob/master/shared/bitutil/bit.go#L22) from set bit assumes there's no aggregation and the bitfield is only represented by *one* validator since it overrides. We'll need to fix this when we support aggregation
Example:
1.) Before `01001111`
2.) After setting bit for validator at index 0: `1000000`
3.) But it should be `11001111` | priority | set bitfield replaces existing bitfield this from set bit assumes there s no aggregation and the bitfield is only represented by one validator since it overrides we ll need to fix this when we support aggregation example before after setting bit for validator at index but it should be | 1 |
284,668 | 24,614,707,465 | IssuesEvent | 2022-10-15 06:38:08 | BoBAdministration/QA-Bug-Reports | https://api.github.com/repos/BoBAdministration/QA-Bug-Reports | closed | Acacia tree asset trunk moves in the wind. | Fixed-PendingTesting | **Describe the Bug**
The broken branch ends on the Acacia tree asset are programmed to move just like the leaves. This is especially noticeable during stormy weather.
**To Reproduce**
Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones.
1. Log onto any server with the asset, Live or Tester.
2. Spawn in as any creature
3. Find tree
4. Change the weather to windy or stormy
5. Angle the camera to the broken branches.
6. It go wiggly wiggly
**Expected behavior**
The cellulose stays put
**Actual behavior**
The cellulose wiggles like a leaf
**Screenshots & Video**
https://www.youtube.com/watch?v=CqhZhAJhhQ0 bottom right corner
https://www.youtube.com/watch?v=qw4CIItPrLQ
**Branch Version**
Tester AND Live
**Character Information**
Not needed. Can be anything
**Additional Information**
Best visibility in wind. | 1.0 | Acacia tree asset trunk moves in the wind. - **Describe the Bug**
The broken branch ends on the Acacia tree asset are programmed to move just like the leaves. This is especially noticeable during stormy weather.
**To Reproduce**
Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones.
1. Log onto any server with the asset, Live or Tester.
2. Spawn in as any creature
3. Find tree
4. Change the weather to windy or stormy
5. Angle the camera to the broken branches.
6. It go wiggly wiggly
**Expected behavior**
The cellulose stays put
**Actual behavior**
The cellulose wiggles like a leaf
**Screenshots & Video**
https://www.youtube.com/watch?v=CqhZhAJhhQ0 bottom right corner
https://www.youtube.com/watch?v=qw4CIItPrLQ
**Branch Version**
Tester AND Live
**Character Information**
Not needed. Can be anything
**Additional Information**
Best visibility in wind. | non_priority | acacia tree asset trunk moves in the wind describe the bug the broken branch ends on the acacia tree asset are programmed to move just like the leaves this is especially noticeable during stormy weather to reproduce steps to reproduce the behavior in detail please include all steps even menial ones log onto any server with the asset live or tester spawn in as any creature find tree change the weather to windy or stormy angle the camera to the broken branches it go wiggly wiggly expected behavior the cellulose stays put actual behavior the cellulose wiggles like a leaf screenshots video bottom right corner branch version tester and live character information not needed can be anything additional information best visibility in wind | 0 |
96,686 | 3,971,910,871 | IssuesEvent | 2016-05-04 13:45:16 | NIHL/Drupal7-Migration | https://api.github.com/repos/NIHL/Drupal7-Migration | closed | Training: Session title reads "Register" and "Registration closed" appears on description page for upcoming sessions | bug instruction PRIORITY | Here's an example: http://nihlibrarystg.prod.acquia-sites.com/node/70286/register.
Also, when registration is closed the form should not appear. | 1.0 | Training: Session title reads "Register" and "Registration closed" appears on description page for upcoming sessions - Here's an example: http://nihlibrarystg.prod.acquia-sites.com/node/70286/register.
Also, when registration is closed the form should not appear. | priority | training session title reads register and registration closed appears on description page for upcoming sessions here s an example also when registration is closed the form should not appear | 1 |
51,234 | 13,633,078,179 | IssuesEvent | 2020-09-24 20:45:47 | luniehq/lunie | https://api.github.com/repos/luniehq/lunie | closed | Check if we can enforce CSP on mobile | security :shield: stale | CSP (Content Security Policy headers) are a great security tool.
We should check if it is possible to enforce them too on Lunie mobile
| True | Check if we can enforce CSP on mobile - CSP (Content Security Policy headers) are a great security tool.
We should check if it is possible to enforce them too on Lunie mobile
| non_priority | check if we can enforce csp on mobile csp content security policy headers are a great security tool we should check if it is possible to enforce them too on lunie mobile | 0 |
138,528 | 18,793,948,747 | IssuesEvent | 2021-11-08 19:54:48 | Dima2022/hygieia-codequality-sonar-collector | https://api.github.com/repos/Dima2022/hygieia-codequality-sonar-collector | opened | CVE-2020-14195 (High) detected in jackson-databind-2.5.0.jar | security vulnerability | ## CVE-2020-14195 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: hygieia-codequality-sonar-collector/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.0/jackson-databind-2.5.0.jar</p>
<p>
Dependency Hierarchy:
- core-3.1.11.jar (Root Library)
- spring-cloud-starter-config-1.3.1.RELEASE.jar
- :x: **jackson-databind-2.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/hygieia-codequality-sonar-collector/commit/d72cd85b78442d6e5c56f0a28d43e8922826f909">d72cd85b78442d6e5c56f0a28d43e8922826f909</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).
<p>Publish Date: 2020-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195>CVE-2020-14195</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195</a></p>
<p>Release Date: 2020-06-16</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.capitalone.dashboard:core:3.1.11;org.springframework.cloud:spring-cloud-starter-config:1.3.1.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.5.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14195","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-14195 (High) detected in jackson-databind-2.5.0.jar - ## CVE-2020-14195 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: hygieia-codequality-sonar-collector/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.0/jackson-databind-2.5.0.jar</p>
<p>
Dependency Hierarchy:
- core-3.1.11.jar (Root Library)
- spring-cloud-starter-config-1.3.1.RELEASE.jar
- :x: **jackson-databind-2.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/hygieia-codequality-sonar-collector/commit/d72cd85b78442d6e5c56f0a28d43e8922826f909">d72cd85b78442d6e5c56f0a28d43e8922826f909</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).
<p>Publish Date: 2020-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195>CVE-2020-14195</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14195</a></p>
<p>Release Date: 2020-06-16</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.capitalone.dashboard:core:3.1.11;org.springframework.cloud:spring-cloud-starter-config:1.3.1.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.5.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14195","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to org.jsecurity.realm.jndi.JndiRealmFactory (aka org.jsecurity).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14195","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file hygieia codequality sonar collector pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy core jar root library spring cloud starter config release jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org jsecurity realm jndi jndirealmfactory aka org jsecurity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com capitalone dashboard core org springframework cloud spring cloud starter config release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org jsecurity realm jndi jndirealmfactory aka org jsecurity vulnerabilityurl | 0 |
137,360 | 20,118,051,327 | IssuesEvent | 2022-02-07 21:50:47 | FreeCAD/FreeCAD | https://api.github.com/repos/FreeCAD/FreeCAD | opened | Pattern not working properly on geometric primitives | 🐛 bug 🚜 PartDesign | Issue imported from https://tracker.freecad.org/view.php?id=4698
* **Reporter:** chrisb
* **Date submitted:** 7/3/2021
* **FreeCAD version:** 0.2
* **Category:** Bug
* **Status:** new
* **Tags:**
# Original report text
In some cases the PartDesign pattern functions don't seem to get the coordinate system of geometric primitives right. The patterned features occur then in the wrong place.
[forum discussion](https://forum.freecadweb.org/viewtopic.php?f=3&t=57658)
# FreeCAD Info
```
OS: macOS 10.16
Word size of FreeCAD: 64-bit
Version: 0.20.25157 (Git)
Build type: Release
Branch: master
Hash: 257cdd8f423f1530b8df6f53e242de629fee4e85
Python version: 3.9.5
Qt version: 5.12.9
Coin version: 4.0.0
OCC version: 7.5.2
Locale: C/Default (C)
```
# Other bug information
* **Priority:** normal
* **Severity:** minor
* **Category:** Bug
* **Updated:** 7/26/2021
# Discussion from Mantis ticket
### Comment by openBrain 2021-07-26 16:22
Just notice this is one of all things that don't work with PD primitives : mirror, ...
Indeed it seems that primitives are by some way messing up the coordinate system.
| 1.0 | Pattern not working properly on geometric primitives - Issue imported from https://tracker.freecad.org/view.php?id=4698
* **Reporter:** chrisb
* **Date submitted:** 7/3/2021
* **FreeCAD version:** 0.2
* **Category:** Bug
* **Status:** new
* **Tags:**
# Original report text
In some cases the PartDesign pattern functions don't seem to get the coordinate system of geometric primitives right. The patterned features occur then in the wrong place.
[forum discussion](https://forum.freecadweb.org/viewtopic.php?f=3&t=57658)
# FreeCAD Info
```
OS: macOS 10.16
Word size of FreeCAD: 64-bit
Version: 0.20.25157 (Git)
Build type: Release
Branch: master
Hash: 257cdd8f423f1530b8df6f53e242de629fee4e85
Python version: 3.9.5
Qt version: 5.12.9
Coin version: 4.0.0
OCC version: 7.5.2
Locale: C/Default (C)
```
# Other bug information
* **Priority:** normal
* **Severity:** minor
* **Category:** Bug
* **Updated:** 7/26/2021
# Discussion from Mantis ticket
### Comment by openBrain 2021-07-26 16:22
Just notice this is one of all things that don't work with PD primitives : mirror, ...
Indeed it seems that primitives are by some way messing up the coordinate system.
| non_priority | pattern not working properly on geometric primitives issue imported from reporter chrisb date submitted freecad version category bug status new tags original report text in some cases the partdesign pattern functions don t seem to get the coordinate system of geometric primitives right the patterned features occur then in the wrong place freecad info os macos word size of freecad bit version git build type release branch master hash python version qt version coin version occ version locale c default c other bug information priority normal severity minor category bug updated discussion from mantis ticket comment by openbrain just notice this is one of all things that don t work with pd primitives mirror indeed it seems that primitives are by some way messing up the coordinate system | 0 |
348,644 | 10,451,042,152 | IssuesEvent | 2019-09-19 12:00:03 | threefoldtech/jumpscaleX_core | https://api.github.com/repos/threefoldtech/jumpscaleX_core | opened | Testing: JSx clients | priority_major | **Description**
We need to test different functions under JumpscaleX_core/JumpscaleCore/clients/X to make sure that it is working as expected.
**Steps to accomplish the task**
- Gather all the requirements for each client and list them in the traceability matrix.
- Design the test cases according to those requirements and report them in TestQuality
- Execute the test cases and report different bugs (Manual execution)
- Automate those test cases
What should be delivered?
1- Traceability matrix
2- Test cases description in testquality
3- Automation test cases | 1.0 | Testing: JSx clients - **Description**
We need to test different functions under JumpscaleX_core/JumpscaleCore/clients/X to make sure that it is working as expected.
**Steps to accomplish the task**
- Gather all the requirements for each client and list them in the traceability matrix.
- Design the test cases according to those requirements and report them in TestQuality
- Execute the test cases and report different bugs (Manual execution)
- Automate those test cases
What should be delivered?
1- Traceability matrix
2- Test cases description in testquality
3- Automation test cases | priority | testing jsx clients description we need to test different functions under jumpscalex core jumpscalecore clients x to make sure that it is working as expected steps to accomplish the task gather all the requirements for each client and list them in the traceability matrix design the test cases according to those requirements and report them in testquality execute the test cases and report different bugs manual execution automate those test cases what should be delivered traceability matrix test cases description in testquality automation test cases | 1 |
217,003 | 24,312,717,690 | IssuesEvent | 2022-09-30 01:11:55 | LalithK90/w3Campus | https://api.github.com/repos/LalithK90/w3Campus | opened | CVE-2021-43980 (High) detected in tomcat-embed-core-9.0.60.jar | security vulnerability | ## CVE-2021-43980 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.60.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.6.6.jar (Root Library)
- spring-boot-starter-tomcat-2.6.6.jar
- :x: **tomcat-embed-core-9.0.60.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/w3Campus/commit/884aaadb9cfd2b17394f6a3444d759579edc8110">884aaadb9cfd2b17394f6a3444d759579edc8110</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The simplified implementation of blocking reads and writes introduced in Tomcat 10 and back-ported to Tomcat 9.0.47 onwards exposed a long standing (but extremely hard to trigger) concurrency bug in Apache Tomcat 10.1.0 to 10.1.0-M12, 10.0.0-M1 to 10.0.18, 9.0.0-M1 to 9.0.60 and 8.5.0 to 8.5.77 that could cause client connections to share an Http11Processor instance resulting in responses, or part responses, to be received by the wrong client.
<p>Publish Date: 2022-09-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43980>CVE-2021-43980</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread/3jjqbsp6j88b198x5rmg99b1qr8ht3g3">https://lists.apache.org/thread/3jjqbsp6j88b198x5rmg99b1qr8ht3g3</a></p>
<p>Release Date: 2022-09-28</p>
<p>Fix Resolution (org.apache.tomcat.embed:tomcat-embed-core): 9.0.62</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.6.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-43980 (High) detected in tomcat-embed-core-9.0.60.jar - ## CVE-2021-43980 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.60.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.6.6.jar (Root Library)
- spring-boot-starter-tomcat-2.6.6.jar
- :x: **tomcat-embed-core-9.0.60.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/w3Campus/commit/884aaadb9cfd2b17394f6a3444d759579edc8110">884aaadb9cfd2b17394f6a3444d759579edc8110</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The simplified implementation of blocking reads and writes introduced in Tomcat 10 and back-ported to Tomcat 9.0.47 onwards exposed a long standing (but extremely hard to trigger) concurrency bug in Apache Tomcat 10.1.0 to 10.1.0-M12, 10.0.0-M1 to 10.0.18, 9.0.0-M1 to 9.0.60 and 8.5.0 to 8.5.77 that could cause client connections to share an Http11Processor instance resulting in responses, or part responses, to be received by the wrong client.
<p>Publish Date: 2022-09-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43980>CVE-2021-43980</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread/3jjqbsp6j88b198x5rmg99b1qr8ht3g3">https://lists.apache.org/thread/3jjqbsp6j88b198x5rmg99b1qr8ht3g3</a></p>
<p>Release Date: 2022-09-28</p>
<p>Fix Resolution (org.apache.tomcat.embed:tomcat-embed-core): 9.0.62</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.6.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href dependency hierarchy spring boot starter web jar root library spring boot starter tomcat jar x tomcat embed core jar vulnerable library found in head commit a href found in base branch master vulnerability details the simplified implementation of blocking reads and writes introduced in tomcat and back ported to tomcat onwards exposed a long standing but extremely hard to trigger concurrency bug in apache tomcat to to to and to that could cause client connections to share an instance resulting in responses or part responses to be received by the wrong client publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core direct dependency fix resolution org springframework boot spring boot starter web step up your open source security game with mend | 0 |
159,936 | 13,776,357,863 | IssuesEvent | 2020-10-08 09:20:58 | bermarte/from-strategy-to-issues | https://api.github.com/repos/bermarte/from-strategy-to-issues | closed | 4. branch kemi | development-strategy documentation enhancement new-branch | **As a site visitor, I want to learn more about Kemi**
## Repo
* [x] developed on a branch named `kemi`
## kemi.md
* [x] create the file
* [x] add their name as the main title
* [x] write a paragraph introduction
* [x] create a link back to the README
## README.md
* [x] change their list item into a link
## development-strategy.md
* [x] write user story
## Github
* [x] push the branch to GitHub
* [ ] create a pull request
* [ ] compare and merge into master
* [ ] close the issue | 1.0 | 4. branch kemi - **As a site visitor, I want to learn more about Kemi**
## Repo
* [x] developed on a branch named `kemi`
## kemi.md
* [x] create the file
* [x] add their name as the main title
* [x] write a paragraph introduction
* [x] create a link back to the README
## README.md
* [x] change their list item into a link
## development-strategy.md
* [x] write user story
## Github
* [x] push the branch to GitHub
* [ ] create a pull request
* [ ] compare and merge into master
* [ ] close the issue | non_priority | branch kemi as a site visitor i want to learn more about kemi repo developed on a branch named kemi kemi md create the file add their name as the main title write a paragraph introduction create a link back to the readme readme md change their list item into a link development strategy md write user story github push the branch to github create a pull request compare and merge into master close the issue | 0 |
49,948 | 6,284,981,042 | IssuesEvent | 2017-07-19 09:11:11 | getsentry/sentry | https://api.github.com/repos/getsentry/sentry | closed | iOS Breadcrumb Design | Design Review Platform: Cocoa | We implemented automatic breadcrumb tracking in iOS, the crumbs itself look very "technical" and could be condensed without loosing any information.
Here is how the breadcrumbs look would look right now ...
<img width="853" alt="screen shot 2016-11-29 at 12 29 07" src="https://cloud.githubusercontent.com/assets/363802/20709150/da0d8e58-b633-11e6-87bb-bdfa25e0c0c5.png">
I've created a quick mockup (of the first 3 entries) what we could do ... Feedback and other suggestions are welcome
<img width="865" alt="screen shot 2016-11-29 at 12 33 37" src="https://cloud.githubusercontent.com/assets/363802/20709163/ed540d02-b633-11e6-9c06-626ba64217f2.png">
Just wanted to clear things up before starting the implementation ... | 1.0 | iOS Breadcrumb Design - We implemented automatic breadcrumb tracking in iOS, the crumbs itself look very "technical" and could be condensed without loosing any information.
Here is how the breadcrumbs look would look right now ...
<img width="853" alt="screen shot 2016-11-29 at 12 29 07" src="https://cloud.githubusercontent.com/assets/363802/20709150/da0d8e58-b633-11e6-87bb-bdfa25e0c0c5.png">
I've created a quick mockup (of the first 3 entries) what we could do ... Feedback and other suggestions are welcome
<img width="865" alt="screen shot 2016-11-29 at 12 33 37" src="https://cloud.githubusercontent.com/assets/363802/20709163/ed540d02-b633-11e6-9c06-626ba64217f2.png">
Just wanted to clear things up before starting the implementation ... | non_priority | ios breadcrumb design we implemented automatic breadcrumb tracking in ios the crumbs itself look very technical and could be condensed without loosing any information here is how the breadcrumbs look would look right now img width alt screen shot at src i ve created a quick mockup of the first entries what we could do feedback and other suggestions are welcome img width alt screen shot at src just wanted to clear things up before starting the implementation | 0 |
38,335 | 8,455,106,140 | IssuesEvent | 2018-10-21 11:43:19 | maesierra/japo | https://api.github.com/repos/maesierra/japo | closed | Set up CI | code quality improvement | A CI to make sure tests are OK it's a good practice.
Travis seems a good candidate to start with, but any other alternatives are fine. | 1.0 | Set up CI - A CI to make sure tests are OK it's a good practice.
Travis seems a good candidate to start with, but any other alternatives are fine. | non_priority | set up ci a ci to make sure tests are ok it s a good practice travis seems a good candidate to start with but any other alternatives are fine | 0 |
254,038 | 19,185,337,229 | IssuesEvent | 2021-12-05 04:35:52 | longturn/freeciv21 | https://api.github.com/repos/longturn/freeciv21 | opened | Add steps to setup KDevelop the first time | documentation Untriaged | **What should be documented? Is there something wrong in the documentation?**
On https://longturn.readthedocs.io/en/latest/Contributing/dev-env.html we talk about KDevelop, but don't give any help for a first time install and configuration.
**Do you have suggestions?**
Add the basics to get a person started and also give some URLs to the KDevelop Manual.
| 1.0 | Add steps to setup KDevelop the first time - **What should be documented? Is there something wrong in the documentation?**
On https://longturn.readthedocs.io/en/latest/Contributing/dev-env.html we talk about KDevelop, but don't give any help for a first time install and configuration.
**Do you have suggestions?**
Add the basics to get a person started and also give some URLs to the KDevelop Manual.
| non_priority | add steps to setup kdevelop the first time what should be documented is there something wrong in the documentation on we talk about kdevelop but don t give any help for a first time install and configuration do you have suggestions add the basics to get a person started and also give some urls to the kdevelop manual | 0 |
99,743 | 16,450,149,029 | IssuesEvent | 2021-05-21 03:43:21 | ElliotChen/spring_boot_example | https://api.github.com/repos/ElliotChen/spring_boot_example | closed | CVE-2020-28491 (High) detected in jackson-dataformat-cbor-2.11.3.jar - autoclosed | security vulnerability | ## CVE-2020-28491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-dataformat-cbor-2.11.3.jar</b></p></summary>
<p>Support for reading and writing Concise Binary Object Representation
([CBOR](https://www.rfc-editor.org/info/rfc7049)
encoded data using Jackson abstractions (streaming API, data binding, tree model)</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson-dataformats-binary">http://github.com/FasterXML/jackson-dataformats-binary</a></p>
<p>Path to dependency file: spring_boot_example/15elastic/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.11.3/jackson-dataformat-cbor-2.11.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.11.3/jackson-dataformat-cbor-2.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-elasticsearch-2.4.1.jar (Root Library)
- spring-data-elasticsearch-4.0.6.RELEASE.jar
- elasticsearch-rest-high-level-client-7.9.3.jar
- elasticsearch-7.9.3.jar
- elasticsearch-x-content-7.9.3.jar
- :x: **jackson-dataformat-cbor-2.11.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package com.fasterxml.jackson.dataformat:jackson-dataformat-cbor from 0 and before 2.11.4, from 2.12.0-rc1 and before 2.12.1. Unchecked allocation of byte buffer can cause a java.lang.OutOfMemoryError exception.
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28491>CVE-2020-28491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution: com.fasterxml.jackson.dataformat:jackson-dataformat-cbor:2.11.4, com.fasterxml.jackson.dataformat:jackson-dataformat-cbor:2.12.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-28491 (High) detected in jackson-dataformat-cbor-2.11.3.jar - autoclosed - ## CVE-2020-28491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-dataformat-cbor-2.11.3.jar</b></p></summary>
<p>Support for reading and writing Concise Binary Object Representation
([CBOR](https://www.rfc-editor.org/info/rfc7049)
encoded data using Jackson abstractions (streaming API, data binding, tree model)</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson-dataformats-binary">http://github.com/FasterXML/jackson-dataformats-binary</a></p>
<p>Path to dependency file: spring_boot_example/15elastic/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.11.3/jackson-dataformat-cbor-2.11.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.11.3/jackson-dataformat-cbor-2.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-elasticsearch-2.4.1.jar (Root Library)
- spring-data-elasticsearch-4.0.6.RELEASE.jar
- elasticsearch-rest-high-level-client-7.9.3.jar
- elasticsearch-7.9.3.jar
- elasticsearch-x-content-7.9.3.jar
- :x: **jackson-dataformat-cbor-2.11.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package com.fasterxml.jackson.dataformat:jackson-dataformat-cbor from 0 and before 2.11.4, from 2.12.0-rc1 and before 2.12.1. Unchecked allocation of byte buffer can cause a java.lang.OutOfMemoryError exception.
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28491>CVE-2020-28491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution: com.fasterxml.jackson.dataformat:jackson-dataformat-cbor:2.11.4, com.fasterxml.jackson.dataformat:jackson-dataformat-cbor:2.12.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in jackson dataformat cbor jar autoclosed cve high severity vulnerability vulnerable library jackson dataformat cbor jar support for reading and writing concise binary object representation encoded data using jackson abstractions streaming api data binding tree model library home page a href path to dependency file spring boot example pom xml path to vulnerable library home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar dependency hierarchy spring boot starter data elasticsearch jar root library spring data elasticsearch release jar elasticsearch rest high level client jar elasticsearch jar elasticsearch x content jar x jackson dataformat cbor jar vulnerable library found in base branch master vulnerability details this affects the package com fasterxml jackson dataformat jackson dataformat cbor from and before from and before unchecked allocation of byte buffer can cause a java lang outofmemoryerror exception publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson dataformat jackson dataformat cbor com fasterxml jackson dataformat jackson dataformat cbor step up your open source security game with whitesource | 0 |
148,846 | 11,868,098,596 | IssuesEvent | 2020-03-26 08:31:46 | nielsen-oss/docs | https://api.github.com/repos/nielsen-oss/docs | closed | hook testing should be best practice for SHARED hooks | enhancement good first issue taken testing-style-guide | currently, we say that hook testing is not preferred but component testing should test it along the way instead. this is true for an internal hook for the sake of the component.
however, when the hook is to be shared across components, it should be tested in isolation as a standalone hook and it is the best practice | 1.0 | hook testing should be best practice for SHARED hooks - currently, we say that hook testing is not preferred but component testing should test it along the way instead. this is true for an internal hook for the sake of the component.
however, when the hook is to be shared across components, it should be tested in isolation as a standalone hook and it is the best practice | non_priority | hook testing should be best practice for shared hooks currently we say that hook testing is not preferred but component testing should test it along the way instead this is true for an internal hook for the sake of the component however when the hook is to be shared across components it should be tested in isolation as a standalone hook and it is the best practice | 0 |
727,660 | 25,043,244,398 | IssuesEvent | 2022-11-05 00:25:22 | ak2yny/Official-Character-Pack-v2-for-Ultimate-Alliance | https://api.github.com/repos/ak2yny/Official-Character-Pack-v2-for-Ultimate-Alliance | closed | Loading Screen Bugs | Bug Low Priority | Reports by UltraMegaMagnus
- [ ] add [Hawkeye & Mephisto loadscreen fix](https://discord.com/channels/449510825385000960/929457958159548506/929607343564533810)
- [ ] Moon Knight loadscreen apparently not showing. Maybe we can test if it loads by finishing his simulator mission? | 1.0 | Loading Screen Bugs - Reports by UltraMegaMagnus
- [ ] add [Hawkeye & Mephisto loadscreen fix](https://discord.com/channels/449510825385000960/929457958159548506/929607343564533810)
- [ ] Moon Knight loadscreen apparently not showing. Maybe we can test if it loads by finishing his simulator mission? | priority | loading screen bugs reports by ultramegamagnus add moon knight loadscreen apparently not showing maybe we can test if it loads by finishing his simulator mission | 1 |
102,523 | 4,156,503,951 | IssuesEvent | 2016-06-16 18:11:14 | Taeir/ContextProject-MIGI2 | https://api.github.com/repos/Taeir/ContextProject-MIGI2 | closed | Fix Webinterface - Quit buttons | Priority C review | Playtesting showed that we need a way to quit the game. Without this feature, mobile phones that connected before will reconnect and fill a spot in the max player count, making the setup before a game finicky. This task is to address that problem.
This is a subtask of #429 - Fix Webinterface | 1.0 | Fix Webinterface - Quit buttons - Playtesting showed that we need a way to quit the game. Without this feature, mobile phones that connected before will reconnect and fill a spot in the max player count, making the setup before a game finicky. This task is to address that problem.
This is a subtask of #429 - Fix Webinterface | priority | fix webinterface quit buttons playtesting showed that we need a way to quit the game without this feature mobile phones that connected before will reconnect and fill a spot in the max player count making the setup before a game finicky this task is to address that problem this is a subtask of fix webinterface | 1 |
318,506 | 27,308,702,692 | IssuesEvent | 2023-02-24 10:22:57 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [APM] Alert link in service inventory should link to the active alerts not all | Team:APM apm:alerting apm:test-plan-regression v8.7.0 | ## Problem
In the service inventory, we display the total number of active alert for the service
Clicking on the alert count leads the user to the Alerts page. However, the alert page by default shows all the alerts (Recovered, Active).
I need one additional click to see the active alerts.
https://user-images.githubusercontent.com/3369346/220392534-b3934e5d-2b4c-41c4-b96e-29dfc45b6d1e.mov
## Proposal
1. Pass the alert status in the `query`
https://github.com/yngrdyn/kibana/blob/main/x-pack/plugins/apm/public/components/routing/templates/apm_service_template/index.tsx#L357-L374
2. or display the active alerts by default on the alerts page
related
https://github.com/elastic/kibana/issues/146702
https://github.com/elastic/kibana/pull/149610 | 1.0 | [APM] Alert link in service inventory should link to the active alerts not all - ## Problem
In the service inventory, we display the total number of active alert for the service
Clicking on the alert count leads the user to the Alerts page. However, the alert page by default shows all the alerts (Recovered, Active).
I need one additional click to see the active alerts.
https://user-images.githubusercontent.com/3369346/220392534-b3934e5d-2b4c-41c4-b96e-29dfc45b6d1e.mov
## Proposal
1. Pass the alert status in the `query`
https://github.com/yngrdyn/kibana/blob/main/x-pack/plugins/apm/public/components/routing/templates/apm_service_template/index.tsx#L357-L374
2. or display the active alerts by default on the alerts page
related
https://github.com/elastic/kibana/issues/146702
https://github.com/elastic/kibana/pull/149610 | non_priority | alert link in service inventory should link to the active alerts not all problem in the service inventory we display the total number of active alert for the service clicking on the alert count leads the user to the alerts page however the alert page by default shows all the alerts recovered active i need one additional click to see the active alerts proposal pass the alert status in the query or display the active alerts by default on the alerts page related | 0 |
46,830 | 7,292,499,562 | IssuesEvent | 2018-02-25 01:54:59 | alexrj/Slic3r | https://api.github.com/repos/alexrj/Slic3r | closed | Permalink to the latest stable and the latest continuous builds | Feature request UI/Documentation | What would be awesome would be a link to the latest stable and the latest continuous build that would redirect to the respective file... that would help for the AppImageHub central directory of available AppImages, and other app collections. | 1.0 | Permalink to the latest stable and the latest continuous builds - What would be awesome would be a link to the latest stable and the latest continuous build that would redirect to the respective file... that would help for the AppImageHub central directory of available AppImages, and other app collections. | non_priority | permalink to the latest stable and the latest continuous builds what would be awesome would be a link to the latest stable and the latest continuous build that would redirect to the respective file that would help for the appimagehub central directory of available appimages and other app collections | 0 |
9,987 | 3,348,333,422 | IssuesEvent | 2015-11-17 01:08:47 | krs-world/bridges-cxx | https://api.github.com/repos/krs-world/bridges-cxx | closed | Validation.h document throws | documentation | All of the validate functions in Validation.h throw strings, but none are documented.
Validation.h's validate methods' throws should be documented. | 1.0 | Validation.h document throws - All of the validate functions in Validation.h throw strings, but none are documented.
Validation.h's validate methods' throws should be documented. | non_priority | validation h document throws all of the validate functions in validation h throw strings but none are documented validation h s validate methods throws should be documented | 0 |
162,629 | 25,568,601,297 | IssuesEvent | 2022-11-30 15:59:00 | ramp4-pcar4/story-ramp | https://api.github.com/repos/ramp4-pcar4/story-ramp | closed | Create table of contents for navigating between Storylines editor slides | StoryRAMP Design R.E.S.P.E.C.T. Editor Priority: Low | *As a Storylines Editor, it would be nice to jump between slides, for example, work on slide 2 and then jump straight to slide 12.*
Implement something similar to the existing ToC navigator for Storylines viewer to jump between editor slides. Should be saving existing work before jumping to a new slide.
In-flight Figma mockups (still in progress by UX Team as of this writing):
https://www.figma.com/file/uFHBYxq5t3zg3Vv3lIxfok/Ramp-Storylines-Editor | 1.0 | Create table of contents for navigating between Storylines editor slides - *As a Storylines Editor, it would be nice to jump between slides, for example, work on slide 2 and then jump straight to slide 12.*
Implement something similar to the existing ToC navigator for Storylines viewer to jump between editor slides. Should be saving existing work before jumping to a new slide.
In-flight Figma mockups (still in progress by UX Team as of this writing):
https://www.figma.com/file/uFHBYxq5t3zg3Vv3lIxfok/Ramp-Storylines-Editor | non_priority | create table of contents for navigating between storylines editor slides as a storylines editor it would be nice to jump between slides for example work on slide and then jump straight to slide implement something similar to the existing toc navigator for storylines viewer to jump between editor slides should be saving existing work before jumping to a new slide in flight figma mockups still in progress by ux team as of this writing | 0 |
308,252 | 9,436,857,506 | IssuesEvent | 2019-04-13 10:29:01 | xmichelo/Beeftext | https://api.github.com/repos/xmichelo/Beeftext | closed | last letter of keyword intermittently replaces char before expanded snippet | bug high priority | I run into this quite a lot with just about all my combos, but here's an example.
### Expected behavior
Say I have a combo in which the keyword "pprox" expands to "approximately." Therefore, if I type "Map scale pprox 1:10,000", I should get "Map scale approximately 1:10,000".
### Actual behavior
Intermittently, the combo works as expected. I haven't yet figured out how or why. Most of the time however, I'll get "Map scalexapproximately 1:10,000", i.e. the space before "pprox" was deleted and replaced with the "x" from the end of the keyword. New lines prior to a keyword get deleted as well. I then have to go back and edit manually, which gets annoying fast. :)
### Steps to reproduce the behavior
I've had trouble *not* reproducing it, actually. It happens all the dang time just in normal use! And then sometimes it doesn't for a little while. I haven't been able to figure out why. :/
I *think* it's more likely to happen if I've deleted a character right before typing the keyword, but it can happen if I just click and type, too. Clipboard expansion can be on or off, it happens either way.
### Operating system
Windows 7 Enterprise, Service Pack 1, 32-bit and 4GB of RAM. I'm a librarian, so I usually encounter the bug in OCLC Connexion, but I've seen it happening in Notepad and Firefox as well (heck, I triggered it on purpose while writing this post!). I don't think it's application-specific.
If it makes a difference, I tend to have several programs open at a time and I use the Dvorak keyboard. | 1.0 | last letter of keyword intermittently replaces char before expanded snippet - I run into this quite a lot with just about all my combos, but here's an example.
### Expected behavior
Say I have a combo in which the keyword "pprox" expands to "approximately." Therefore, if I type "Map scale pprox 1:10,000", I should get "Map scale approximately 1:10,000".
### Actual behavior
Intermittently, the combo works as expected. I haven't yet figured out how or why. Most of the time however, I'll get "Map scalexapproximately 1:10,000", i.e. the space before "pprox" was deleted and replaced with the "x" from the end of the keyword. New lines prior to a keyword get deleted as well. I then have to go back and edit manually, which gets annoying fast. :)
### Steps to reproduce the behavior
I've had trouble *not* reproducing it, actually. It happens all the dang time just in normal use! And then sometimes it doesn't for a little while. I haven't been able to figure out why. :/
I *think* it's more likely to happen if I've deleted a character right before typing the keyword, but it can happen if I just click and type, too. Clipboard expansion can be on or off, it happens either way.
### Operating system
Windows 7 Enterprise, Service Pack 1, 32-bit and 4GB of RAM. I'm a librarian, so I usually encounter the bug in OCLC Connexion, but I've seen it happening in Notepad and Firefox as well (heck, I triggered it on purpose while writing this post!). I don't think it's application-specific.
If it makes a difference, I tend to have several programs open at a time and I use the Dvorak keyboard. | priority | last letter of keyword intermittently replaces char before expanded snippet i run into this quite a lot with just about all my combos but here s an example expected behavior say i have a combo in which the keyword pprox expands to approximately therefore if i type map scale pprox i should get map scale approximately actual behavior intermittently the combo works as expected i haven t yet figured out how or why most of the time however i ll get map scalexapproximately i e the space before pprox was deleted and replaced with the x from the end of the keyword new lines prior to a keyword get deleted as well i then have to go back and edit manually which gets annoying fast steps to reproduce the behavior i ve had trouble not reproducing it actually it happens all the dang time just in normal use and then sometimes it doesn t for a little while i haven t been able to figure out why i think it s more likely to happen if i ve deleted a character right before typing the keyword but it can happen if i just click and type too clipboard expansion can be on or off it happens either way operating system windows enterprise service pack bit and of ram i m a librarian so i usually encounter the bug in oclc connexion but i ve seen it happening in notepad and firefox as well heck i triggered it on purpose while writing this post i don t think it s application specific if it makes a difference i tend to have several programs open at a time and i use the dvorak keyboard | 1 |
90,615 | 8,250,217,963 | IssuesEvent | 2018-09-12 01:29:52 | nltk/nltk | https://api.github.com/repos/nltk/nltk | opened | Cyclic imports in nltk.ccg.lexicon | need-help pleaseverify tests | Pylint seems to be complaining about `nltk.ccg.lexicon` having cyclic imports but I'm not familiar with R0401 to understand what is the error.
Anyone knows how to resolve this issue?
```
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.corpus -> nltk.tokenize -> nltk.tokenize.texttiling)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.draw.tree -> nltk.tree)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tree -> nltk.treeprettyprinter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.grammar -> nltk.parse.pchart)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.stem -> nltk.stem.porter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.classify.maxent -> nltk.classify.tadm)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tag -> nltk.tag.brill)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.shiftreduce)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.cluster -> nltk.cluster.gaac)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.chunk -> nltk.chunk.regexp)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm5)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.mwe -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.treeprettyprinter -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tree -> nltk.treetransforms)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.draw.table -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.collocations -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.corpus -> nltk.tokenize -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.cfg -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.cfg -> nltk.tree -> nltk.treetransforms -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.corenlp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.wsd -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.help -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.featstruct -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.misc -> nltk.misc.wordfinder -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.confusionmatrix -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.agreement -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.scores -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.gaac -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.cluster -> nltk.cluster.kmeans)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.em -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.kmeans -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.tnt -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.senna -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.classify.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.weka -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.corpus.util -> nltk.corpus.reader.api -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.classify.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.scikitlearn -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.classify.megam -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.textcat)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.positivenaivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.decisiontree -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.classify.tadm -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.chunk.api -> nltk.chunk.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.chunk.api -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.featstruct -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.malt -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.inference.mace -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.tnt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.corenlp -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.rslp -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.hmm -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.translate -> nltk.translate.ribes_score -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.translate -> nltk.translate.bleu_score -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm2)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm3)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm4)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm1)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.boxer -> nltk.sem.drt -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.skolemize -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.util -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.boxer -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.lfg -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem.glue -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.relextract)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem.glue -> nltk.sem.linearlogic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.evaluate -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.corpus.reader.util -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.parse.dependencygraph -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.bllip -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.recursivedescent -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.util -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.parse.chart -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.shiftreduce -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.recursivedescent)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.nonprojectivedependencyparser -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.transitionparser)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.earleychart -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.viterbi -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.stem -> nltk.stem.snowball -> nltk.stem.porter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.tableau -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem -> nltk.sem.lfg -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.inference.mace -> nltk.inference.prover9)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.resolution -> nltk.sem.skolemize -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.sentiment -> nltk.sentiment.sentiment_analyzer -> nltk.sentiment.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.sentiment -> nltk.sentiment.vader -> nltk.sentiment.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tree -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.table -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.dispersion -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.perceptron -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.hunpos -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.crf -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.stanford -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.nonprojectivedependencyparser -> nltk.classify -> nltk.classify.naivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tag -> nltk.tag.brill_trainer)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.mapping -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.stanford_segmenter -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.texttiling -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.mwe -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.repp -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.sexpr -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.treebank -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.simple -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.toktok -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.lexicon -> nltk.ccg.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.combinator -> nltk.ccg.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.lexicon -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.lexicon -> nltk.internals)
``` | 1.0 | Cyclic imports in nltk.ccg.lexicon - Pylint seems to be complaining about `nltk.ccg.lexicon` having cyclic imports but I'm not familiar with R0401 to understand what is the error.
Anyone knows how to resolve this issue?
```
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.corpus -> nltk.tokenize -> nltk.tokenize.texttiling)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.draw.tree -> nltk.tree)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tree -> nltk.treeprettyprinter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.grammar -> nltk.parse.pchart)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.stem -> nltk.stem.porter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.classify.maxent -> nltk.classify.tadm)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tag -> nltk.tag.brill)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.shiftreduce)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.cluster -> nltk.cluster.gaac)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.chunk -> nltk.chunk.regexp)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm5)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.mwe -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.treeprettyprinter -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tree -> nltk.treetransforms)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.draw.table -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.collocations -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.corpus -> nltk.tokenize -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.cfg -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.cfg -> nltk.tree -> nltk.treetransforms -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.corenlp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.wsd -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.help -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.featstruct -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.misc -> nltk.misc.wordfinder -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.confusionmatrix -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.agreement -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.metrics -> nltk.metrics.scores -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.gaac -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.cluster -> nltk.cluster.kmeans)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.em -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.cluster -> nltk.cluster.kmeans -> nltk.cluster.util -> nltk.cluster.api -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.tnt -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.senna -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.classify.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.weka -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.corpus.util -> nltk.corpus.reader.api -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.naivebayes -> nltk.classify.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.scikitlearn -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.classify.megam -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.textcat)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.positivenaivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.decisiontree -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.sequential -> nltk.classify -> nltk.classify.rte_classify -> nltk.classify.maxent -> nltk.classify.tadm -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.chunk.api -> nltk.chunk.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.chunk -> nltk.chunk.regexp -> nltk.chunk.api -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.featstruct -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.malt -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.inference.mace -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.tnt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.corenlp -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.rslp -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.hmm -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.translate -> nltk.translate.ribes_score -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.translate -> nltk.translate.bleu_score -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm2)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm3)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm4)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.translate -> nltk.translate.ibm1)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.boxer -> nltk.sem.drt -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.skolemize -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.util -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.boxer -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.lfg -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem.glue -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.relextract)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem.glue -> nltk.sem.linearlogic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.sem -> nltk.sem.evaluate -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.downloader -> nltk.corpus.reader.util -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.projectivedependencyparser -> nltk.parse.dependencygraph -> nltk.tree -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.bllip -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.recursivedescent -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.util -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.tree -> nltk.tree -> nltk.grammar -> nltk.parse.pchart -> nltk.parse.chart -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.shiftreduce -> nltk.parse.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.recursivedescent)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.nonprojectivedependencyparser -> nltk.grammar -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.parse -> nltk.parse.transitionparser)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.earleychart -> nltk.parse.featurechart -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.viterbi -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.stem -> nltk.stem.snowball -> nltk.stem.porter)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.tableau -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.sem -> nltk.sem.lfg -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.discourse -> nltk.inference.mace -> nltk.inference.prover9)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.inference -> nltk.inference.resolution -> nltk.sem.skolemize -> nltk.sem.logic -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.sentiment -> nltk.sentiment.sentiment_analyzer -> nltk.sentiment.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.sentiment -> nltk.sentiment.vader -> nltk.sentiment.util)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tree -> nltk.draw.tree -> nltk.draw.util -> nltk.util -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.table -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.draw -> nltk.draw.dispersion -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.regexp -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.perceptron -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.hunpos -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.crf -> nltk.tag.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.stanford -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.parse -> nltk.parse.nonprojectivedependencyparser -> nltk.classify -> nltk.classify.naivebayes -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk.tag -> nltk.tag.brill_trainer)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.stem -> nltk.stem.wordnet -> nltk.corpus.reader.wordnet -> nltk.corpus.reader -> nltk.corpus.reader.bracket_parse -> nltk.tag -> nltk.tag.mapping -> nltk.data)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.stanford_segmenter -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.texttiling -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.mwe -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.repp -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.sexpr -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.punkt -> nltk.probability -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.treebank -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.simple -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.text -> nltk.collocations -> nltk.probability -> nltk.corpus -> nltk.tokenize -> nltk.tokenize.toktok -> nltk.tokenize.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.lexicon -> nltk.ccg.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.combinator -> nltk.ccg.api -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.chart -> nltk.ccg.lexicon -> nltk.internals)
nltk/nltk/ccg/lexicon.py:1: [R0401(cyclic-import), ] Cyclic import (nltk -> nltk.ccg -> nltk.ccg.lexicon -> nltk.internals)
``` | non_priority | cyclic imports in nltk ccg lexicon pylint seems to be complaining about nltk ccg lexicon having cyclic imports but i m not familiar with to understand what is the error anyone knows how to resolve this issue nltk nltk ccg lexicon py cyclic import nltk nltk internals nltk nltk ccg lexicon py cyclic import nltk corpus nltk tokenize nltk tokenize punkt nltk probability nltk nltk ccg lexicon py cyclic import nltk corpus nltk tokenize nltk tokenize texttiling nltk nltk ccg lexicon py cyclic import nltk draw tree nltk tree nltk nltk ccg lexicon py cyclic import nltk tree nltk treeprettyprinter nltk nltk ccg lexicon py cyclic import nltk grammar nltk parse pchart nltk nltk ccg lexicon py cyclic import nltk stem nltk stem porter nltk nltk ccg lexicon py cyclic import nltk classify maxent nltk classify tadm nltk nltk ccg lexicon py cyclic import nltk tag nltk tag brill nltk nltk ccg lexicon py cyclic import nltk parse nltk parse shiftreduce nltk nltk ccg lexicon py cyclic import nltk cluster nltk cluster gaac nltk nltk ccg lexicon py cyclic import nltk chunk nltk chunk regexp nltk nltk ccg lexicon py cyclic import nltk translate nltk translate nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize mwe nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk draw util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk downloader nltk draw util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk treeprettyprinter nltk internals nltk nltk ccg lexicon py cyclic import nltk tree nltk treetransforms nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk downloader nltk nltk ccg lexicon py cyclic import nltk nltk downloader nltk draw table nltk corpus nltk tokenize nltk tokenize punkt nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk collocations nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk corpus nltk tokenize nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw cfg nltk draw util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw cfg nltk tree nltk treetransforms nltk draw tree nltk draw util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk parse pchart nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse corenlp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk wsd nltk corpus nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk help nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk featstruct nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk misc nltk misc wordfinder nltk corpus nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk metrics nltk metrics confusionmatrix nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk metrics nltk metrics agreement nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk metrics nltk metrics scores nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk cluster nltk cluster gaac nltk cluster util nltk cluster api nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk cluster nltk cluster util nltk cluster api nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk cluster nltk cluster kmeans nltk nltk ccg lexicon py cyclic import nltk nltk cluster nltk cluster em nltk cluster util nltk cluster api nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk cluster nltk cluster kmeans nltk cluster util nltk cluster api nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag tnt nltk tag api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify senna nltk tag api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify naivebayes nltk classify util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify weka nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify naivebayes nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk corpus util nltk corpus reader api nltk data nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify rte classify nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify naivebayes nltk classify api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify rte classify nltk classify maxent nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify scikitlearn nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify rte classify nltk classify maxent nltk classify megam nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify textcat nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify positivenaivebayes nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify decisiontree nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag sequential nltk classify nltk classify rte classify nltk classify maxent nltk classify tadm nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk chunk nltk chunk regexp nltk chunk api nltk chunk util nltk nltk ccg lexicon py cyclic import nltk nltk chunk nltk chunk regexp nltk tree nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse projectivedependencyparser nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk chunk nltk chunk regexp nltk chunk api nltk parse nltk parse projectivedependencyparser nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk featstruct nltk sem logic nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse malt nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference discourse nltk inference mace nltk sem logic nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tree nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag tnt nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse corenlp nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem rslp nltk data nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag hmm nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk translate nltk translate ribes score nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk translate nltk translate bleu score nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk translate nltk translate nltk nltk ccg lexicon py cyclic import nltk translate nltk translate nltk nltk ccg lexicon py cyclic import nltk translate nltk translate nltk nltk ccg lexicon py cyclic import nltk translate nltk translate nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem boxer nltk sem drt nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem skolemize nltk sem logic nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem util nltk grammar nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem boxer nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem lfg nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference discourse nltk sem glue nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem relextract nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference discourse nltk sem glue nltk sem linearlogic nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk sem nltk sem evaluate nltk sem logic nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse featurechart nltk data nltk nltk ccg lexicon py cyclic import nltk nltk downloader nltk corpus reader util nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse projectivedependencyparser nltk parse dependencygraph nltk tree nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse bllip nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk parse pchart nltk parse api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse recursivedescent nltk parse api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse util nltk parse featurechart nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw tree nltk tree nltk grammar nltk parse pchart nltk parse chart nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse shiftreduce nltk parse api nltk internals nltk nltk ccg lexicon py cyclic import nltk parse nltk parse recursivedescent nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse nonprojectivedependencyparser nltk grammar nltk internals nltk nltk ccg lexicon py cyclic import nltk parse nltk parse transitionparser nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse earleychart nltk parse featurechart nltk data nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse viterbi nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk stem nltk stem snowball nltk stem porter nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference tableau nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference discourse nltk sem nltk sem lfg nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference discourse nltk inference mace nltk inference nltk nltk ccg lexicon py cyclic import nltk nltk inference nltk inference resolution nltk sem skolemize nltk sem logic nltk internals nltk nltk ccg lexicon py cyclic import nltk sentiment nltk sentiment sentiment analyzer nltk sentiment util nltk nltk ccg lexicon py cyclic import nltk sentiment nltk sentiment vader nltk sentiment util nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tree nltk draw tree nltk draw util nltk util nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw table nltk corpus nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk draw nltk draw dispersion nltk corpus nltk tokenize nltk tokenize regexp nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag perceptron nltk tag api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag hunpos nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag crf nltk tag api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag stanford nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk parse nltk parse nonprojectivedependencyparser nltk classify nltk classify naivebayes nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk tag nltk tag brill trainer nltk nltk ccg lexicon py cyclic import nltk nltk stem nltk stem wordnet nltk corpus reader wordnet nltk corpus reader nltk corpus reader bracket parse nltk tag nltk tag mapping nltk data nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize stanford segmenter nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize texttiling nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize mwe nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize repp nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize sexpr nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk corpus nltk tokenize nltk tokenize punkt nltk probability nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize treebank nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize simple nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk text nltk collocations nltk probability nltk corpus nltk tokenize nltk tokenize toktok nltk tokenize api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk ccg lexicon nltk ccg api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk ccg combinator nltk ccg api nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg chart nltk ccg lexicon nltk internals nltk nltk ccg lexicon py cyclic import nltk nltk ccg nltk ccg lexicon nltk internals | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.