Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
338,627 | 10,232,453,098 | IssuesEvent | 2019-08-18 17:37:04 | futurismo-org/titan | https://api.github.com/repos/futurismo-org/titan | closed | りんごリジェクト対応(5回目) | high priority | ちゃんと見ていないので、解説する。
1. 1 Safety: Objectionable Content
1. 2 Safety: User Generated Content | 1.0 | りんごリジェクト対応(5回目) - ちゃんと見ていないので、解説する。
1. 1 Safety: Objectionable Content
1. 2 Safety: User Generated Content | priority | りんごリジェクト対応 ちゃんと見ていないので、解説する。 safety objectionable content safety user generated content | 1 |
43,879 | 2,893,715,337 | IssuesEvent | 2015-06-15 19:26:02 | SCIInstitute/shapeworks | https://api.github.com/repos/SCIInstitute/shapeworks | reopened | Final Release | High Priority IBBM | No more features to add at this point. Only bugs that are important for IBBM allowed after this. | 1.0 | Final Release - No more features to add at this point. Only bugs that are important for IBBM allowed after this. | priority | final release no more features to add at this point only bugs that are important for ibbm allowed after this | 1 |
186,230 | 6,734,519,625 | IssuesEvent | 2017-10-18 18:20:16 | resin-io/resin-cli | https://api.github.com/repos/resin-io/resin-cli | closed | resin device register only does the old long UUIDs | priority:high type:bug | With resinOS 2.0, the device UUIDs became half the length (~31~ 32 chars down from 62 characters). Running `resin device register AppName` only able to do the original 62 character version. Not sure if it's a problem, just highlight it, because this way images freshly downloaded from the Dashboard and images configured with the CLI will have different behaviour for 2.0.
What would be the right default behaviour? Or this just doesn't have any real consequences?
<img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_20dt) | 1.0 | resin device register only does the old long UUIDs - With resinOS 2.0, the device UUIDs became half the length (~31~ 32 chars down from 62 characters). Running `resin device register AppName` only able to do the original 62 character version. Not sure if it's a problem, just highlight it, because this way images freshly downloaded from the Dashboard and images configured with the CLI will have different behaviour for 2.0.
What would be the right default behaviour? Or this just doesn't have any real consequences?
<img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_20dt) | priority | resin device register only does the old long uuids with resinos the device uuids became half the length chars down from characters running resin device register appname only able to do the original character version not sure if it s a problem just highlight it because this way images freshly downloaded from the dashboard and images configured with the cli will have different behaviour for what would be the right default behaviour or this just doesn t have any real consequences | 1 |
3,912 | 2,542,061,046 | IssuesEvent | 2015-01-28 14:06:33 | bethlakshmi/GBE2 | https://api.github.com/repos/bethlakshmi/GBE2 | closed | Update Your Bio | High Priority | Your bio, which is now live on the site, reads:
Betty has been doing this too long.
She also ported the data.
You might want to change that... | 1.0 | Update Your Bio - Your bio, which is now live on the site, reads:
Betty has been doing this too long.
She also ported the data.
You might want to change that... | priority | update your bio your bio which is now live on the site reads betty has been doing this too long she also ported the data you might want to change that | 1 |
628,072 | 19,974,919,871 | IssuesEvent | 2022-01-29 00:52:10 | rstudio/gt | https://api.github.com/repos/rstudio/gt | closed | Option to set locale globally | Difficulty: [2] Intermediate Effort: [3] High Priority: ♨︎ Critical Type: ★ Enhancement | ### Setting locale for fmt_\* globally
Im using gt increasingly in html reports in Europe, therefore I would like to set the locale globally, for example via an option that switches every standard gt fmt_\* to use the locale specified in the option.
That would make (working with) gt code in non-US locales less verbose and reduces the need to use purrr::partial
| 1.0 | Option to set locale globally - ### Setting locale for fmt_\* globally
Im using gt increasingly in html reports in Europe, therefore I would like to set the locale globally, for example via an option that switches every standard gt fmt_\* to use the locale specified in the option.
That would make (working with) gt code in non-US locales less verbose and reduces the need to use purrr::partial
| priority | option to set locale globally setting locale for fmt globally im using gt increasingly in html reports in europe therefore i would like to set the locale globally for example via an option that switches every standard gt fmt to use the locale specified in the option that would make working with gt code in non us locales less verbose and reduces the need to use purrr partial | 1 |
528,599 | 15,370,530,770 | IssuesEvent | 2021-03-02 08:55:35 | Mobsya/aseba | https://api.github.com/repos/Mobsya/aseba | opened | ThymioSuite on Big Sur do not seen robots, problem of Discovery | Mac OS specific Thymio Device Manager bug high priority | Some user reported that on Big Sur (some on 10.15.7), robot are not seen. Lot of investigation was made with users and apple around a correct packaging and Notarization. That's resolved the problem for some user because TDM was blocked by gatekeeper.
Finally bug remains with some Big Sur user where the package was not the problem. Finally we discovered that TDM cannot show itself with the Discovery process (Bonjour) a error came back from the library with "bad parameters". [https://github.com/Mobsya/aseba/blob/6a557425761f494f2536ba725889acad54840215/aseba/thymio-device-manager/aseba_node_registery.cpp#L138](url).
This behaviour has been seen on two user computer but cannot be reproduced.
| 1.0 | ThymioSuite on Big Sur do not seen robots, problem of Discovery - Some user reported that on Big Sur (some on 10.15.7), robot are not seen. Lot of investigation was made with users and apple around a correct packaging and Notarization. That's resolved the problem for some user because TDM was blocked by gatekeeper.
Finally bug remains with some Big Sur user where the package was not the problem. Finally we discovered that TDM cannot show itself with the Discovery process (Bonjour) a error came back from the library with "bad parameters". [https://github.com/Mobsya/aseba/blob/6a557425761f494f2536ba725889acad54840215/aseba/thymio-device-manager/aseba_node_registery.cpp#L138](url).
This behaviour has been seen on two user computer but cannot be reproduced.
| priority | thymiosuite on big sur do not seen robots problem of discovery some user reported that on big sur some on robot are not seen lot of investigation was made with users and apple around a correct packaging and notarization that s resolved the problem for some user because tdm was blocked by gatekeeper finally bug remains with some big sur user where the package was not the problem finally we discovered that tdm cannot show itself with the discovery process bonjour a error came back from the library with bad parameters url this behaviour has been seen on two user computer but cannot be reproduced | 1 |
533,233 | 15,586,878,098 | IssuesEvent | 2021-03-18 02:59:41 | AY2021S2-CS2103T-T12-3/tp | https://api.github.com/repos/AY2021S2-CS2103T-T12-3/tp | opened | Alias command is not found in autocomplete panel | priority.High | A possible reason is that **getAutoCompleteCommands()** is not pulling that command yet.
**Alias command as found in UG.**

**Alias command not found.**

| 1.0 | Alias command is not found in autocomplete panel - A possible reason is that **getAutoCompleteCommands()** is not pulling that command yet.
**Alias command as found in UG.**

**Alias command not found.**

| priority | alias command is not found in autocomplete panel a possible reason is that getautocompletecommands is not pulling that command yet alias command as found in ug alias command not found | 1 |
478,479 | 13,780,041,736 | IssuesEvent | 2020-10-08 14:28:05 | carbon-design-system/ibm-dotcom-library | https://api.github.com/repos/carbon-design-system/ibm-dotcom-library | closed | Web Component: Develop Table of Contents of the React version - Group 2 | Airtable Done dev package: web components priority: high | #### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
IBM.com Library developer
> I need to:
create the `Table of Contents`
> so that I can:
provide ibm.com adopter developers a web component version for every react version available in the ibm.com Library
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- Story within Storybook with corresponding knobs
- Utilize Carbon
- Create with Shadow DOM and Custom Elements standards
- **See the Epic for the Design and Functional specs information**
- [React canary environment](https://ibmdotcom-react-canary.mybluemix.net/?path=/docs/overview-getting-started--page)
- Prod QA testing issue (#3631)
#### Acceptance criteria
- [ ] Include README for the web component and corresponding styles
- [ ] Create Web Components styles in styles package
- [ ] No custom styles in web-components package
- [ ] Do not create knobs in Storybook that include JSON objects
- [ ] Break out Storybook stories into multiple variation stories, if applicable
- [ ] Create codesandbox example under `/packages/web-components/examples/codesandbox` and include in README
- [ ] Minimum 80% unit test coverage
- [ ] A comment is posted in the Prod QA issue, tagging Praveen when development is finished
| 1.0 | Web Component: Develop Table of Contents of the React version - Group 2 - #### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
IBM.com Library developer
> I need to:
create the `Table of Contents`
> so that I can:
provide ibm.com adopter developers a web component version for every react version available in the ibm.com Library
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- Story within Storybook with corresponding knobs
- Utilize Carbon
- Create with Shadow DOM and Custom Elements standards
- **See the Epic for the Design and Functional specs information**
- [React canary environment](https://ibmdotcom-react-canary.mybluemix.net/?path=/docs/overview-getting-started--page)
- Prod QA testing issue (#3631)
#### Acceptance criteria
- [ ] Include README for the web component and corresponding styles
- [ ] Create Web Components styles in styles package
- [ ] No custom styles in web-components package
- [ ] Do not create knobs in Storybook that include JSON objects
- [ ] Break out Storybook stories into multiple variation stories, if applicable
- [ ] Create codesandbox example under `/packages/web-components/examples/codesandbox` and include in README
- [ ] Minimum 80% unit test coverage
- [ ] A comment is posted in the Prod QA issue, tagging Praveen when development is finished
| priority | web component develop table of contents of the react version group user story as a ibm com library developer i need to create the table of contents so that i can provide ibm com adopter developers a web component version for every react version available in the ibm com library additional information story within storybook with corresponding knobs utilize carbon create with shadow dom and custom elements standards see the epic for the design and functional specs information prod qa testing issue acceptance criteria include readme for the web component and corresponding styles create web components styles in styles package no custom styles in web components package do not create knobs in storybook that include json objects break out storybook stories into multiple variation stories if applicable create codesandbox example under packages web components examples codesandbox and include in readme minimum unit test coverage a comment is posted in the prod qa issue tagging praveen when development is finished | 1 |
759,362 | 26,591,525,305 | IssuesEvent | 2023-01-23 09:06:52 | codelab-app/builder | https://api.github.com/repos/codelab-app/builder | closed | Proposal for component slots | priority: high | ## The Problem
The current component system works for only basic templating. For example, you can't create a useful layout component right now.
## The solution
We need what's the equivalent of slots in templating tools. In Vue they are called slots, in Rails this is done through partials, in Laravel you have component slots. And in React this functionality is filled mostly by render props or by passing components as props/in context.
## Implementation
I can imagine 2 ways to do this
### 1. The explicit way
Users explicitly define an API for their components, similar to how we have an api for Atoms props. For example:
<img width="962" alt="image" src="https://user-images.githubusercontent.com/57956282/187932904-ee36b004-bbb9-4671-9119-4faf59705218.png">
For slots we can use existing types, like RenderPropsType, ReactNodeType, ElementType.
This api serves as the place of truth for defining the inputs that a component takes.
The benefit of this is that that's not only applicable for slots, but we can also assign other props to the component, like strings, numbers, etc.
We use this interface to generate a form for the component, just like we do for atoms.
The next part is to be able to assign this slot to a particular element
One way to do that is to bind it to props. Say that we have a Div atom with this API
<img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187933534-e081da3a-0039-45b5-99b1-8e831d7adfdf.png">
Now we only need to connect `heroContent` from the Layout's API to the `children` of the Divs API. The easiest way I imagine is to bind it as we bind global state variables.
<img width="1908" alt="image" src="https://user-images.githubusercontent.com/57956282/187933790-70b0db34-6773-4a60-9dda-604fc74fb4db.png">
This would require modifying the prop evaluating code to take into account the current component that the element is in and its props.
### The implicit way
We create a new Atom Type, for example named `Slot`.
The user creates a new element as usual and assigns it an Atom with type Slot:
<img width="1919" alt="image" src="https://user-images.githubusercontent.com/57956282/187934585-3e701b16-624b-48ab-aef7-989c6c449acc.png">
Then on the component instance, we render a form that has all of the elements inside it with atom type Slot and we allow the user to pick a Component to render for them.
<img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187935140-50fc67cc-ab9d-425f-8c5b-dd23bd7f4494.png">
The data from this form is stored on the component instance either as a separate field or as a special prop. It has the shape of a key-value object where the key is the id of the Slot-atomed element and the value is the component id to render. This is then used when evaluating the props to render the specific component instead of the slot-atomed element.
This approach seems simpler, but it's less flexible since the user can't define other component props other than slots.
#### Note
In both implementations, we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it.
Any thoughts or other ideas?
| 1.0 | Proposal for component slots - ## The Problem
The current component system works for only basic templating. For example, you can't create a useful layout component right now.
## The solution
We need what's the equivalent of slots in templating tools. In Vue they are called slots, in Rails this is done through partials, in Laravel you have component slots. And in React this functionality is filled mostly by render props or by passing components as props/in context.
## Implementation
I can imagine 2 ways to do this
### 1. The explicit way
Users explicitly define an API for their components, similar to how we have an api for Atoms props. For example:
<img width="962" alt="image" src="https://user-images.githubusercontent.com/57956282/187932904-ee36b004-bbb9-4671-9119-4faf59705218.png">
For slots we can use existing types, like RenderPropsType, ReactNodeType, ElementType.
This api serves as the place of truth for defining the inputs that a component takes.
The benefit of this is that that's not only applicable for slots, but we can also assign other props to the component, like strings, numbers, etc.
We use this interface to generate a form for the component, just like we do for atoms.
The next part is to be able to assign this slot to a particular element
One way to do that is to bind it to props. Say that we have a Div atom with this API
<img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187933534-e081da3a-0039-45b5-99b1-8e831d7adfdf.png">
Now we only need to connect `heroContent` from the Layout's API to the `children` of the Divs API. The easiest way I imagine is to bind it as we bind global state variables.
<img width="1908" alt="image" src="https://user-images.githubusercontent.com/57956282/187933790-70b0db34-6773-4a60-9dda-604fc74fb4db.png">
This would require modifying the prop evaluating code to take into account the current component that the element is in and its props.
### The implicit way
We create a new Atom Type, for example named `Slot`.
The user creates a new element as usual and assigns it an Atom with type Slot:
<img width="1919" alt="image" src="https://user-images.githubusercontent.com/57956282/187934585-3e701b16-624b-48ab-aef7-989c6c449acc.png">
Then on the component instance, we render a form that has all of the elements inside it with atom type Slot and we allow the user to pick a Component to render for them.
<img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187935140-50fc67cc-ab9d-425f-8c5b-dd23bd7f4494.png">
The data from this form is stored on the component instance either as a separate field or as a special prop. It has the shape of a key-value object where the key is the id of the Slot-atomed element and the value is the component id to render. This is then used when evaluating the props to render the specific component instead of the slot-atomed element.
This approach seems simpler, but it's less flexible since the user can't define other component props other than slots.
#### Note
In both implementations, we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it.
Any thoughts or other ideas?
| priority | proposal for component slots the problem the current component system works for only basic templating for example you can t create a useful layout component right now the solution we need what s the equivalent of slots in templating tools in vue they are called slots in rails this is done through partials in laravel you have component slots and in react this functionality is filled mostly by render props or by passing components as props in context implementation i can imagine ways to do this the explicit way users explicitly define an api for their components similar to how we have an api for atoms props for example img width alt image src for slots we can use existing types like renderpropstype reactnodetype elementtype this api serves as the place of truth for defining the inputs that a component takes the benefit of this is that that s not only applicable for slots but we can also assign other props to the component like strings numbers etc we use this interface to generate a form for the component just like we do for atoms the next part is to be able to assign this slot to a particular element one way to do that is to bind it to props say that we have a div atom with this api img width alt image src now we only need to connect herocontent from the layout s api to the children of the divs api the easiest way i imagine is to bind it as we bind global state variables img width alt image src this would require modifying the prop evaluating code to take into account the current component that the element is in and its props the implicit way we create a new atom type for example named slot the user creates a new element as usual and assigns it an atom with type slot img width alt image src then on the component instance we render a form that has all of the elements inside it with atom type slot and we allow the user to pick a component to render for them img width alt image src the data from this form is stored on the component instance either as a separate field or as a special prop it has the shape of a key value object where the key is the id of the slot atomed element and the value is the component id to render this is then used when evaluating the props to render the specific component instead of the slot atomed element this approach seems simpler but it s less flexible since the user can t define other component props other than slots note in both implementations we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it any thoughts or other ideas | 1 |
687,341 | 23,522,443,739 | IssuesEvent | 2022-08-19 07:34:34 | roq-trading/roq-issues | https://api.github.com/repos/roq-trading/roq-issues | closed | [roq-server] Using an invalid account causes crash | bug high priority support | Validation is done correctly and an `OrderAck` with the reject is prepared for sending.
When sending the `OrderAck`, the gateway needs to find the `account_id` and that's where it fails.
There's a low-level optimization that allows clients to only process updates where `account_id`'s (as known to the gateway) are used for filtering.
The client filtering is based on the subscription configuration and managed inside the `roq-client` library, i.e. not in the "user" code.
This is therefore a problem -- an `account_id` is needed for the filtering.
Somehow we need to allow for missing `account_id` as well. | 1.0 | [roq-server] Using an invalid account causes crash - Validation is done correctly and an `OrderAck` with the reject is prepared for sending.
When sending the `OrderAck`, the gateway needs to find the `account_id` and that's where it fails.
There's a low-level optimization that allows clients to only process updates where `account_id`'s (as known to the gateway) are used for filtering.
The client filtering is based on the subscription configuration and managed inside the `roq-client` library, i.e. not in the "user" code.
This is therefore a problem -- an `account_id` is needed for the filtering.
Somehow we need to allow for missing `account_id` as well. | priority | using an invalid account causes crash validation is done correctly and an orderack with the reject is prepared for sending when sending the orderack the gateway needs to find the account id and that s where it fails there s a low level optimization that allows clients to only process updates where account id s as known to the gateway are used for filtering the client filtering is based on the subscription configuration and managed inside the roq client library i e not in the user code this is therefore a problem an account id is needed for the filtering somehow we need to allow for missing account id as well | 1 |
393,638 | 11,622,676,081 | IssuesEvent | 2020-02-27 07:10:26 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Cannot commit edits to spatialite layers in QGIS 3.12 | Bug Data Provider High Priority Regression | <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
When I try to commit edits to a spatialite layer I get a yellow warning saying the following:
`Could not commit changes to layer Facilities`
`Errors: ERROR: 5 feature(s) not added.`
`Provider errors:`
`SQLite error: unknown cause`
`SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)`
and a red warning saying the following:
`Layer Facilities: SQLite error: unknown cause SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)`
This prevents me from making any edits to spatialite layers.
**How to Reproduce**
1. Create a new spatialite database and layer
2. Enable editing, add features, click `save edits`
3. See error(s)
**QGIS and OS versions**
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
QGIS version
3.12.0-București
QGIS code revision
cd141490ec
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
3.0.4
Running against GDAL/OGR
3.0.4
Compiled against GEOS
3.8.0-CAPI-1.13.1
Running against GEOS
3.8.0-CAPI-1.13.1
Compiled against SQLite
3.29.0
Running against SQLite
3.29.0
PostgreSQL Client Version
11.5
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
6.3.1
Running against PROJ
Rel. 6.3.1, February 10th, 2020
OS Version
Windows 10 (10.0)
**Test Project and Database**
[test_project_and_db.zip](https://github.com/qgis/QGIS/files/4256866/test_project_and_db.zip)
| 1.0 | Cannot commit edits to spatialite layers in QGIS 3.12 - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
When I try to commit edits to a spatialite layer I get a yellow warning saying the following:
`Could not commit changes to layer Facilities`
`Errors: ERROR: 5 feature(s) not added.`
`Provider errors:`
`SQLite error: unknown cause`
`SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)`
and a red warning saying the following:
`Layer Facilities: SQLite error: unknown cause SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)`
This prevents me from making any edits to spatialite layers.
**How to Reproduce**
1. Create a new spatialite database and layer
2. Enable editing, add features, click `save edits`
3. See error(s)
**QGIS and OS versions**
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
QGIS version
3.12.0-București
QGIS code revision
cd141490ec
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
3.0.4
Running against GDAL/OGR
3.0.4
Compiled against GEOS
3.8.0-CAPI-1.13.1
Running against GEOS
3.8.0-CAPI-1.13.1
Compiled against SQLite
3.29.0
Running against SQLite
3.29.0
PostgreSQL Client Version
11.5
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
6.3.1
Running against PROJ
Rel. 6.3.1, February 10th, 2020
OS Version
Windows 10 (10.0)
**Test Project and Database**
[test_project_and_db.zip](https://github.com/qgis/QGIS/files/4256866/test_project_and_db.zip)
| priority | cannot commit edits to spatialite layers in qgis bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when i try to commit edits to a spatialite layer i get a yellow warning saying the following could not commit changes to layer facilities errors error feature s not added provider errors sqlite error unknown cause sql insert into facilities geometry type comment values geomfromwkb and a red warning saying the following layer facilities sqlite error unknown cause sql insert into facilities geometry type comment values geomfromwkb this prevents me from making any edits to spatialite layers how to reproduce create a new spatialite database and layer enable editing add features click save edits see error s qgis and os versions qgis version bucurești qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel february os version windows test project and database | 1 |
444,298 | 12,809,386,262 | IssuesEvent | 2020-07-03 15:32:04 | cds-snc/covid-shield-mobile | https://api.github.com/repos/cds-snc/covid-shield-mobile | reopened | "Share your random codes" notification - Screen in background on infinite load state | bug high priority | Environment :
Pixel 3XL, Android v9 (BrowserStack), version v14
Scenario :
- Go to Enter you code screen from the menu
- Enter a 8-digit code and submit
- Click on "Agree" on the upload code screen
Expected :
A notification about "share you random codes" is displayed on top of the upload code screen
Issue :
The screen in the background of the notification is on an infinite load state

| 1.0 | "Share your random codes" notification - Screen in background on infinite load state - Environment :
Pixel 3XL, Android v9 (BrowserStack), version v14
Scenario :
- Go to Enter you code screen from the menu
- Enter a 8-digit code and submit
- Click on "Agree" on the upload code screen
Expected :
A notification about "share you random codes" is displayed on top of the upload code screen
Issue :
The screen in the background of the notification is on an infinite load state

| priority | share your random codes notification screen in background on infinite load state environment pixel android browserstack version scenario go to enter you code screen from the menu enter a digit code and submit click on agree on the upload code screen expected a notification about share you random codes is displayed on top of the upload code screen issue the screen in the background of the notification is on an infinite load state | 1 |
354,862 | 10,573,840,280 | IssuesEvent | 2019-10-07 12:53:28 | AY1920S1-CS2103T-W11-1/main | https://api.github.com/repos/AY1920S1-CS2103T-W11-1/main | opened | As a user, I want to mark tasks as done/undone | priority.High status.Ongoing type.Story | ... so I can manage my progress in the training plans. | 1.0 | As a user, I want to mark tasks as done/undone - ... so I can manage my progress in the training plans. | priority | as a user i want to mark tasks as done undone so i can manage my progress in the training plans | 1 |
394,718 | 11,647,939,654 | IssuesEvent | 2020-03-01 17:53:29 | Rammelkast/AntiCheatReloaded | https://api.github.com/repos/Rammelkast/AntiCheatReloaded | closed | Speed bypass | bypass help wanted high priority | Video:
https://youtu.be/H9u6GS0jNes
Code:
```
package AppleClient.modules.movement;
import org.lwjgl.input.Keyboard;
import AppleClient.events.EventTarget;
import AppleClient.events.events.EventMove;
import AppleClient.events.events.EventTick;
import AppleClient.modules.Module;
import net.minecraft.util.MovementInput;
public class MemeSpeed extends Module
{
public MemeSpeed() {
super("ACRSpeed", Keyboard.KEY_Z, 7733063, Category.MOVEMENT, "memes", new String[] {"2fast4uboi"}, true);
}
@Override
public void onEnable() {
if(mc.thePlayer != null) {
}
super.onEnable();
}
public void setSpeed(double speed) {
final MovementInput movementInput = mc.thePlayer.movementInput;
float forward = movementInput.moveForward;
float strafe = movementInput.moveStrafe;
float yaw = mc.thePlayer.rotationYaw;
if (forward == 0.0f && strafe == 0.0f) {
mc.thePlayer.motionX = 0.0;
mc.thePlayer.motionZ = 0.0;
}
else if (forward != 0.0f) {
if (strafe >= 1.0f) {
yaw += ((forward > 0.0f) ? -45 : 45);
strafe = 0.0f;
}
else if (strafe <= -1.0f) {
yaw += ((forward > 0.0f) ? 45 : -45);
strafe = 0.0f;
}
if (forward > 0.0f) {
forward = 1.0f;
}
else if (forward < 0.0f) {
forward = -1.0f;
}
}
final double mx = Math.cos(Math.toRadians(yaw + 90.0f));
final double mz = Math.sin(Math.toRadians(yaw + 90.0f));
mc.thePlayer.motionX = forward * speed * mx + strafe * speed * mz;
mc.thePlayer.motionZ = forward * speed * mz - strafe * speed * mx;
if (forward == 0.0f && strafe == 0.0f) {
mc.thePlayer.motionX = 0.0;
mc.thePlayer.motionZ = 0.0;
}
}
@EventTarget
private void onUpdate(EventTick event) {
if((mc.thePlayer.moveForward != 0.0D || mc.thePlayer.moveStrafing != 0.0D) && mc.thePlayer.onGround) {
mc.thePlayer.motionY = 0.4D;
}
}
@EventTarget
public void onMove(EventMove event) {
//heres how it changes between speeds to bypass.
boolean hack = mc.thePlayer.ticksExisted % 2 == 0;
MemeSpeed.this.setSpeed(hack ? 0.06D : 1.6D);
}
}
```
What it does is change between a slow speed, then goes fast for a short amount of time, before the anticheat can recognize it. | 1.0 | Speed bypass - Video:
https://youtu.be/H9u6GS0jNes
Code:
```
package AppleClient.modules.movement;
import org.lwjgl.input.Keyboard;
import AppleClient.events.EventTarget;
import AppleClient.events.events.EventMove;
import AppleClient.events.events.EventTick;
import AppleClient.modules.Module;
import net.minecraft.util.MovementInput;
public class MemeSpeed extends Module
{
public MemeSpeed() {
super("ACRSpeed", Keyboard.KEY_Z, 7733063, Category.MOVEMENT, "memes", new String[] {"2fast4uboi"}, true);
}
@Override
public void onEnable() {
if(mc.thePlayer != null) {
}
super.onEnable();
}
public void setSpeed(double speed) {
final MovementInput movementInput = mc.thePlayer.movementInput;
float forward = movementInput.moveForward;
float strafe = movementInput.moveStrafe;
float yaw = mc.thePlayer.rotationYaw;
if (forward == 0.0f && strafe == 0.0f) {
mc.thePlayer.motionX = 0.0;
mc.thePlayer.motionZ = 0.0;
}
else if (forward != 0.0f) {
if (strafe >= 1.0f) {
yaw += ((forward > 0.0f) ? -45 : 45);
strafe = 0.0f;
}
else if (strafe <= -1.0f) {
yaw += ((forward > 0.0f) ? 45 : -45);
strafe = 0.0f;
}
if (forward > 0.0f) {
forward = 1.0f;
}
else if (forward < 0.0f) {
forward = -1.0f;
}
}
final double mx = Math.cos(Math.toRadians(yaw + 90.0f));
final double mz = Math.sin(Math.toRadians(yaw + 90.0f));
mc.thePlayer.motionX = forward * speed * mx + strafe * speed * mz;
mc.thePlayer.motionZ = forward * speed * mz - strafe * speed * mx;
if (forward == 0.0f && strafe == 0.0f) {
mc.thePlayer.motionX = 0.0;
mc.thePlayer.motionZ = 0.0;
}
}
@EventTarget
private void onUpdate(EventTick event) {
if((mc.thePlayer.moveForward != 0.0D || mc.thePlayer.moveStrafing != 0.0D) && mc.thePlayer.onGround) {
mc.thePlayer.motionY = 0.4D;
}
}
@EventTarget
public void onMove(EventMove event) {
//heres how it changes between speeds to bypass.
boolean hack = mc.thePlayer.ticksExisted % 2 == 0;
MemeSpeed.this.setSpeed(hack ? 0.06D : 1.6D);
}
}
```
What it does is change between a slow speed, then goes fast for a short amount of time, before the anticheat can recognize it. | priority | speed bypass video code package appleclient modules movement import org lwjgl input keyboard import appleclient events eventtarget import appleclient events events eventmove import appleclient events events eventtick import appleclient modules module import net minecraft util movementinput public class memespeed extends module public memespeed super acrspeed keyboard key z category movement memes new string true override public void onenable if mc theplayer null super onenable public void setspeed double speed final movementinput movementinput mc theplayer movementinput float forward movementinput moveforward float strafe movementinput movestrafe float yaw mc theplayer rotationyaw if forward strafe mc theplayer motionx mc theplayer motionz else if forward if strafe yaw forward strafe else if strafe yaw forward strafe if forward forward else if forward forward final double mx math cos math toradians yaw final double mz math sin math toradians yaw mc theplayer motionx forward speed mx strafe speed mz mc theplayer motionz forward speed mz strafe speed mx if forward strafe mc theplayer motionx mc theplayer motionz eventtarget private void onupdate eventtick event if mc theplayer moveforward mc theplayer movestrafing mc theplayer onground mc theplayer motiony eventtarget public void onmove eventmove event heres how it changes between speeds to bypass boolean hack mc theplayer ticksexisted memespeed this setspeed hack what it does is change between a slow speed then goes fast for a short amount of time before the anticheat can recognize it | 1 |
153,971 | 5,906,750,331 | IssuesEvent | 2017-05-19 15:52:52 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | [Request] Add jmespath | High Priority in progress Library - Request to Add/Update | **Library name:** jmespath
**Git repository url:** https://github.com/jmespath/jmespath.js
**npm package name or url** (if there is one): https://www.npmjs.com/package/jmespath
**License (List them all if it's multiple):** Apache License, Version 2.0
**Official homepage:** http://jmespath.org/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first.
We encourage you to add a library via sending pull request,
it'll be faster than just opening a request issue,
since there are tons of issues, please wait with patience,
and please don't forget to read the guidelines for contributing, thanks!!
| 1.0 | [Request] Add jmespath - **Library name:** jmespath
**Git repository url:** https://github.com/jmespath/jmespath.js
**npm package name or url** (if there is one): https://www.npmjs.com/package/jmespath
**License (List them all if it's multiple):** Apache License, Version 2.0
**Official homepage:** http://jmespath.org/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first.
We encourage you to add a library via sending pull request,
it'll be faster than just opening a request issue,
since there are tons of issues, please wait with patience,
and please don't forget to read the guidelines for contributing, thanks!!
| priority | add jmespath library name jmespath git repository url npm package name or url if there is one license list them all if it s multiple apache license version official homepage wanna say something leave message here notes from cdnjs maintainer please read the and document first we encourage you to add a library via sending pull request it ll be faster than just opening a request issue since there are tons of issues please wait with patience and please don t forget to read the guidelines for contributing thanks | 1 |
161,676 | 6,132,993,337 | IssuesEvent | 2017-06-25 09:21:29 | play2-maven-plugin/play2-maven-plugin | https://api.github.com/repos/play2-maven-plugin/play2-maven-plugin | closed | Upgrade Play! version from 2.6.0-RC2 to 2.6.0 | Component-Maven-Plugin Component-Provider26 Priority-High Type-Task | Upgrade:
- `play` version from `2.6.0-RC2` to `2.6.0`
- `twirl` version from `1.3.0` to `1.3.2`
- `ebean-agent` version from `10.1.7` to `10.3.1` (used in `play-ebean` version `4.0.2`)
Upgrade in documentation and test projects:
- `akka-*` dependencies versions to `2.5.3`
- `play-ebean` version to `4.0.2`
- `play-json` version to `2.6.0`
- `play-slick` version to `3.0.0`
- `hibernate-entitymanager` version to `5.2.10.Final`
- `scalatestplus-play` version to `3.0.0`
| 1.0 | Upgrade Play! version from 2.6.0-RC2 to 2.6.0 - Upgrade:
- `play` version from `2.6.0-RC2` to `2.6.0`
- `twirl` version from `1.3.0` to `1.3.2`
- `ebean-agent` version from `10.1.7` to `10.3.1` (used in `play-ebean` version `4.0.2`)
Upgrade in documentation and test projects:
- `akka-*` dependencies versions to `2.5.3`
- `play-ebean` version to `4.0.2`
- `play-json` version to `2.6.0`
- `play-slick` version to `3.0.0`
- `hibernate-entitymanager` version to `5.2.10.Final`
- `scalatestplus-play` version to `3.0.0`
| priority | upgrade play version from to upgrade play version from to twirl version from to ebean agent version from to used in play ebean version upgrade in documentation and test projects akka dependencies versions to play ebean version to play json version to play slick version to hibernate entitymanager version to final scalatestplus play version to | 1 |
681,150 | 23,298,773,054 | IssuesEvent | 2022-08-07 01:45:32 | zot4plan/Zot4Plan | https://api.github.com/repos/zot4plan/Zot4Plan | opened | INS-15 Tracking taken GE courses | Priority: high Type: feature request | **Story**
Users want to know how many and which GE courses have been taken.
**Requirement**
1. Each GE category has a badge which display the number of taken courses
2. Show list of taken courses when hovering the badge
| 1.0 | INS-15 Tracking taken GE courses - **Story**
Users want to know how many and which GE courses have been taken.
**Requirement**
1. Each GE category has a badge which display the number of taken courses
2. Show list of taken courses when hovering the badge
| priority | ins tracking taken ge courses story users want to know how many and which ge courses have been taken requirement each ge category has a badge which display the number of taken courses show list of taken courses when hovering the badge | 1 |
449,301 | 12,966,629,447 | IssuesEvent | 2020-07-21 01:08:52 | dhowe/Website | https://api.github.com/repos/dhowe/Website | closed | Ready new design to go live | priority: high | lets get the new website design pushed live:
- [x] check resolution of all images and replace any pixelated/low-res
- [x] check display in all mobile | 1.0 | Ready new design to go live - lets get the new website design pushed live:
- [x] check resolution of all images and replace any pixelated/low-res
- [x] check display in all mobile | priority | ready new design to go live lets get the new website design pushed live check resolution of all images and replace any pixelated low res check display in all mobile | 1 |
374,480 | 11,091,183,010 | IssuesEvent | 2019-12-15 10:28:58 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | excise.wb.gov.in - see bug description | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal | <!-- @browser: Firefox 72.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx
**Browser / Version**: Firefox 72.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: the page whichi was working dssapeared and new page started
**Steps to Reproduce**:
i lost the page which iwas working
[](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191210230245</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
[{'level': 'error', 'log': ["SyntaxError: expected expression, got '}'"], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/CSS/menu.js', 'pos': '216:10'}, {'level': 'warn', 'log': ['This page uses the non standard property zoom. Consider using calc() in the relevant property values, or using transform along with transform-origin: 0 0.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '9:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '10:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '11:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '12:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/js/demos.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '21:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/themeroller/themeswitchertool/.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '22:1'}, {'level': 'error', 'log': ['TypeError: $(...).simplyScroll is not a function'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '153:28'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'error', 'log': ['ReferenceError: theme_path is not defined'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/nic9e1a.js?oo10gf', 'pos': '221:5'}, {'level': 'error', 'log': ['TypeError: this.mqo is null'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Javascript/marquee.js', 'pos': '14:202'}]
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | excise.wb.gov.in - see bug description - <!-- @browser: Firefox 72.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx
**Browser / Version**: Firefox 72.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: the page whichi was working dssapeared and new page started
**Steps to Reproduce**:
i lost the page which iwas working
[](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191210230245</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
[{'level': 'error', 'log': ["SyntaxError: expected expression, got '}'"], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/CSS/menu.js', 'pos': '216:10'}, {'level': 'warn', 'log': ['This page uses the non standard property zoom. Consider using calc() in the relevant property values, or using transform along with transform-origin: 0 0.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '9:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '10:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '11:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '12:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/js/demos.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '21:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/themeroller/themeswitchertool/.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '22:1'}, {'level': 'error', 'log': ['TypeError: $(...).simplyScroll is not a function'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '153:28'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'error', 'log': ['ReferenceError: theme_path is not defined'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/nic9e1a.js?oo10gf', 'pos': '221:5'}, {'level': 'error', 'log': ['TypeError: this.mqo is null'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Javascript/marquee.js', 'pos': '14:202'}]
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | excise wb gov in see bug description url browser version firefox operating system windows tested another browser unknown problem type something else description the page whichi was working dssapeared and new page started steps to reproduce i lost the page which iwas working browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false console messages uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level error log uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level error log uri pos level error log uri pos from with ❤️ | 1 |
545,354 | 15,948,800,697 | IssuesEvent | 2021-04-15 06:29:08 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.3 release-184] Can't migrate White tiger cycle 10 | Category: Tech Priority: High Squad: Wild Turkey Status: Fixed Type: Bug | https://drive.google.com/file/d/1vBiHPylX02eiHy53IfosJ8ZJSu-62x9p/view?usp=sharing
[log_210408120622.log](https://github.com/StrangeLoopGames/EcoIssues/files/6277913/log_210408120622.log)
```
Failed to start the server. Exception was Exception: NullReferenceException
Message:Object reference not set to an instance of an object.
Source:Eco.Gameplay
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Items.WorkOrder.get_Product()
at Eco.Gameplay.Items.WorkOrder.UILinkContent()
at Eco.Gameplay.Systems.TextLinks.UILinkExtensions.UILink(ILinkable linkable)
at Eco.Gameplay.Utils.SimpleEntry.get_MarkedUpName()
at Eco.Core.Utils.PropertyScanning.PropertyScanner.SetupValidity(ScanScope scope, ScanSettings settings, ScanResults results)
at Eco.Core.Utils.PropertyScanning.PropertyScanner.ScanObj(Object root, ScanSettings settings)
at Eco.Core.Utils.PropertyScanning.PropertyScanner.Scan(ScanSettings settings)
at Eco.Gameplay.Utils.SimpleEntry.Initialize()
at Eco.Gameplay.Economy.WorkParties.WorkParty.Initialize()
at Eco.Core.Systems.Registrar.Initialize()
at Eco.Core.Systems.Registrars.Init()
at Eco.Core.Utils.Initializer.Initialize()
at Eco.Server.PluginManager.InitializeAsync(StartupInfo startupInfo)
at Eco.Server.Startup.StartAsync(StartupInfo startupInfo)
``` | 1.0 | [0.9.3 release-184] Can't migrate White tiger cycle 10 - https://drive.google.com/file/d/1vBiHPylX02eiHy53IfosJ8ZJSu-62x9p/view?usp=sharing
[log_210408120622.log](https://github.com/StrangeLoopGames/EcoIssues/files/6277913/log_210408120622.log)
```
Failed to start the server. Exception was Exception: NullReferenceException
Message:Object reference not set to an instance of an object.
Source:Eco.Gameplay
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Items.WorkOrder.get_Product()
at Eco.Gameplay.Items.WorkOrder.UILinkContent()
at Eco.Gameplay.Systems.TextLinks.UILinkExtensions.UILink(ILinkable linkable)
at Eco.Gameplay.Utils.SimpleEntry.get_MarkedUpName()
at Eco.Core.Utils.PropertyScanning.PropertyScanner.SetupValidity(ScanScope scope, ScanSettings settings, ScanResults results)
at Eco.Core.Utils.PropertyScanning.PropertyScanner.ScanObj(Object root, ScanSettings settings)
at Eco.Core.Utils.PropertyScanning.PropertyScanner.Scan(ScanSettings settings)
at Eco.Gameplay.Utils.SimpleEntry.Initialize()
at Eco.Gameplay.Economy.WorkParties.WorkParty.Initialize()
at Eco.Core.Systems.Registrar.Initialize()
at Eco.Core.Systems.Registrars.Init()
at Eco.Core.Utils.Initializer.Initialize()
at Eco.Server.PluginManager.InitializeAsync(StartupInfo startupInfo)
at Eco.Server.Startup.StartAsync(StartupInfo startupInfo)
``` | priority | can t migrate white tiger cycle failed to start the server exception was exception nullreferenceexception message object reference not set to an instance of an object source eco gameplay system nullreferenceexception object reference not set to an instance of an object at eco gameplay items workorder get product at eco gameplay items workorder uilinkcontent at eco gameplay systems textlinks uilinkextensions uilink ilinkable linkable at eco gameplay utils simpleentry get markedupname at eco core utils propertyscanning propertyscanner setupvalidity scanscope scope scansettings settings scanresults results at eco core utils propertyscanning propertyscanner scanobj object root scansettings settings at eco core utils propertyscanning propertyscanner scan scansettings settings at eco gameplay utils simpleentry initialize at eco gameplay economy workparties workparty initialize at eco core systems registrar initialize at eco core systems registrars init at eco core utils initializer initialize at eco server pluginmanager initializeasync startupinfo startupinfo at eco server startup startasync startupinfo startupinfo | 1 |
585,670 | 17,514,187,860 | IssuesEvent | 2021-08-11 03:42:12 | encorelab/ck-board | https://api.github.com/repos/encorelab/ck-board | closed | Move post modification buttons on post | bug high priority | Move edit and delete buttons on the post objects themselves, instead of having them on the side toolbar. | 1.0 | Move post modification buttons on post - Move edit and delete buttons on the post objects themselves, instead of having them on the side toolbar. | priority | move post modification buttons on post move edit and delete buttons on the post objects themselves instead of having them on the side toolbar | 1 |
582,870 | 17,372,779,126 | IssuesEvent | 2021-07-30 16:07:18 | MaibornWolff/codecharta | https://api.github.com/repos/MaibornWolff/codecharta | opened | Legend improvements | difficulty:low feature javascript pr-visualization priority:high | # Feature request
## Description
As a user, I want the legend to be as useful as possible so that I have all necessary information without any information that is not needed or being distracted by anything.
## Acceptance criteria
- Change the background color of the legend to be identical to the background of the map (it should be 100% transparent)
- Remove the bounding box of the legend so that the content is visible without anything distracting the user
- Remove "Outgoing" and "Incoming" Edge from the legend in case there is no such metric available
- Add all used metrics to the legend and add a description in case we know what the abbreviation stands for


## Development notes (optional Task Breakdown)
- [ ] Remove the bounding box and changing the background color
- [ ] Remove "outgoing" and "incoming" edges from the legend, if not applicable
- [ ] Add all used metrics to the legend
- [ ] Add a list of known metric descriptions to the frontend and show a translation next to the entry
| 1.0 | Legend improvements - # Feature request
## Description
As a user, I want the legend to be as useful as possible so that I have all necessary information without any information that is not needed or being distracted by anything.
## Acceptance criteria
- Change the background color of the legend to be identical to the background of the map (it should be 100% transparent)
- Remove the bounding box of the legend so that the content is visible without anything distracting the user
- Remove "Outgoing" and "Incoming" Edge from the legend in case there is no such metric available
- Add all used metrics to the legend and add a description in case we know what the abbreviation stands for


## Development notes (optional Task Breakdown)
- [ ] Remove the bounding box and changing the background color
- [ ] Remove "outgoing" and "incoming" edges from the legend, if not applicable
- [ ] Add all used metrics to the legend
- [ ] Add a list of known metric descriptions to the frontend and show a translation next to the entry
| priority | legend improvements feature request description as a user i want the legend to be as useful as possible so that i have all necessary information without any information that is not needed or being distracted by anything acceptance criteria change the background color of the legend to be identical to the background of the map it should be transparent remove the bounding box of the legend so that the content is visible without anything distracting the user remove outgoing and incoming edge from the legend in case there is no such metric available add all used metrics to the legend and add a description in case we know what the abbreviation stands for development notes optional task breakdown remove the bounding box and changing the background color remove outgoing and incoming edges from the legend if not applicable add all used metrics to the legend add a list of known metric descriptions to the frontend and show a translation next to the entry | 1 |
563,454 | 16,685,275,210 | IssuesEvent | 2021-06-08 07:22:38 | nlpsandbox/nlpsandbox.io | https://api.github.com/repos/nlpsandbox/nlpsandbox.io | closed | Create multi-site compatible leaderboard | Priority: High | - [x] Select the columns in the leaderboard (@tschaffter )
- [ ] Implement the leaderboard (@thomasyu888 ) | 1.0 | Create multi-site compatible leaderboard - - [x] Select the columns in the leaderboard (@tschaffter )
- [ ] Implement the leaderboard (@thomasyu888 ) | priority | create multi site compatible leaderboard select the columns in the leaderboard tschaffter implement the leaderboard | 1 |
794,481 | 28,037,790,952 | IssuesEvent | 2023-03-28 16:11:17 | asastats/channel | https://api.github.com/repos/asastats/channel | opened | ASA Stats is displaying a different swap price for Tinyman | bug high priority | Typing the same ASA amount in Tinyman gives a different swap value. | 1.0 | ASA Stats is displaying a different swap price for Tinyman - Typing the same ASA amount in Tinyman gives a different swap value. | priority | asa stats is displaying a different swap price for tinyman typing the same asa amount in tinyman gives a different swap value | 1 |
524,799 | 15,223,629,826 | IssuesEvent | 2021-02-18 03:09:21 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Data copy from CPU to GPU use default stream in nightly version. | high priority module: cuda oncall: distributed | ## 🐛 Bug
After https://github.com/pytorch/pytorch/pull/46304, data copy per iteration from cpu to gpu can't overlap with last iteration's computation/communication.
There is my test result which shows different behaviors by literally same code.
nightly:

1.5.1 release:

## To Reproduce
Steps to reproduce the behavior:
1. Run the same code in nightly build and 1.5.1 release build. Use nsys profiling it.
My test code is listed as below, you can use it do some tests. But I think the reason that incurs different behavior is clear. Is it a new feature? If so, I might not think it's a good one.
```
import os
import time
import sys
import random
import traceback
import numpy as np
import torch
import torch.nn as nn
import torch.nn.parallel
import torch.distributed as dist
import torch.optim
import torch.utils.data
import torch.utils.data.distributed
import torchvision.transforms as transforms
import torchvision.datasets as datasets
import torchvision.models as models
from torch.multiprocessing import Pool, Process
import argparse
class AverageMeter(object):
"""Computes and stores the average and current value"""
def __init__(self):
self.reset()
def reset(self):
self.val = 0
self.avg = 0
self.sum = 0
self.count = 0
def update(self, val, n=1):
self.val = val
self.sum += val * n
self.count += n
self.avg = self.sum / self.count
def accuracy(output, target, topk=(1,)):
"""Computes the precision@k for the specified values of k"""
with torch.no_grad():
maxk = max(topk)
batch_size = target.size(0)
_, pred = output.topk(maxk, 1, True, True)
pred = pred.t()
correct = pred.eq(target.view(1, -1).expand_as(pred))
res = []
for k in topk:
correct_k = correct[:k].contiguous().view(-1).float().sum(0, keepdim=True)
res.append(correct_k.mul_(100.0 / batch_size))
return res
def train(train_loader, model, criterion, optimizer, epoch, batch_size):
batch_time = AverageMeter()
data_time = AverageMeter()
losses = AverageMeter()
top1 = AverageMeter()
top5 = AverageMeter()
local_rank = os.environ['LOCAL_RANK']
# switch to train mode
model.train()
end = time.time()
for i, (input, target) in enumerate(train_loader):
# measure data loading time
data_time.update(time.time() - end)
# Create non_blocking tensors for distributed training
# input = input.cuda(non_blocking=True)
target = target.cuda(non_blocking=True)
# compute output
output = model(input)
loss = criterion(output, target)
# measure accuracy and record loss
prec1, prec5 = accuracy(output, target, topk=(1, 5))
losses.update(loss.item(), input.size(0))
top1.update(prec1[0], input.size(0))
top5.update(prec5[0], input.size(0))
# compute gradients in a backward pass
optimizer.zero_grad()
loss.backward()
# Call step of optimizer to update model params
optimizer.step()
# measure elapsed time
if i >= 10:
batch_time.update(time.time() - end)
end = time.time()
if local_rank == '0' and i % 10 == 0 and i > 10:
# if local_rank == '0' and i > 0:
print('Epoch: [{0}][{1}/{2}]\t'
'Speed {speed_now:.3f} ({speed_avg:.3f})\t'
'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t'
'Data {data_time.val:.3f} ({data_time.avg:.3f})\t'
'Loss {loss.val:.4f} ({loss.avg:.4f})\t'
'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t'
'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format(
epoch, i, len(train_loader), batch_time=batch_time,
speed_now=batch_size/batch_time.val, speed_avg=batch_size/batch_time.avg,
data_time=data_time, loss=losses, top1=top1, top5=top5))
def adjust_learning_rate(initial_lr, optimizer, epoch):
"""Sets the learning rate to the initial LR decayed by 10 every 30 epochs"""
lr = initial_lr * (0.1 ** (epoch // 30))
for param_group in optimizer.param_groups:
param_group['lr'] = lr
def validate(val_loader, model, criterion):
batch_time = AverageMeter()
losses = AverageMeter()
top1 = AverageMeter()
top5 = AverageMeter()
# switch to evaluate mode
model.eval()
with torch.no_grad():
end = time.time()
for i, (input, target) in enumerate(val_loader):
input = input.cuda(non_blocking=True)
target = target.cuda(non_blocking=True)
# compute output
output = model(input)
loss = criterion(output, target)
# measure accuracy and record loss
prec1, prec5 = accuracy(output, target, topk=(1, 5))
losses.update(loss.item(), input.size(0))
top1.update(prec1[0], input.size(0))
top5.update(prec5[0], input.size(0))
# measure elapsed time
torch.cuda.synchronize()
batch_time.update(time.time() - end)
end = time.time()
if i % 100 == 0:
print('Test: [{0}/{1}]\t'
'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t'
'Loss {loss.val:.4f} ({loss.avg:.4f})\t'
'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t'
'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format(
i, len(val_loader), batch_time=batch_time, loss=losses,
top1=top1, top5=top5))
print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}'
.format(top1=top1, top5=top5))
return top1.avg
def start(backend):
print("Collect Inputs...")
# Batch Size for training and testing
batch_size = 64
# Number of additional worker processes for dataloading
workers = 8
# Number of epochs to train for
num_epochs = 1
# Starting Learning Rate
starting_lr = 0.01
# Distributed backend type
# dist_backend = 'nccl'
# dist_backend = 'esl'
dist_backend = backend
# Url used to setup distributed training
dist.init_process_group(dist_backend, init_method='env://')
print("Wordsize: ", os.environ['WORLD_SIZE'])
local_rank = int(os.environ['LOCAL_RANK'])
print("Local_Rank: ", local_rank)
dp_device_ids = [local_rank]
torch.cuda.set_device(local_rank)
device = torch.device("cuda", local_rank)
print("Initialize Model...")
# Construct Model
model = models.vgg16(pretrained=False).to(device)
# Make model DistributedDataParallel
model = torch.nn.parallel.DistributedDataParallel(
model, device_ids=dp_device_ids, output_device=local_rank)
# define loss function (criterion) and optimizer
criterion = nn.CrossEntropyLoss().cuda()
optimizer = torch.optim.SGD(
model.parameters(), starting_lr, momentum=0.9, weight_decay=1e-4)
print("Initialize Dataloaders...")
# Define the transform for the data. Notice, we must resize to 224x224 with this dataset and model.
transform = transforms.Compose(
[transforms.Resize(224),
transforms.ToTensor(),
transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))])
# Initialize Datasets. STL10 will automatically download if not present
# trainset = datasets.STL10(
# root='/workspace/dataset', split='train', download=True, transform=transform)
# valset = datasets.STL10(root='/workspace/dataset',
# split='test', download=True, transform=transform)
trainset = datasets.FakeData(size=batch_size*int(os.environ['WORLD_SIZE'])*100, num_classes=1000, ,transform=transform)
valset = datasets.FakeData(size=1000, transform=transform)
# Create DistributedSampler to handle distributing the dataset across nodes when training
# This can only be called after torch.distributed.init_process_group is called
train_sampler = torch.utils.data.distributed.DistributedSampler(trainset)
# Create the Dataloaders to feed data to the training and validation steps
train_loader = torch.utils.data.DataLoader(trainset, batch_size=batch_size, shuffle=(train_sampler is None),
num_workers=workers, pin_memory=True, sampler=train_sampler)
val_loader = torch.utils.data.DataLoader(valset, batch_size=batch_size, shuffle=False, num_workers=workers,
pin_memory=True)
best_prec1 = 0
for epoch in range(num_epochs):
# Set epoch count for DistributedSampler
train_sampler.set_epoch(epoch)
# Adjust learning rate according to schedule
adjust_learning_rate(starting_lr, optimizer, epoch)
# train for one epoch
print("\nBegin Training Epoch {}".format(epoch + 1))
train(train_loader, model, criterion, optimizer, epoch, batch_size*int(os.environ['WORLD_SIZE']))
# # evaluate on validation set
# print("Begin Validation @ Epoch {}".format(epoch + 1))
# prec1 = validate(val_loader, model, criterion)
# # remember best prec@1 and save checkpoint if desired
# # is_best = prec1 > best_prec1
# best_prec1 = max(prec1, best_prec1)
# print("Epoch Summary: ")
# print("\tEpoch Accuracy: {}".format(prec1))
# print("\tBest Accuracy: {}".format(best_prec1))
def set_seed(self, seed=0):
random.seed(seed)
np.random.seed(seed)
torch.manual_seed(seed) #为CPU设置种子用于生成随机数,以使得结果是确定的
torch.cuda.manual_seed(seed) #为当前GPU设置随机种子;
torch.cuda.manual_seed_all(seed) #如果使用多个GPU,为所有的GPU设置种子。
torch.backends.cudnn.deterministic = True # CPU和GPU结果一致
torch.backends.cudnn.benchmark = False
if __name__ == '__main__':
# set_seed(42)
parser = argparse.ArgumentParser()
parser.add_argument("--local_rank", type=int)
args = parser.parse_args()
os.environ['LOCAL_RANK'] = str(args.local_rank)
backend = nccl
start(backend)
# usage: python -m torch.distributed.launch --nproc_per_node=NUM_GPUS_YOU_HAVE --nnodes=2 --node_rank=1 --master_addr="127.0.0.1" --master_port=1997 YOUR_TRAINING_SCRIPT.py (--arg1 --arg2 --arg3 and all other arguments of your training script)
```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
## Environment
```
PyTorch version: 1.5.1 │12:02:37 eth0 600B/s 0B/s
Is debug build: No │12:02:37 eth0 600B/s 0B/s
CUDA used to build PyTorch: 10.0 │12:02:37 eth0 1.17KB/s 14.02KB/s
│12:02:38 eth0 600B/s 0B/s
OS: Ubuntu 16.04.6 LTS │12:02:38 eth0 600B/s 0B/s
GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.12) 5.4.0 20160609 │12:02:38 eth0 600B/s 0B/s
CMake version: version 3.5.1 │12:02:38 eth0 600B/s 0B/s
│12:02:38 eth0 600B/s 0B/s
Python version: 3.7 │12:02:38 eth0 600B/s 0B/s
Is CUDA available: Yes │12:02:38 eth0 600B/s 0B/s
CUDA runtime version: 10.0.130 │12:02:38 eth0 600B/s 0B/s
GPU models and configuration: │12:02:39 eth0 600B/s 0B/s
GPU 0: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 1: Tesla V100-SXM2-16GB │time Interface RX TX
GPU 2: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 3: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s
GPU 4: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 5: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s
GPU 6: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 7: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
│12:02:39 eth0 600B/s 0B/s
Nvidia driver version: 418.74 │12:02:40 eth0 600B/s 0B/s
cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 │12:02:40 eth0 600B/s 0B/s
│12:02:40 eth0 600B/s 0B/s
Versions of relevant libraries: │12:02:40 eth0 600B/s 0B/s
[pip3] numpy==1.19.2 │12:02:40 eth0 600B/s 0B/s
[pip3] torch==1.5.1 │12:02:40 eth0 600B/s 0B/s
[pip3] torchsummary==1.5.1 │12:02:40 eth0 600B/s 0B/s
[pip3] torchvision==0.6.1 │12:02:40 eth0 600B/s 0B/s
[conda] blas 1.0 mkl defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl 2020.2 256 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl-service 2.3.0 py37he904b0f_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl_fft 1.2.0 py37h23d657b_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl_random 1.1.1 py37h0573a6f_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] torch 1.5.1 pypi_0 pypi │time Interface RX TX
[conda] torchsummary 1.5.1 pypi_0 pypi │12:02:41 eth0 600B/s 0B/s
[conda] torchvision 0.6.1 pypi_0 pypi
```
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @ngimel @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @agolynski @SciPioneer @H-Huang @mrzzd | 1.0 | Data copy from CPU to GPU use default stream in nightly version. - ## 🐛 Bug
After https://github.com/pytorch/pytorch/pull/46304, data copy per iteration from cpu to gpu can't overlap with last iteration's computation/communication.
There is my test result which shows different behaviors by literally same code.
nightly:

1.5.1 release:

## To Reproduce
Steps to reproduce the behavior:
1. Run the same code in nightly build and 1.5.1 release build. Use nsys profiling it.
My test code is listed as below, you can use it do some tests. But I think the reason that incurs different behavior is clear. Is it a new feature? If so, I might not think it's a good one.
```
import os
import time
import sys
import random
import traceback
import numpy as np
import torch
import torch.nn as nn
import torch.nn.parallel
import torch.distributed as dist
import torch.optim
import torch.utils.data
import torch.utils.data.distributed
import torchvision.transforms as transforms
import torchvision.datasets as datasets
import torchvision.models as models
from torch.multiprocessing import Pool, Process
import argparse
class AverageMeter(object):
"""Computes and stores the average and current value"""
def __init__(self):
self.reset()
def reset(self):
self.val = 0
self.avg = 0
self.sum = 0
self.count = 0
def update(self, val, n=1):
self.val = val
self.sum += val * n
self.count += n
self.avg = self.sum / self.count
def accuracy(output, target, topk=(1,)):
"""Computes the precision@k for the specified values of k"""
with torch.no_grad():
maxk = max(topk)
batch_size = target.size(0)
_, pred = output.topk(maxk, 1, True, True)
pred = pred.t()
correct = pred.eq(target.view(1, -1).expand_as(pred))
res = []
for k in topk:
correct_k = correct[:k].contiguous().view(-1).float().sum(0, keepdim=True)
res.append(correct_k.mul_(100.0 / batch_size))
return res
def train(train_loader, model, criterion, optimizer, epoch, batch_size):
batch_time = AverageMeter()
data_time = AverageMeter()
losses = AverageMeter()
top1 = AverageMeter()
top5 = AverageMeter()
local_rank = os.environ['LOCAL_RANK']
# switch to train mode
model.train()
end = time.time()
for i, (input, target) in enumerate(train_loader):
# measure data loading time
data_time.update(time.time() - end)
# Create non_blocking tensors for distributed training
# input = input.cuda(non_blocking=True)
target = target.cuda(non_blocking=True)
# compute output
output = model(input)
loss = criterion(output, target)
# measure accuracy and record loss
prec1, prec5 = accuracy(output, target, topk=(1, 5))
losses.update(loss.item(), input.size(0))
top1.update(prec1[0], input.size(0))
top5.update(prec5[0], input.size(0))
# compute gradients in a backward pass
optimizer.zero_grad()
loss.backward()
# Call step of optimizer to update model params
optimizer.step()
# measure elapsed time
if i >= 10:
batch_time.update(time.time() - end)
end = time.time()
if local_rank == '0' and i % 10 == 0 and i > 10:
# if local_rank == '0' and i > 0:
print('Epoch: [{0}][{1}/{2}]\t'
'Speed {speed_now:.3f} ({speed_avg:.3f})\t'
'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t'
'Data {data_time.val:.3f} ({data_time.avg:.3f})\t'
'Loss {loss.val:.4f} ({loss.avg:.4f})\t'
'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t'
'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format(
epoch, i, len(train_loader), batch_time=batch_time,
speed_now=batch_size/batch_time.val, speed_avg=batch_size/batch_time.avg,
data_time=data_time, loss=losses, top1=top1, top5=top5))
def adjust_learning_rate(initial_lr, optimizer, epoch):
"""Sets the learning rate to the initial LR decayed by 10 every 30 epochs"""
lr = initial_lr * (0.1 ** (epoch // 30))
for param_group in optimizer.param_groups:
param_group['lr'] = lr
def validate(val_loader, model, criterion):
batch_time = AverageMeter()
losses = AverageMeter()
top1 = AverageMeter()
top5 = AverageMeter()
# switch to evaluate mode
model.eval()
with torch.no_grad():
end = time.time()
for i, (input, target) in enumerate(val_loader):
input = input.cuda(non_blocking=True)
target = target.cuda(non_blocking=True)
# compute output
output = model(input)
loss = criterion(output, target)
# measure accuracy and record loss
prec1, prec5 = accuracy(output, target, topk=(1, 5))
losses.update(loss.item(), input.size(0))
top1.update(prec1[0], input.size(0))
top5.update(prec5[0], input.size(0))
# measure elapsed time
torch.cuda.synchronize()
batch_time.update(time.time() - end)
end = time.time()
if i % 100 == 0:
print('Test: [{0}/{1}]\t'
'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t'
'Loss {loss.val:.4f} ({loss.avg:.4f})\t'
'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t'
'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format(
i, len(val_loader), batch_time=batch_time, loss=losses,
top1=top1, top5=top5))
print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}'
.format(top1=top1, top5=top5))
return top1.avg
def start(backend):
print("Collect Inputs...")
# Batch Size for training and testing
batch_size = 64
# Number of additional worker processes for dataloading
workers = 8
# Number of epochs to train for
num_epochs = 1
# Starting Learning Rate
starting_lr = 0.01
# Distributed backend type
# dist_backend = 'nccl'
# dist_backend = 'esl'
dist_backend = backend
# Url used to setup distributed training
dist.init_process_group(dist_backend, init_method='env://')
print("Wordsize: ", os.environ['WORLD_SIZE'])
local_rank = int(os.environ['LOCAL_RANK'])
print("Local_Rank: ", local_rank)
dp_device_ids = [local_rank]
torch.cuda.set_device(local_rank)
device = torch.device("cuda", local_rank)
print("Initialize Model...")
# Construct Model
model = models.vgg16(pretrained=False).to(device)
# Make model DistributedDataParallel
model = torch.nn.parallel.DistributedDataParallel(
model, device_ids=dp_device_ids, output_device=local_rank)
# define loss function (criterion) and optimizer
criterion = nn.CrossEntropyLoss().cuda()
optimizer = torch.optim.SGD(
model.parameters(), starting_lr, momentum=0.9, weight_decay=1e-4)
print("Initialize Dataloaders...")
# Define the transform for the data. Notice, we must resize to 224x224 with this dataset and model.
transform = transforms.Compose(
[transforms.Resize(224),
transforms.ToTensor(),
transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))])
# Initialize Datasets. STL10 will automatically download if not present
# trainset = datasets.STL10(
# root='/workspace/dataset', split='train', download=True, transform=transform)
# valset = datasets.STL10(root='/workspace/dataset',
# split='test', download=True, transform=transform)
trainset = datasets.FakeData(size=batch_size*int(os.environ['WORLD_SIZE'])*100, num_classes=1000, ,transform=transform)
valset = datasets.FakeData(size=1000, transform=transform)
# Create DistributedSampler to handle distributing the dataset across nodes when training
# This can only be called after torch.distributed.init_process_group is called
train_sampler = torch.utils.data.distributed.DistributedSampler(trainset)
# Create the Dataloaders to feed data to the training and validation steps
train_loader = torch.utils.data.DataLoader(trainset, batch_size=batch_size, shuffle=(train_sampler is None),
num_workers=workers, pin_memory=True, sampler=train_sampler)
val_loader = torch.utils.data.DataLoader(valset, batch_size=batch_size, shuffle=False, num_workers=workers,
pin_memory=True)
best_prec1 = 0
for epoch in range(num_epochs):
# Set epoch count for DistributedSampler
train_sampler.set_epoch(epoch)
# Adjust learning rate according to schedule
adjust_learning_rate(starting_lr, optimizer, epoch)
# train for one epoch
print("\nBegin Training Epoch {}".format(epoch + 1))
train(train_loader, model, criterion, optimizer, epoch, batch_size*int(os.environ['WORLD_SIZE']))
# # evaluate on validation set
# print("Begin Validation @ Epoch {}".format(epoch + 1))
# prec1 = validate(val_loader, model, criterion)
# # remember best prec@1 and save checkpoint if desired
# # is_best = prec1 > best_prec1
# best_prec1 = max(prec1, best_prec1)
# print("Epoch Summary: ")
# print("\tEpoch Accuracy: {}".format(prec1))
# print("\tBest Accuracy: {}".format(best_prec1))
def set_seed(self, seed=0):
random.seed(seed)
np.random.seed(seed)
torch.manual_seed(seed) #为CPU设置种子用于生成随机数,以使得结果是确定的
torch.cuda.manual_seed(seed) #为当前GPU设置随机种子;
torch.cuda.manual_seed_all(seed) #如果使用多个GPU,为所有的GPU设置种子。
torch.backends.cudnn.deterministic = True # CPU和GPU结果一致
torch.backends.cudnn.benchmark = False
if __name__ == '__main__':
# set_seed(42)
parser = argparse.ArgumentParser()
parser.add_argument("--local_rank", type=int)
args = parser.parse_args()
os.environ['LOCAL_RANK'] = str(args.local_rank)
backend = nccl
start(backend)
# usage: python -m torch.distributed.launch --nproc_per_node=NUM_GPUS_YOU_HAVE --nnodes=2 --node_rank=1 --master_addr="127.0.0.1" --master_port=1997 YOUR_TRAINING_SCRIPT.py (--arg1 --arg2 --arg3 and all other arguments of your training script)
```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
## Environment
```
PyTorch version: 1.5.1 │12:02:37 eth0 600B/s 0B/s
Is debug build: No │12:02:37 eth0 600B/s 0B/s
CUDA used to build PyTorch: 10.0 │12:02:37 eth0 1.17KB/s 14.02KB/s
│12:02:38 eth0 600B/s 0B/s
OS: Ubuntu 16.04.6 LTS │12:02:38 eth0 600B/s 0B/s
GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.12) 5.4.0 20160609 │12:02:38 eth0 600B/s 0B/s
CMake version: version 3.5.1 │12:02:38 eth0 600B/s 0B/s
│12:02:38 eth0 600B/s 0B/s
Python version: 3.7 │12:02:38 eth0 600B/s 0B/s
Is CUDA available: Yes │12:02:38 eth0 600B/s 0B/s
CUDA runtime version: 10.0.130 │12:02:38 eth0 600B/s 0B/s
GPU models and configuration: │12:02:39 eth0 600B/s 0B/s
GPU 0: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 1: Tesla V100-SXM2-16GB │time Interface RX TX
GPU 2: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 3: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s
GPU 4: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 5: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s
GPU 6: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
GPU 7: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s
│12:02:39 eth0 600B/s 0B/s
Nvidia driver version: 418.74 │12:02:40 eth0 600B/s 0B/s
cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 │12:02:40 eth0 600B/s 0B/s
│12:02:40 eth0 600B/s 0B/s
Versions of relevant libraries: │12:02:40 eth0 600B/s 0B/s
[pip3] numpy==1.19.2 │12:02:40 eth0 600B/s 0B/s
[pip3] torch==1.5.1 │12:02:40 eth0 600B/s 0B/s
[pip3] torchsummary==1.5.1 │12:02:40 eth0 600B/s 0B/s
[pip3] torchvision==0.6.1 │12:02:40 eth0 600B/s 0B/s
[conda] blas 1.0 mkl defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl 2020.2 256 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl-service 2.3.0 py37he904b0f_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl_fft 1.2.0 py37h23d657b_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] mkl_random 1.1.1 py37h0573a6f_0 defaults │12:02:41 eth0 600B/s 0B/s
[conda] torch 1.5.1 pypi_0 pypi │time Interface RX TX
[conda] torchsummary 1.5.1 pypi_0 pypi │12:02:41 eth0 600B/s 0B/s
[conda] torchvision 0.6.1 pypi_0 pypi
```
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @ngimel @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @agolynski @SciPioneer @H-Huang @mrzzd | priority | data copy from cpu to gpu use default stream in nightly version 🐛 bug after data copy per iteration from cpu to gpu can t overlap with last iteration s computation communication there is my test result which shows different behaviors by literally same code nightly release to reproduce steps to reproduce the behavior run the same code in nightly build and release build use nsys profiling it my test code is listed as below you can use it do some tests but i think the reason that incurs different behavior is clear is it a new feature if so i might not think it s a good one import os import time import sys import random import traceback import numpy as np import torch import torch nn as nn import torch nn parallel import torch distributed as dist import torch optim import torch utils data import torch utils data distributed import torchvision transforms as transforms import torchvision datasets as datasets import torchvision models as models from torch multiprocessing import pool process import argparse class averagemeter object computes and stores the average and current value def init self self reset def reset self self val self avg self sum self count def update self val n self val val self sum val n self count n self avg self sum self count def accuracy output target topk computes the precision k for the specified values of k with torch no grad maxk max topk batch size target size pred output topk maxk true true pred pred t correct pred eq target view expand as pred res for k in topk correct k correct contiguous view float sum keepdim true res append correct k mul batch size return res def train train loader model criterion optimizer epoch batch size batch time averagemeter data time averagemeter losses averagemeter averagemeter averagemeter local rank os environ switch to train mode model train end time time for i input target in enumerate train loader measure data loading time data time update time time end create non blocking tensors for distributed training input input cuda non blocking true target target cuda non blocking true compute output output model input loss criterion output target measure accuracy and record loss accuracy output target topk losses update loss item input size update input size update input size compute gradients in a backward pass optimizer zero grad loss backward call step of optimizer to update model params optimizer step measure elapsed time if i batch time update time time end end time time if local rank and i and i if local rank and i print epoch t speed speed now speed avg t time batch time val batch time avg t data data time val data time avg t loss loss val loss avg t prec val avg t prec val avg format epoch i len train loader batch time batch time speed now batch size batch time val speed avg batch size batch time avg data time data time loss losses def adjust learning rate initial lr optimizer epoch sets the learning rate to the initial lr decayed by every epochs lr initial lr epoch for param group in optimizer param groups param group lr def validate val loader model criterion batch time averagemeter losses averagemeter averagemeter averagemeter switch to evaluate mode model eval with torch no grad end time time for i input target in enumerate val loader input input cuda non blocking true target target cuda non blocking true compute output output model input loss criterion output target measure accuracy and record loss accuracy output target topk losses update loss item input size update input size update input size measure elapsed time torch cuda synchronize batch time update time time end end time time if i print test t time batch time val batch time avg t loss loss val loss avg t prec val avg t prec val avg format i len val loader batch time batch time loss losses print prec avg prec avg format return avg def start backend print collect inputs batch size for training and testing batch size number of additional worker processes for dataloading workers number of epochs to train for num epochs starting learning rate starting lr distributed backend type dist backend nccl dist backend esl dist backend backend url used to setup distributed training dist init process group dist backend init method env print wordsize os environ local rank int os environ print local rank local rank dp device ids torch cuda set device local rank device torch device cuda local rank print initialize model construct model model models pretrained false to device make model distributeddataparallel model torch nn parallel distributeddataparallel model device ids dp device ids output device local rank define loss function criterion and optimizer criterion nn crossentropyloss cuda optimizer torch optim sgd model parameters starting lr momentum weight decay print initialize dataloaders define the transform for the data notice we must resize to with this dataset and model transform transforms compose transforms resize transforms totensor transforms normalize initialize datasets will automatically download if not present trainset datasets root workspace dataset split train download true transform transform valset datasets root workspace dataset split test download true transform transform trainset datasets fakedata size batch size int os environ num classes transform transform valset datasets fakedata size transform transform create distributedsampler to handle distributing the dataset across nodes when training this can only be called after torch distributed init process group is called train sampler torch utils data distributed distributedsampler trainset create the dataloaders to feed data to the training and validation steps train loader torch utils data dataloader trainset batch size batch size shuffle train sampler is none num workers workers pin memory true sampler train sampler val loader torch utils data dataloader valset batch size batch size shuffle false num workers workers pin memory true best for epoch in range num epochs set epoch count for distributedsampler train sampler set epoch epoch adjust learning rate according to schedule adjust learning rate starting lr optimizer epoch train for one epoch print nbegin training epoch format epoch train train loader model criterion optimizer epoch batch size int os environ evaluate on validation set print begin validation epoch format epoch validate val loader model criterion remember best prec and save checkpoint if desired is best best best max best print epoch summary print tepoch accuracy format print tbest accuracy format best def set seed self seed random seed seed np random seed seed torch manual seed seed 为cpu设置种子用于生成随机数,以使得结果是确定的 torch cuda manual seed seed 为当前gpu设置随机种子; torch cuda manual seed all seed 如果使用多个gpu,为所有的gpu设置种子。 torch backends cudnn deterministic true cpu和gpu结果一致 torch backends cudnn benchmark false if name main set seed parser argparse argumentparser parser add argument local rank type int args parser parse args os environ str args local rank backend nccl start backend usage python m torch distributed launch nproc per node num gpus you have nnodes node rank master addr master port your training script py and all other arguments of your training script expected behavior environment pytorch version │ s s is debug build no │ s s cuda used to build pytorch │ s s │ s s os ubuntu lts │ s s gcc version ubuntu │ s s cmake version version │ s s │ s s python version │ s s is cuda available yes │ s s cuda runtime version │ s s gpu models and configuration │ s s gpu tesla │ s s gpu tesla │time interface rx tx gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s │ s s nvidia driver version │ s s cudnn version usr lib linux gnu libcudnn so │ s s │ s s versions of relevant libraries │ s s numpy │ s s torch │ s s torchsummary │ s s torchvision │ s s blas mkl defaults │ s s mkl defaults │ s s mkl service defaults │ s s mkl fft defaults │ s s mkl random defaults │ s s torch pypi pypi │time interface rx tx torchsummary pypi pypi │ s s torchvision pypi pypi cc ezyang gchanan bdhirsh jbschlosser ngimel pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jiayisuse agolynski scipioneer h huang mrzzd | 1 |
454,599 | 13,104,449,841 | IssuesEvent | 2020-08-04 10:17:48 | dnd-mentee-3rd/dnd-mentee-3rd-1-planto | https://api.github.com/repos/dnd-mentee-3rd/dnd-mentee-3rd-1-planto | closed | 로그인 View 구현 | Android Priority: High iOS | ## 필요한 요소
- Email
- [x] iOS
- [x] Android
- 비밀번호
- [x] iOS
- [x] Android
- 자동 로그인 체크박스
- [x] iOS
- [x] Android
- 로그인 버튼
- [x] iOS
- [x] Android
- 비밀번호 찾기 버튼
- [x] iOS
- [x] Android
## 필요한 기능
- 이메일 형식 검사
- [x] iOS
- [x] Android
- 이메일과 비밀번호가 등록된 정보와 맞는지 검사
- [x] iOS
- [x] Android
- 자동 로그인 버튼 토글
- [x] iOS
- [x] Android
- 로그인 버튼 액션
- 정보 일치 시 인증 정보 저장
- [x] iOS
- [x] Android
- 정보 불일치 시 정보 확인 요청 Alert
- [x] iOS
- [x] Android
- 비밀번호 찾기 버튼 액션 (추후 진행)
| 1.0 | 로그인 View 구현 - ## 필요한 요소
- Email
- [x] iOS
- [x] Android
- 비밀번호
- [x] iOS
- [x] Android
- 자동 로그인 체크박스
- [x] iOS
- [x] Android
- 로그인 버튼
- [x] iOS
- [x] Android
- 비밀번호 찾기 버튼
- [x] iOS
- [x] Android
## 필요한 기능
- 이메일 형식 검사
- [x] iOS
- [x] Android
- 이메일과 비밀번호가 등록된 정보와 맞는지 검사
- [x] iOS
- [x] Android
- 자동 로그인 버튼 토글
- [x] iOS
- [x] Android
- 로그인 버튼 액션
- 정보 일치 시 인증 정보 저장
- [x] iOS
- [x] Android
- 정보 불일치 시 정보 확인 요청 Alert
- [x] iOS
- [x] Android
- 비밀번호 찾기 버튼 액션 (추후 진행)
| priority | 로그인 view 구현 필요한 요소 email ios android 비밀번호 ios android 자동 로그인 체크박스 ios android 로그인 버튼 ios android 비밀번호 찾기 버튼 ios android 필요한 기능 이메일 형식 검사 ios android 이메일과 비밀번호가 등록된 정보와 맞는지 검사 ios android 자동 로그인 버튼 토글 ios android 로그인 버튼 액션 정보 일치 시 인증 정보 저장 ios android 정보 불일치 시 정보 확인 요청 alert ios android 비밀번호 찾기 버튼 액션 추후 진행 | 1 |
746,796 | 26,046,512,838 | IssuesEvent | 2022-12-22 14:50:07 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | opened | [PTR][Dungeon] Zul Aman | Dungeon Item Loot NPC Spell Priority: High Status: Confirmed | **How to reproduce:**
1. akil'zon gauntlet event chest got no loot, tested it a lot even with all bosses dead i didnt see any loots. (beside gold satchel)

2. Jan alai if you stay in a corner out of range of him he wont engage you in melee combat
also his firebomb spell is not working as intended he is not dropping any bombs


3. Zandalari hierophant trash pack before hex lord, hierophant simply wont react after pulling his trash pack until you attack him directly

4. halazzi water totem got too much hp it should be 100k hp based on this https://cata-twinhead.twinstar.cz/?npc=52755 and this video https://youtu.be/w80XUSB3ls4?t=234 right now it has more than 1mil hp

4. on hexlord malacrass shackling the darkheart wont pull the boss

| 1.0 | [PTR][Dungeon] Zul Aman - **How to reproduce:**
1. akil'zon gauntlet event chest got no loot, tested it a lot even with all bosses dead i didnt see any loots. (beside gold satchel)

2. Jan alai if you stay in a corner out of range of him he wont engage you in melee combat
also his firebomb spell is not working as intended he is not dropping any bombs


3. Zandalari hierophant trash pack before hex lord, hierophant simply wont react after pulling his trash pack until you attack him directly

4. halazzi water totem got too much hp it should be 100k hp based on this https://cata-twinhead.twinstar.cz/?npc=52755 and this video https://youtu.be/w80XUSB3ls4?t=234 right now it has more than 1mil hp

4. on hexlord malacrass shackling the darkheart wont pull the boss

| priority | zul aman how to reproduce akil zon gauntlet event chest got no loot tested it a lot even with all bosses dead i didnt see any loots beside gold satchel jan alai if you stay in a corner out of range of him he wont engage you in melee combat also his firebomb spell is not working as intended he is not dropping any bombs zandalari hierophant trash pack before hex lord hierophant simply wont react after pulling his trash pack until you attack him directly halazzi water totem got too much hp it should be hp based on this and this video right now it has more than hp on hexlord malacrass shackling the darkheart wont pull the boss | 1 |
757,883 | 26,533,517,563 | IssuesEvent | 2023-01-19 14:11:15 | fractal-analytics-platform/fractal-server | https://api.github.com/repos/fractal-analytics-platform/fractal-server | closed | JobExecutionError without relevant error message | High Priority | I'm trying to run a new example and my workflow fails during the `Convert Yokogawa to OME-Zarr` with a `JobExecutionError`.
The weird thing is that the .err file of the job says the job finished successfully:
```
Matplotlib created a temporary config/cache directory at /tmp/matplotlib-wrbssjh0 because the default path (/data/homes/fractal/.cache/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing.
2023-01-16 13:49:38,878; INFO; START yokogawa_to_ome_zarr task
2023-01-16 13:49:39,263; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C01*.png
2023-01-16 13:49:55,731; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C02*.png
2023-01-16 13:50:09,745; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A02*C03*.png
2023-01-16 13:50:39,890; INFO; END yokogawa_to_ome_zarr task
```
And the slurm .err file only contains this:
```
Content of /net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.9445845.err:
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/0.slurm.submit.sbatch': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.submit.sbatch': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/cfut.in.uoaLSyH2xsDzvmmasowgIC7ahBxNWwbF.pickle': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/workflow.log': Operation not permitted
```
But that also appears in jobs that don't fail.
Thus, I don't understand why this job is failing (currently looking into this) and it's worrying that the job fails without a clear error message...
@tcompa the chmod failing can't actually fail the job, right? Otherwise, I wouldn't understand why this reproducibly happens in the workflow here, but not in other examples...
I'll check with a different executor with more memory (shouldn't actually be an issue, but who knows) and a different example script. | 1.0 | JobExecutionError without relevant error message - I'm trying to run a new example and my workflow fails during the `Convert Yokogawa to OME-Zarr` with a `JobExecutionError`.
The weird thing is that the .err file of the job says the job finished successfully:
```
Matplotlib created a temporary config/cache directory at /tmp/matplotlib-wrbssjh0 because the default path (/data/homes/fractal/.cache/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing.
2023-01-16 13:49:38,878; INFO; START yokogawa_to_ome_zarr task
2023-01-16 13:49:39,263; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C01*.png
2023-01-16 13:49:55,731; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C02*.png
2023-01-16 13:50:09,745; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A02*C03*.png
2023-01-16 13:50:39,890; INFO; END yokogawa_to_ome_zarr task
```
And the slurm .err file only contains this:
```
Content of /net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.9445845.err:
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/0.slurm.submit.sbatch': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.submit.sbatch': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/cfut.in.uoaLSyH2xsDzvmmasowgIC7ahBxNWwbF.pickle': Operation not permitted
chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/workflow.log': Operation not permitted
```
But that also appears in jobs that don't fail.
Thus, I don't understand why this job is failing (currently looking into this) and it's worrying that the job fails without a clear error message...
@tcompa the chmod failing can't actually fail the job, right? Otherwise, I wouldn't understand why this reproducibly happens in the workflow here, but not in other examples...
I'll check with a different executor with more memory (shouldn't actually be an issue, but who knows) and a different example script. | priority | jobexecutionerror without relevant error message i m trying to run a new example and my workflow fails during the convert yokogawa to ome zarr with a jobexecutionerror the weird thing is that the err file of the job says the job finished successfully matplotlib created a temporary config cache directory at tmp matplotlib because the default path data homes fractal cache matplotlib is not a writable directory it is highly recommended to set the mplconfigdir environment variable to a writable directory in particular to speed up the import of matplotlib and to better support multiprocessing info start yokogawa to ome zarr task info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info end yokogawa to ome zarr task and the slurm err file only contains this content of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job par zarr b slurm err chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job slurm submit sbatch operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job par zarr b slurm submit sbatch operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job cfut in pickle operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job workflow log operation not permitted but that also appears in jobs that don t fail thus i don t understand why this job is failing currently looking into this and it s worrying that the job fails without a clear error message tcompa the chmod failing can t actually fail the job right otherwise i wouldn t understand why this reproducibly happens in the workflow here but not in other examples i ll check with a different executor with more memory shouldn t actually be an issue but who knows and a different example script | 1 |
429,635 | 12,426,663,328 | IssuesEvent | 2020-05-24 22:25:42 | stevenwaterman/musetree | https://api.github.com/repos/stevenwaterman/musetree | closed | Load starter samples into the app | High Priority enhancement good first issue help wanted | In the [official musenet app](https://openai.com/blog/musenet/), you can specify what you want the generator to start with.
Add the 12 samples in the official app to musetree.
It's fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song.
Loading the MP3 for those samples may be difficult - 2 options:
* Hit the `/audio` musenet endpoint to get the mp3 for the encoding
* Hard-code the mp3 base64 into the app alongside the encoding. (Temporarily)
Custom uploads are more complex and therefore in a separate issue. | 1.0 | Load starter samples into the app - In the [official musenet app](https://openai.com/blog/musenet/), you can specify what you want the generator to start with.
Add the 12 samples in the official app to musetree.
It's fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song.
Loading the MP3 for those samples may be difficult - 2 options:
* Hit the `/audio` musenet endpoint to get the mp3 for the encoding
* Hard-code the mp3 base64 into the app alongside the encoding. (Temporarily)
Custom uploads are more complex and therefore in a separate issue. | priority | load starter samples into the app in the you can specify what you want the generator to start with add the samples in the official app to musetree it s fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song loading the for those samples may be difficult options hit the audio musenet endpoint to get the for the encoding hard code the into the app alongside the encoding temporarily custom uploads are more complex and therefore in a separate issue | 1 |
486,436 | 14,009,063,038 | IssuesEvent | 2020-10-29 01:21:29 | AFlyingCar/WarringStates | https://api.github.com/repos/AFlyingCar/WarringStates | closed | Declaiming a Chunk in the Capitol removes Claimer | High Priority bug | Despite a message "You cannot rescind you claim on your capitol" | 1.0 | Declaiming a Chunk in the Capitol removes Claimer - Despite a message "You cannot rescind you claim on your capitol" | priority | declaiming a chunk in the capitol removes claimer despite a message you cannot rescind you claim on your capitol | 1 |
156,091 | 5,964,210,149 | IssuesEvent | 2017-05-30 08:14:44 | mkdo/kapow-theme | https://api.github.com/repos/mkdo/kapow-theme | closed | FE assets re-build | Priority: High Status: Completed Type: Maintenance | The compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules.
Can be completed when the full-build testing is finished at the end of Phase 2.0. | 1.0 | FE assets re-build - The compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules.
Can be completed when the full-build testing is finished at the end of Phase 2.0. | priority | fe assets re build the compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules can be completed when the full build testing is finished at the end of phase | 1 |
6,899 | 2,595,479,709 | IssuesEvent | 2015-02-20 14:36:23 | centre-for-educational-technology/edidaktikum | https://api.github.com/repos/centre-for-educational-technology/edidaktikum | opened | Kogumik (täiendamisel) | High Priority New feature | Kogumik on õppematerjalide kollektsioon. Kõik õppejõu rollis kasutajad saavad luua kogumikke.
Kogumiku loomiseks on kaks võimalust:
... | 1.0 | Kogumik (täiendamisel) - Kogumik on õppematerjalide kollektsioon. Kõik õppejõu rollis kasutajad saavad luua kogumikke.
Kogumiku loomiseks on kaks võimalust:
... | priority | kogumik täiendamisel kogumik on õppematerjalide kollektsioon kõik õppejõu rollis kasutajad saavad luua kogumikke kogumiku loomiseks on kaks võimalust | 1 |
331,998 | 10,083,062,497 | IssuesEvent | 2019-07-25 12:51:43 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | `with torch.enable_grad` also works outside a `no_grad` context | high priority module: autograd module: docs small triaged | ## 📚 Documentation
The [`torch.autograd.enable_grad` documentation](https://pytorch.org/docs/stable/autograd.html#torch.autograd.enable_grad) says:
> Enables gradient calculation inside a `no_grad` context. This has no effect outside of `no_grad`.
This implies:
```
torch.set_grad_enabled(False)
with torch.enable_grad:
# Gradient tracking will NOT be enabled here.
torch.set_grad_enabled(True)
vs:
with torch.no_grad():
with torch.enable_grad:
# Gradient tracking IS enabled here.
```
However the observed behaviour (`1.0.1.post2`) is:
```
x = torch.tensor([.1], requires_grad=True)
with torch.no_grad():
with torch.enable_grad():
y = x * 2
print(y.requires_grad) # True (as expected)
with torch.set_grad_enabled(False):
y = x * 2
print(y.requires_grad) # False (as expected)
with torch.set_grad_enabled(False):
with torch.enable_grad():
y = x * 2
print(y.requires_grad) # True, but False expected from doc quote
```
Note the last example is not "inside a `no_grad` context", but it still works.
Other prior art: [PyTorch set_grad_enabled(False) vs with no_grad():](https://stackoverflow.com/a/53447634/5353461)
-----------
I'm assuming the documentation is incorrect, and it should say simply:
> Enables gradient calculation.
Would you accept a PR? | 1.0 | `with torch.enable_grad` also works outside a `no_grad` context - ## 📚 Documentation
The [`torch.autograd.enable_grad` documentation](https://pytorch.org/docs/stable/autograd.html#torch.autograd.enable_grad) says:
> Enables gradient calculation inside a `no_grad` context. This has no effect outside of `no_grad`.
This implies:
```
torch.set_grad_enabled(False)
with torch.enable_grad:
# Gradient tracking will NOT be enabled here.
torch.set_grad_enabled(True)
vs:
with torch.no_grad():
with torch.enable_grad:
# Gradient tracking IS enabled here.
```
However the observed behaviour (`1.0.1.post2`) is:
```
x = torch.tensor([.1], requires_grad=True)
with torch.no_grad():
with torch.enable_grad():
y = x * 2
print(y.requires_grad) # True (as expected)
with torch.set_grad_enabled(False):
y = x * 2
print(y.requires_grad) # False (as expected)
with torch.set_grad_enabled(False):
with torch.enable_grad():
y = x * 2
print(y.requires_grad) # True, but False expected from doc quote
```
Note the last example is not "inside a `no_grad` context", but it still works.
Other prior art: [PyTorch set_grad_enabled(False) vs with no_grad():](https://stackoverflow.com/a/53447634/5353461)
-----------
I'm assuming the documentation is incorrect, and it should say simply:
> Enables gradient calculation.
Would you accept a PR? | priority | with torch enable grad also works outside a no grad context 📚 documentation the says enables gradient calculation inside a no grad context this has no effect outside of no grad this implies torch set grad enabled false with torch enable grad gradient tracking will not be enabled here torch set grad enabled true vs with torch no grad with torch enable grad gradient tracking is enabled here however the observed behaviour is x torch tensor requires grad true with torch no grad with torch enable grad y x print y requires grad true as expected with torch set grad enabled false y x print y requires grad false as expected with torch set grad enabled false with torch enable grad y x print y requires grad true but false expected from doc quote note the last example is not inside a no grad context but it still works other prior art i m assuming the documentation is incorrect and it should say simply enables gradient calculation would you accept a pr | 1 |
632,465 | 20,198,013,114 | IssuesEvent | 2022-02-11 12:32:06 | AlpsBTE/Plot-System | https://api.github.com/repos/AlpsBTE/Plot-System | opened | [Feature] - Make the plot outline usable in code | enhancement high priority | **Is your feature request related to a problem? Please describe.**
Currently, there is no way to get the polygon outline of the plot in the PlotSystem plugin.
I first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region, not a poly region.
For the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot.
This would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for /tpll etc.
**Describe the solution you'd like**
Add a column "outline" to the table plotsystem_plots which contains a list of points of the polygon from the plot.
When the plot is created the PlotSystemTerra PlotCreator.java adds the outline data additionally to the other data that is inserted.
**Additional context**
The "one-world" branch is on hold until this is released | 1.0 | [Feature] - Make the plot outline usable in code - **Is your feature request related to a problem? Please describe.**
Currently, there is no way to get the polygon outline of the plot in the PlotSystem plugin.
I first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region, not a poly region.
For the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot.
This would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for /tpll etc.
**Describe the solution you'd like**
Add a column "outline" to the table plotsystem_plots which contains a list of points of the polygon from the plot.
When the plot is created the PlotSystemTerra PlotCreator.java adds the outline data additionally to the other data that is inserted.
**Additional context**
The "one-world" branch is on hold until this is released | priority | make the plot outline usable in code is your feature request related to a problem please describe currently there is no way to get the polygon outline of the plot in the plotsystem plugin i first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region not a poly region for the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot this would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for tpll etc describe the solution you d like add a column outline to the table plotsystem plots which contains a list of points of the polygon from the plot when the plot is created the plotsystemterra plotcreator java adds the outline data additionally to the other data that is inserted additional context the one world branch is on hold until this is released | 1 |
552,677 | 16,246,774,495 | IssuesEvent | 2021-05-07 15:20:41 | DCS-LCSR/SignStream3 | https://api.github.com/repos/DCS-LCSR/SignStream3 | closed | Overlap that should not be allowed - gloss items - est. 8 hours | bug cannot reproduce priority HIGH | I'm working with file 112, which I emailed you very recently -- utterance 82.
The endpoint of one sign (GROUP/TOGETHER) overlaps with the initial hold of the following sign.
This should not be allowed...
<img width="713" alt="screen shot 2018-06-06 at 2 10 10 pm" src="https://user-images.githubusercontent.com/13629122/41056894-9146a3ac-6993-11e8-9f5f-f316643d70a9.png">
I just deleted that initial hold, and I am (correctly) no longer allowed to add a hold, because it would involve disallowed overlap. So perhaps this is a holdover from an older version of the program that did not enforce things properly. Just letting you know, in any case. | 1.0 | Overlap that should not be allowed - gloss items - est. 8 hours - I'm working with file 112, which I emailed you very recently -- utterance 82.
The endpoint of one sign (GROUP/TOGETHER) overlaps with the initial hold of the following sign.
This should not be allowed...
<img width="713" alt="screen shot 2018-06-06 at 2 10 10 pm" src="https://user-images.githubusercontent.com/13629122/41056894-9146a3ac-6993-11e8-9f5f-f316643d70a9.png">
I just deleted that initial hold, and I am (correctly) no longer allowed to add a hold, because it would involve disallowed overlap. So perhaps this is a holdover from an older version of the program that did not enforce things properly. Just letting you know, in any case. | priority | overlap that should not be allowed gloss items est hours i m working with file which i emailed you very recently utterance the endpoint of one sign group together overlaps with the initial hold of the following sign this should not be allowed img width alt screen shot at pm src i just deleted that initial hold and i am correctly no longer allowed to add a hold because it would involve disallowed overlap so perhaps this is a holdover from an older version of the program that did not enforce things properly just letting you know in any case | 1 |
178,034 | 6,598,077,340 | IssuesEvent | 2017-09-16 00:02:32 | how-is/how_is | https://api.github.com/repos/how-is/how_is | closed | move away from slop bc of conflicts | high priority | this is like the 3rd time i've had slop interfere with something, one of which prompted a new release only to encounter a different conflict the next time I tried to use it(!).
so, as much as I like it, I think it'd be best to use something else. probably just Ruby's built in OptionParser, even if it's not as nice imo.
this is blocking work on https://github.com/rubytogether/rubytogether.org/pull/91 | 1.0 | move away from slop bc of conflicts - this is like the 3rd time i've had slop interfere with something, one of which prompted a new release only to encounter a different conflict the next time I tried to use it(!).
so, as much as I like it, I think it'd be best to use something else. probably just Ruby's built in OptionParser, even if it's not as nice imo.
this is blocking work on https://github.com/rubytogether/rubytogether.org/pull/91 | priority | move away from slop bc of conflicts this is like the time i ve had slop interfere with something one of which prompted a new release only to encounter a different conflict the next time i tried to use it so as much as i like it i think it d be best to use something else probably just ruby s built in optionparser even if it s not as nice imo this is blocking work on | 1 |
519,775 | 15,057,030,623 | IssuesEvent | 2021-02-03 21:03:20 | protofire/omen-exchange | https://api.github.com/repos/protofire/omen-exchange | opened | Market list item in Market overview is not showing the correct "current prediction" | bug priority:high | Market: https://xdai.dxtest.eth.link/#/0xd3f61725f4a1905223160abec29b98e81cd569ba
if you see the market details:
<img width="610" alt="Bildschirmfoto 2021-02-03 um 22 02 02" src="https://user-images.githubusercontent.com/5337809/106809020-76775d00-666b-11eb-9d3f-8dc673645294.png">
Current Prediction shows: `1,920.94 USD`
If you go back to the market details view, we see the current prediction is `1,657.76 USD`
<img width="613" alt="Bildschirmfoto 2021-02-03 um 22 02 39" src="https://user-images.githubusercontent.com/5337809/106809150-9e66c080-666b-11eb-94e6-984af0b129f3.png">
| 1.0 | Market list item in Market overview is not showing the correct "current prediction" - Market: https://xdai.dxtest.eth.link/#/0xd3f61725f4a1905223160abec29b98e81cd569ba
if you see the market details:
<img width="610" alt="Bildschirmfoto 2021-02-03 um 22 02 02" src="https://user-images.githubusercontent.com/5337809/106809020-76775d00-666b-11eb-9d3f-8dc673645294.png">
Current Prediction shows: `1,920.94 USD`
If you go back to the market details view, we see the current prediction is `1,657.76 USD`
<img width="613" alt="Bildschirmfoto 2021-02-03 um 22 02 39" src="https://user-images.githubusercontent.com/5337809/106809150-9e66c080-666b-11eb-94e6-984af0b129f3.png">
| priority | market list item in market overview is not showing the correct current prediction market if you see the market details img width alt bildschirmfoto um src current prediction shows usd if you go back to the market details view we see the current prediction is usd img width alt bildschirmfoto um src | 1 |
157,835 | 6,016,754,062 | IssuesEvent | 2017-06-07 07:56:29 | buggins/dlangui | https://api.github.com/repos/buggins/dlangui | closed | Folder and file icons disappeared from FileDialog | bug high_priority | Since this commit https://github.com/buggins/dlangui/commit/dfe707322254e75e362853aac3226ed337a33714 no folder or file icon shown in file dialog | 1.0 | Folder and file icons disappeared from FileDialog - Since this commit https://github.com/buggins/dlangui/commit/dfe707322254e75e362853aac3226ed337a33714 no folder or file icon shown in file dialog | priority | folder and file icons disappeared from filedialog since this commit no folder or file icon shown in file dialog | 1 |
343,378 | 10,329,021,746 | IssuesEvent | 2019-09-02 11:03:10 | python/mypy | https://api.github.com/repos/python/mypy | closed | Error codes in # type: ignore don't suppress related notes | bug priority-0-high topic-usability | For example:
```python
import nostubs # type: ignore[import]
```
Causes
```
mypy/report.py:30: note: (Stub files are from https://github.com/python/typeshed)
```
and error code 1. I didn't check carefully, but I think this affects all error messages that also emit additional notes (there are quite many of them). | 1.0 | Error codes in # type: ignore don't suppress related notes - For example:
```python
import nostubs # type: ignore[import]
```
Causes
```
mypy/report.py:30: note: (Stub files are from https://github.com/python/typeshed)
```
and error code 1. I didn't check carefully, but I think this affects all error messages that also emit additional notes (there are quite many of them). | priority | error codes in type ignore don t suppress related notes for example python import nostubs type ignore causes mypy report py note stub files are from and error code i didn t check carefully but i think this affects all error messages that also emit additional notes there are quite many of them | 1 |
415,685 | 12,133,028,555 | IssuesEvent | 2020-04-23 08:20:54 | arkhn/pyrog | https://api.github.com/repos/arkhn/pyrog | closed | Slicing hierarchy error | Bug High Priority | When importing a profile with extensions (ex : observation-bp https://www.hl7.org/fhir/bp.html or variant http://hl7.org/fhir/uv/genomics-reporting/variant.html), the slice structure is not as expected.
Example of observation-bp:
* Actual state

* Expected
see https://www.hl7.org/fhir/bp.html#profile

| 1.0 | Slicing hierarchy error - When importing a profile with extensions (ex : observation-bp https://www.hl7.org/fhir/bp.html or variant http://hl7.org/fhir/uv/genomics-reporting/variant.html), the slice structure is not as expected.
Example of observation-bp:
* Actual state

* Expected
see https://www.hl7.org/fhir/bp.html#profile

| priority | slicing hierarchy error when importing a profile with extensions ex observation bp or variant the slice structure is not as expected example of observation bp actual state expected see | 1 |
189,065 | 6,793,598,703 | IssuesEvent | 2017-11-01 08:23:20 | dalaranwow/dalaran-wow | https://api.github.com/repos/dalaranwow/dalaran-wow | closed | Warrior Talent: Sweeping Strikes -- Defensive Stance cancels SS. | Class - Warrior Fixed - Waiting for PTR update Priority - High | **Current behaviour**: If you have Sweeping Strikes up, switching to Defensive Stance cancels it.
**Expected behaviour**: SS shouldn't be cancelled by switching into DS.
**Steps to reproduce the problem**:
1. Use Sweeping Strikes;
2. Go into Defensive Stance -- SS is gone.
**Include proofs for this behaviour**
Sweeping Strikes ID: https://wotlk-twinhead.twinstar.cz/?spell=12328
Retail proof: https://youtu.be/nWUkwISLPPI?t=290 -- watch the activation of Sweeping Strikes in correspondence to the Stance and keep an eye on both. See that upon switching to Defensive Stance, SS remains. | 1.0 | Warrior Talent: Sweeping Strikes -- Defensive Stance cancels SS. - **Current behaviour**: If you have Sweeping Strikes up, switching to Defensive Stance cancels it.
**Expected behaviour**: SS shouldn't be cancelled by switching into DS.
**Steps to reproduce the problem**:
1. Use Sweeping Strikes;
2. Go into Defensive Stance -- SS is gone.
**Include proofs for this behaviour**
Sweeping Strikes ID: https://wotlk-twinhead.twinstar.cz/?spell=12328
Retail proof: https://youtu.be/nWUkwISLPPI?t=290 -- watch the activation of Sweeping Strikes in correspondence to the Stance and keep an eye on both. See that upon switching to Defensive Stance, SS remains. | priority | warrior talent sweeping strikes defensive stance cancels ss current behaviour if you have sweeping strikes up switching to defensive stance cancels it expected behaviour ss shouldn t be cancelled by switching into ds steps to reproduce the problem use sweeping strikes go into defensive stance ss is gone include proofs for this behaviour sweeping strikes id retail proof watch the activation of sweeping strikes in correspondence to the stance and keep an eye on both see that upon switching to defensive stance ss remains | 1 |
650,404 | 21,390,822,545 | IssuesEvent | 2022-04-21 06:54:06 | eclipse/dirigible | https://api.github.com/repos/eclipse/dirigible | opened | [EDM] Put all the generated artefacts under a 'gen' folder | web-ide usability priority-high efforts-medium component-template | Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned. | 1.0 | [EDM] Put all the generated artefacts under a 'gen' folder - Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned. | priority | put all the generated artefacts under a gen folder put all the generated artefacts under a gen folder it will be easy and definitive when regeneration is needed what to be cleaned | 1 |
474,503 | 13,671,187,719 | IssuesEvent | 2020-09-29 06:31:51 | fossasia/open-event-frontend | https://api.github.com/repos/fossasia/open-event-frontend | closed | Event Wizard publishing with deleted tickets | Priority: High bug | Sometimes the ticket display shows “there are no tickets” even though a free ticket is defined | 1.0 | Event Wizard publishing with deleted tickets - Sometimes the ticket display shows “there are no tickets” even though a free ticket is defined | priority | event wizard publishing with deleted tickets sometimes the ticket display shows “there are no tickets” even though a free ticket is defined | 1 |
63,626 | 3,197,112,323 | IssuesEvent | 2015-10-01 01:18:13 | MarcNo/lifelines | https://api.github.com/repos/MarcNo/lifelines | closed | docs directory missing from dist | auto-migrated high priority sourceforge | distribution lifelines-3.0.11.tar.gz file
does not include docs directory
Reported by: puddle
Original Ticket: "bugs/71":https://sourceforge.net/p/bugs/71 | 1.0 | docs directory missing from dist - distribution lifelines-3.0.11.tar.gz file
does not include docs directory
Reported by: puddle
Original Ticket: "bugs/71":https://sourceforge.net/p/bugs/71 | priority | docs directory missing from dist distribution lifelines tar gz file does not include docs directory reported by puddle original ticket bugs | 1 |
630,597 | 20,114,377,407 | IssuesEvent | 2022-02-07 17:55:55 | webcat12345/ngx-ui-switch | https://api.github.com/repos/webcat12345/ngx-ui-switch | closed | reverse attribute question | Priority: High Status: Available | Hello!
What is the intended way of "reverse"?
I currently have a boolean coming from a backend with "false" as the value, which causes the switch to be in "off" mode.
I want to reverse this, so that the switch is "on" when its value is false.
Is this what "reverse" should do? If that's the case, then I think it's broken. | 1.0 | reverse attribute question - Hello!
What is the intended way of "reverse"?
I currently have a boolean coming from a backend with "false" as the value, which causes the switch to be in "off" mode.
I want to reverse this, so that the switch is "on" when its value is false.
Is this what "reverse" should do? If that's the case, then I think it's broken. | priority | reverse attribute question hello what is the intended way of reverse i currently have a boolean coming from a backend with false as the value which causes the switch to be in off mode i want to reverse this so that the switch is on when its value is false is this what reverse should do if that s the case then i think it s broken | 1 |
752,703 | 26,321,357,735 | IssuesEvent | 2023-01-10 00:12:24 | microsoft/fluentui | https://api.github.com/repos/microsoft/fluentui | closed | Input: `contentBefore` and `contentAfter` don't work with `bundleIcons` | Type: Bug :bug: Priority 1: High Partner Ask Status: In PR Component: Input | **Repro:** https://codesandbox.io/s/mystifying-austin-02jpd4?file=/example.tsx

The specificity of the content styles in `Input` are more specific than the styles applied in bundled icons, the result is that both filled and regular icons are always visible
https://github.com/microsoft/fluentui/blob/66bf89f634cad4a275e957d7a2214c7e73ff8c2e/packages/react-components/react-input/src/components/Input/useInputStyles.ts#L215-L235 | 1.0 | Input: `contentBefore` and `contentAfter` don't work with `bundleIcons` - **Repro:** https://codesandbox.io/s/mystifying-austin-02jpd4?file=/example.tsx

The specificity of the content styles in `Input` are more specific than the styles applied in bundled icons, the result is that both filled and regular icons are always visible
https://github.com/microsoft/fluentui/blob/66bf89f634cad4a275e957d7a2214c7e73ff8c2e/packages/react-components/react-input/src/components/Input/useInputStyles.ts#L215-L235 | priority | input contentbefore and contentafter don t work with bundleicons repro the specificity of the content styles in input are more specific than the styles applied in bundled icons the result is that both filled and regular icons are always visible | 1 |
152,119 | 5,833,346,104 | IssuesEvent | 2017-05-09 01:10:37 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | ALP: add support for `#replaceUrl` for cache hit improvement | Category: Ads P1: High Priority Type: Bug | ALPs have many client-resolvable query parameters which greatly reduce cache-hit by multiplying cache keys. E.g. `https://host/alp?adpos=X` will multiply on the possible space of `adpos` query parameter, even though it's never used by the page itself.
The proposal is to use `https://host/alp` as the cache URL/key and replace it when AMP runtime starts up to `https://host/alp?adpos=X`. The replacing URL will be specified via viewer param as `#replaceUrl=...`.
Some important notes:
1. Replacement should be done, ideally, very early on and definitely before `installDocumentStateService` is called.
2. The replacement URL must match the original URLs origin _and_ source origin
| 1.0 | ALP: add support for `#replaceUrl` for cache hit improvement - ALPs have many client-resolvable query parameters which greatly reduce cache-hit by multiplying cache keys. E.g. `https://host/alp?adpos=X` will multiply on the possible space of `adpos` query parameter, even though it's never used by the page itself.
The proposal is to use `https://host/alp` as the cache URL/key and replace it when AMP runtime starts up to `https://host/alp?adpos=X`. The replacing URL will be specified via viewer param as `#replaceUrl=...`.
Some important notes:
1. Replacement should be done, ideally, very early on and definitely before `installDocumentStateService` is called.
2. The replacement URL must match the original URLs origin _and_ source origin
| priority | alp add support for replaceurl for cache hit improvement alps have many client resolvable query parameters which greatly reduce cache hit by multiplying cache keys e g will multiply on the possible space of adpos query parameter even though it s never used by the page itself the proposal is to use as the cache url key and replace it when amp runtime starts up to the replacing url will be specified via viewer param as replaceurl some important notes replacement should be done ideally very early on and definitely before installdocumentstateservice is called the replacement url must match the original urls origin and source origin | 1 |
506,485 | 14,666,161,824 | IssuesEvent | 2020-12-29 15:43:57 | netlify/next-on-netlify | https://api.github.com/repos/netlify/next-on-netlify | reopened | Preview mode broken in Next 10.0.2 with target: "serverless": TypeError: e[t] is not a function | Next 10 experimental-serverless-trace priority: high | Since we've switched to the latest version of next (10.0.2), [preview mode](https://nextjs.org/docs/advanced-features/preview-mode) has been broken. Everything builds successfully, but when we load our `/api/preview` endpoint we get the following error:
```
[request] /api/preview/
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:116
throw ex;
^
abort({}) at Error
at jsStackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1070:13)
at stackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1087:12)
at process.abort (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:8502:44)
at process.emit (events.js:310:20)
at processPromiseRejections (internal/process/promises.js:209:33)
at processTicksAndRejections (internal/process/task_queues.js:98:32)
(Use `node --trace-uncaught ...` to show where the exception was thrown)
``` | 1.0 | Preview mode broken in Next 10.0.2 with target: "serverless": TypeError: e[t] is not a function - Since we've switched to the latest version of next (10.0.2), [preview mode](https://nextjs.org/docs/advanced-features/preview-mode) has been broken. Everything builds successfully, but when we load our `/api/preview` endpoint we get the following error:
```
[request] /api/preview/
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
TypeError: e[t] is not a function
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901)
at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130)
at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078)
/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:116
throw ex;
^
abort({}) at Error
at jsStackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1070:13)
at stackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1087:12)
at process.abort (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:8502:44)
at process.emit (events.js:310:20)
at processPromiseRejections (internal/process/promises.js:209:33)
at processTicksAndRejections (internal/process/task_queues.js:98:32)
(Use `node --trace-uncaught ...` to show where the exception was thrown)
``` | priority | preview mode broken in next with target serverless typeerror e is not a function since we ve switched to the latest version of next has been broken everything builds successfully but when we load our api preview endpoint we get the following error api preview typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js usr local lib node modules netlify cli node modules netlify redirector lib redirects js throw ex abort at error at jsstacktrace usr local lib node modules netlify cli node modules netlify redirector lib redirects js at stacktrace usr local lib node modules netlify cli node modules netlify redirector lib redirects js at process abort usr local lib node modules netlify cli node modules netlify redirector lib redirects js at process emit events js at processpromiserejections internal process promises js at processticksandrejections internal process task queues js use node trace uncaught to show where the exception was thrown | 1 |
602,482 | 18,470,078,286 | IssuesEvent | 2021-10-17 15:34:24 | AY2122S1-CS2103-T14-2/tp | https://api.github.com/repos/AY2122S1-CS2103-T14-2/tp | closed | Store CCA enrolments data locally | type.Story priority.High | As a user, after enrolling contacts into CCAs, I want to have it saved locally so I do not have to re-enrol my contacts every time. | 1.0 | Store CCA enrolments data locally - As a user, after enrolling contacts into CCAs, I want to have it saved locally so I do not have to re-enrol my contacts every time. | priority | store cca enrolments data locally as a user after enrolling contacts into ccas i want to have it saved locally so i do not have to re enrol my contacts every time | 1 |
537,099 | 15,723,056,021 | IssuesEvent | 2021-03-29 06:55:01 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.youtube.com - desktop site instead of mobile site | browser-chrome ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Chrome 89.0.4389 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68920 -->
**URL**: https://www.youtube.com/watch?v=u0KQqOai-uA
**Browser / Version**: Chrome 89.0.4389
**Operating System**: Windows 10
**Tested Another Browser**: Yes Safari
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.youtube.com - desktop site instead of mobile site - <!-- @browser: Chrome 89.0.4389 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68920 -->
**URL**: https://www.youtube.com/watch?v=u0KQqOai-uA
**Browser / Version**: Chrome 89.0.4389
**Operating System**: Windows 10
**Tested Another Browser**: Yes Safari
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | desktop site instead of mobile site url browser version chrome operating system windows tested another browser yes safari problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww browser configuration none from with ❤️ | 1 |
368,313 | 10,877,474,279 | IssuesEvent | 2019-11-16 10:27:27 | trustwallet/blockatlas | https://api.github.com/repos/trustwallet/blockatlas | closed | reformat response for GET {chain}/staking/delegations/{addr} endpoint | Priority: High Task Size: S Type: Enhancement | Make response consistent with POST /staking/delegations/ endpoint | 1.0 | reformat response for GET {chain}/staking/delegations/{addr} endpoint - Make response consistent with POST /staking/delegations/ endpoint | priority | reformat response for get chain staking delegations addr endpoint make response consistent with post staking delegations endpoint | 1 |
556,516 | 16,485,213,813 | IssuesEvent | 2021-05-24 16:55:46 | sopra-fs21-group-11/sopra-client | https://api.github.com/repos/sopra-fs21-group-11/sopra-client | closed | S16: As a player I want to configure the game before it starts in order to get the most suitable game setup. | high priority user story | Before the player starts a game he must be able to:
- [ ] edit the guess countdown before the evaluation.
- [ ] edit the number of cards before an evaluation happens.
- [ ] Edit the number of evaluations.
- [ ] Edit the countdown for doubting a placement.
- [ ] Edit the maximal number of players participating
- [ ] Edit, how card is visible after it gets involved in a doubting process.
- [ ] Edit how long a players turn lasts until he loses a token (S8) | 1.0 | S16: As a player I want to configure the game before it starts in order to get the most suitable game setup. - Before the player starts a game he must be able to:
- [ ] edit the guess countdown before the evaluation.
- [ ] edit the number of cards before an evaluation happens.
- [ ] Edit the number of evaluations.
- [ ] Edit the countdown for doubting a placement.
- [ ] Edit the maximal number of players participating
- [ ] Edit, how card is visible after it gets involved in a doubting process.
- [ ] Edit how long a players turn lasts until he loses a token (S8) | priority | as a player i want to configure the game before it starts in order to get the most suitable game setup before the player starts a game he must be able to edit the guess countdown before the evaluation edit the number of cards before an evaluation happens edit the number of evaluations edit the countdown for doubting a placement edit the maximal number of players participating edit how card is visible after it gets involved in a doubting process edit how long a players turn lasts until he loses a token | 1 |
271,621 | 8,485,895,492 | IssuesEvent | 2018-10-26 09:14:26 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Report the reason why a workflow failed to be acquired | High Priority | When we hit a terminal exception in WorkQueue, we should somehow propagate this information up to CompOps. Maybe in the ReqMgr2 template where it has the state transitions and etc. | 1.0 | Report the reason why a workflow failed to be acquired - When we hit a terminal exception in WorkQueue, we should somehow propagate this information up to CompOps. Maybe in the ReqMgr2 template where it has the state transitions and etc. | priority | report the reason why a workflow failed to be acquired when we hit a terminal exception in workqueue we should somehow propagate this information up to compops maybe in the template where it has the state transitions and etc | 1 |
639,089 | 20,746,471,634 | IssuesEvent | 2022-03-14 23:54:45 | SSWConsulting/SSW.SophieBot | https://api.github.com/repos/SSWConsulting/SSW.SophieBot | closed | ⭐ Query - When is John Smith's next availability | Priority: high | <!-- These comments automatically delete -->
<!-- **Tip:** Delete parts that are not relevant -->
<!-- Next to Cc:, @ mention users who should be in the loop -->
Cc: @theixdev @AttackOnMorty
<!-- add intended user next to **Hi** -->
### Pain
<!-- Explain the pain you are experiencing -->
People want to know when is someone available
### Tasks (only do the easy ones)
<!--Add GitHub tasks-->
- [x] When is John Smith's next availability
- [x] When is John next available
- [x] When will John next be free
- [x] When is John next Free/Available
- [x] When will John finish his current booking
- [x] When is John's next bookable
### Hard Tasks
- [x] When is John next free for 3 days?
Feedback from @theixdev
Thanks!
| 1.0 | ⭐ Query - When is John Smith's next availability - <!-- These comments automatically delete -->
<!-- **Tip:** Delete parts that are not relevant -->
<!-- Next to Cc:, @ mention users who should be in the loop -->
Cc: @theixdev @AttackOnMorty
<!-- add intended user next to **Hi** -->
### Pain
<!-- Explain the pain you are experiencing -->
People want to know when is someone available
### Tasks (only do the easy ones)
<!--Add GitHub tasks-->
- [x] When is John Smith's next availability
- [x] When is John next available
- [x] When will John next be free
- [x] When is John next Free/Available
- [x] When will John finish his current booking
- [x] When is John's next bookable
### Hard Tasks
- [x] When is John next free for 3 days?
Feedback from @theixdev
Thanks!
| priority | ⭐ query when is john smith s next availability cc theixdev attackonmorty pain people want to know when is someone available tasks only do the easy ones when is john smith s next availability when is john next available when will john next be free when is john next free available when will john finish his current booking when is john s next bookable hard tasks when is john next free for days feedback from theixdev thanks | 1 |
498,880 | 14,434,965,239 | IssuesEvent | 2020-12-07 08:01:19 | ahmedkaludi/accelerated-mobile-pages | https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages | closed | The "amphtml" is not generating on the custom taxonomy and the custom post types | NEXT UPDATE [Priority: HIGH] bug | Ref: https://secure.helpscout.net/conversation/1331694976/164006?folderId=2770545
The amphtml is not generating for the custom taxonomy and the custom post types after the 1.0.63 version update.
Custom post type: https://demo6.petclassifieds.com/ad/1-year-old-basset-hound/
Custom taxonomy: https://demo6.petclassifieds.com/ad_category/dogs/
The user is creating the post types and the custom taxonomy using the Adforest theme.
| 1.0 | The "amphtml" is not generating on the custom taxonomy and the custom post types - Ref: https://secure.helpscout.net/conversation/1331694976/164006?folderId=2770545
The amphtml is not generating for the custom taxonomy and the custom post types after the 1.0.63 version update.
Custom post type: https://demo6.petclassifieds.com/ad/1-year-old-basset-hound/
Custom taxonomy: https://demo6.petclassifieds.com/ad_category/dogs/
The user is creating the post types and the custom taxonomy using the Adforest theme.
| priority | the amphtml is not generating on the custom taxonomy and the custom post types ref the amphtml is not generating for the custom taxonomy and the custom post types after the version update custom post type custom taxonomy the user is creating the post types and the custom taxonomy using the adforest theme | 1 |
509,292 | 14,727,737,864 | IssuesEvent | 2021-01-06 08:59:15 | MikeVedsted/JoinMe | https://api.github.com/repos/MikeVedsted/JoinMe | opened | [FEAT] Add search event functionality in homepage | Priority: High :bomb: Status: Received :inbox_tray: Status: Selected for development :pencil: Type: Enhancement :rocket: |
**🚀 I'm ready for take off**
Before submitting, please mark if you:
- [ ] Checked that this feature doesn't already exists
- [ ] Checked that a feature request doesn't already exists
- [ ] Went through the user flow, and understand the impact
- [ ] Made sure the request shows why it is important to users but doesn't exaggerate the value
| 1.0 | [FEAT] Add search event functionality in homepage -
**🚀 I'm ready for take off**
Before submitting, please mark if you:
- [ ] Checked that this feature doesn't already exists
- [ ] Checked that a feature request doesn't already exists
- [ ] Went through the user flow, and understand the impact
- [ ] Made sure the request shows why it is important to users but doesn't exaggerate the value
| priority | add search event functionality in homepage 🚀 i m ready for take off before submitting please mark if you checked that this feature doesn t already exists checked that a feature request doesn t already exists went through the user flow and understand the impact made sure the request shows why it is important to users but doesn t exaggerate the value | 1 |
689,352 | 23,617,802,756 | IssuesEvent | 2022-08-24 17:27:50 | Unity-Technologies/com.unity.netcode.gameobjects | https://api.github.com/repos/Unity-Technologies/com.unity.netcode.gameobjects | closed | When a client connects to a server all clients receive all NetworkVariables | type:bug priority:high stat:imported | Thanks @mraue for the update. I get your issue now. Please let me update it and re-open as it seems pretty important.
### Description
Every time a client connects to a server all `NetworkVariable`s on all `NetworkObject`s get synced to **all** clients.
### Reproduce Steps
1. Create a NetworkObject with a NetworkVariable. Add a log that tracks `OnValueChanged`
2. Start a server and connect a client
3. Connect a second client
4. The first and the second client show the `OnValueChanged` log
### Actual Outcome
All connected clients receive a `NetworkVariable` update.
### Expected Outcome
Only the second client connecting should receive the `NetworkVariable` data.
### Additional Context
During the `internal void SendSpawnCallForObject(ulong clientId, NetworkObject networkObject)` for the NetworkObject to spawn, we mark:
`networkObject.MarkVariablesDirty(true);`
This causes the NetworkVariable to be considered dirty for all clients. As a result, the NetworkVariable gets serialized to all clients.
The difficulty lies with the fact that current Netcode for GameObjects doesn't have a way to mark dirtiness for some clients only. This will require a bit of refactor.
| 1.0 | When a client connects to a server all clients receive all NetworkVariables - Thanks @mraue for the update. I get your issue now. Please let me update it and re-open as it seems pretty important.
### Description
Every time a client connects to a server all `NetworkVariable`s on all `NetworkObject`s get synced to **all** clients.
### Reproduce Steps
1. Create a NetworkObject with a NetworkVariable. Add a log that tracks `OnValueChanged`
2. Start a server and connect a client
3. Connect a second client
4. The first and the second client show the `OnValueChanged` log
### Actual Outcome
All connected clients receive a `NetworkVariable` update.
### Expected Outcome
Only the second client connecting should receive the `NetworkVariable` data.
### Additional Context
During the `internal void SendSpawnCallForObject(ulong clientId, NetworkObject networkObject)` for the NetworkObject to spawn, we mark:
`networkObject.MarkVariablesDirty(true);`
This causes the NetworkVariable to be considered dirty for all clients. As a result, the NetworkVariable gets serialized to all clients.
The difficulty lies with the fact that current Netcode for GameObjects doesn't have a way to mark dirtiness for some clients only. This will require a bit of refactor.
| priority | when a client connects to a server all clients receive all networkvariables thanks mraue for the update i get your issue now please let me update it and re open as it seems pretty important description every time a client connects to a server all networkvariable s on all networkobject s get synced to all clients reproduce steps create a networkobject with a networkvariable add a log that tracks onvaluechanged start a server and connect a client connect a second client the first and the second client show the onvaluechanged log actual outcome all connected clients receive a networkvariable update expected outcome only the second client connecting should receive the networkvariable data additional context during the internal void sendspawncallforobject ulong clientid networkobject networkobject for the networkobject to spawn we mark networkobject markvariablesdirty true this causes the networkvariable to be considered dirty for all clients as a result the networkvariable gets serialized to all clients the difficulty lies with the fact that current netcode for gameobjects doesn t have a way to mark dirtiness for some clients only this will require a bit of refactor | 1 |
541,259 | 15,823,859,508 | IssuesEvent | 2021-04-06 01:51:00 | istopwg/ippsample | https://api.github.com/repos/istopwg/ippsample | reopened | ipptool: Add directive to monitor "printer-state" or "printer-state-reasons" in the background during a test | enhancement priority-high | Recent issues reported against the IPP Everywhere Self Certification Test I-20 and I-20.1 where I-20 deadlocks because PWG Raster is handled as a streaming document format, which prevents I-20.1 from checking the "printer-state-reasons", causes a condition where the tests cannot complete successfully. What is needed is a way for the tests to be run in parallel. But we don't want to create a complex general purpose parallel testing facility - that could create a monster.
However, from discussions on the IPP WG reflector and in recent teleconferences, it was discussed that, if "ipptool" were able to monitor "printer-state-reasons" in a background thread while performing the test in the foreground thread, this would be useful for this purpose and for monitoring the completion of the action triggered by the operation.
Add to ipptool support for a new "MONITOR-PRINTER-STATE" directive like so:
MONITOR-PRINTER-STATE predicate [timeout]
where ipptool would perform a polling Get-Printer-Attributes operation looking for the matching Printer state and state reason conditions defined by "predicate", and "timeout" would specify a hard time limit after which the MONITOR-PRINTER-STATE would fail if the matching status wasn't achieved. (We can noodle on the design of this...)
Also define a MONITOR-JOB-STATE that could be used to monitor a Job state for completion:
MONITOR-JOB-STATE jobid predicate [timeout]
The logic would be basically the same but would poll the Job specified by jobid with Get-Job-Attributes until the matching conditions were achieved or the timeout occurred. | 1.0 | ipptool: Add directive to monitor "printer-state" or "printer-state-reasons" in the background during a test - Recent issues reported against the IPP Everywhere Self Certification Test I-20 and I-20.1 where I-20 deadlocks because PWG Raster is handled as a streaming document format, which prevents I-20.1 from checking the "printer-state-reasons", causes a condition where the tests cannot complete successfully. What is needed is a way for the tests to be run in parallel. But we don't want to create a complex general purpose parallel testing facility - that could create a monster.
However, from discussions on the IPP WG reflector and in recent teleconferences, it was discussed that, if "ipptool" were able to monitor "printer-state-reasons" in a background thread while performing the test in the foreground thread, this would be useful for this purpose and for monitoring the completion of the action triggered by the operation.
Add to ipptool support for a new "MONITOR-PRINTER-STATE" directive like so:
MONITOR-PRINTER-STATE predicate [timeout]
where ipptool would perform a polling Get-Printer-Attributes operation looking for the matching Printer state and state reason conditions defined by "predicate", and "timeout" would specify a hard time limit after which the MONITOR-PRINTER-STATE would fail if the matching status wasn't achieved. (We can noodle on the design of this...)
Also define a MONITOR-JOB-STATE that could be used to monitor a Job state for completion:
MONITOR-JOB-STATE jobid predicate [timeout]
The logic would be basically the same but would poll the Job specified by jobid with Get-Job-Attributes until the matching conditions were achieved or the timeout occurred. | priority | ipptool add directive to monitor printer state or printer state reasons in the background during a test recent issues reported against the ipp everywhere self certification test i and i where i deadlocks because pwg raster is handled as a streaming document format which prevents i from checking the printer state reasons causes a condition where the tests cannot complete successfully what is needed is a way for the tests to be run in parallel but we don t want to create a complex general purpose parallel testing facility that could create a monster however from discussions on the ipp wg reflector and in recent teleconferences it was discussed that if ipptool were able to monitor printer state reasons in a background thread while performing the test in the foreground thread this would be useful for this purpose and for monitoring the completion of the action triggered by the operation add to ipptool support for a new monitor printer state directive like so monitor printer state predicate where ipptool would perform a polling get printer attributes operation looking for the matching printer state and state reason conditions defined by predicate and timeout would specify a hard time limit after which the monitor printer state would fail if the matching status wasn t achieved we can noodle on the design of this also define a monitor job state that could be used to monitor a job state for completion monitor job state jobid predicate the logic would be basically the same but would poll the job specified by jobid with get job attributes until the matching conditions were achieved or the timeout occurred | 1 |
620,606 | 19,565,670,566 | IssuesEvent | 2022-01-03 23:39:35 | bounswe/2021SpringGroup12 | https://api.github.com/repos/bounswe/2021SpringGroup12 | closed | Search a goal in an exact way | priority: high android | **Description**
- Users should be able to see the goals that are published that includes the searched query inside the title, desription or tags
| 1.0 | Search a goal in an exact way - **Description**
- Users should be able to see the goals that are published that includes the searched query inside the title, desription or tags
| priority | search a goal in an exact way description users should be able to see the goals that are published that includes the searched query inside the title desription or tags | 1 |
376,263 | 11,140,471,143 | IssuesEvent | 2019-12-21 14:39:45 | wilcommerce/Wilcommerce.Catalog | https://api.github.com/repos/wilcommerce/Wilcommerce.Catalog | closed | Change model description throws an error if description is empty | priority:high status:done type:bug | The method _ChangeDescription_ in the _Brand_, _Category_ and _Product_ entities throws an _ArgumentNullException_ if the description argument is null or empty.
This behavior doesn't allow to set an empty description for the specified entities. | 1.0 | Change model description throws an error if description is empty - The method _ChangeDescription_ in the _Brand_, _Category_ and _Product_ entities throws an _ArgumentNullException_ if the description argument is null or empty.
This behavior doesn't allow to set an empty description for the specified entities. | priority | change model description throws an error if description is empty the method changedescription in the brand category and product entities throws an argumentnullexception if the description argument is null or empty this behavior doesn t allow to set an empty description for the specified entities | 1 |
311,622 | 9,536,674,501 | IssuesEvent | 2019-04-30 10:21:13 | aartiukh/sph | https://api.github.com/repos/aartiukh/sph | opened | Fix unit tests in sph library | area: c++ priority: high type: bug | **Describe the bug**
Currently unit tests are disabled in `sph` CMake file https://github.com/aartiukh/sph/blob/develop/sph/CMakeLists.txt#L22
**To Reproduce**
Steps to reproduce the behavior:
1. Build `sph-sdk`
2. Run tests
3. Ensure that `sph` tests are not running
**Expected behavior**
The `sph` tests should work.
| 1.0 | Fix unit tests in sph library - **Describe the bug**
Currently unit tests are disabled in `sph` CMake file https://github.com/aartiukh/sph/blob/develop/sph/CMakeLists.txt#L22
**To Reproduce**
Steps to reproduce the behavior:
1. Build `sph-sdk`
2. Run tests
3. Ensure that `sph` tests are not running
**Expected behavior**
The `sph` tests should work.
| priority | fix unit tests in sph library describe the bug currently unit tests are disabled in sph cmake file to reproduce steps to reproduce the behavior build sph sdk run tests ensure that sph tests are not running expected behavior the sph tests should work | 1 |
222,403 | 7,431,931,558 | IssuesEvent | 2018-03-25 19:32:02 | smit-happens/YCP_EVOS | https://api.github.com/repos/smit-happens/YCP_EVOS | closed | Handle Precharge done signal | priority-high size-small stage unitek | <!--
Issue template
To Use this Template:
* Fill out what you can
* Delete what you do not fill out
-->
## End goal
Transition to energized stage when precharge done line is enabled by the Unitek logic board
| 1.0 | Handle Precharge done signal - <!--
Issue template
To Use this Template:
* Fill out what you can
* Delete what you do not fill out
-->
## End goal
Transition to energized stage when precharge done line is enabled by the Unitek logic board
| priority | handle precharge done signal issue template to use this template fill out what you can delete what you do not fill out end goal transition to energized stage when precharge done line is enabled by the unitek logic board | 1 |
50,922 | 3,008,142,753 | IssuesEvent | 2015-07-27 19:41:12 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | PeopleInSpace: decrease or eliminate z-index | Bug Low-Hanging Fruit Priority: High | setting the z-index so high conflicts with other answerbar elements.
set it low, like 2, or find a better way to line up the flags without using z-index.
------
http://duck.co/ia/view/people_in_space | 1.0 | PeopleInSpace: decrease or eliminate z-index - setting the z-index so high conflicts with other answerbar elements.
set it low, like 2, or find a better way to line up the flags without using z-index.
------
http://duck.co/ia/view/people_in_space | priority | peopleinspace decrease or eliminate z index setting the z index so high conflicts with other answerbar elements set it low like or find a better way to line up the flags without using z index | 1 |
257,279 | 8,135,296,338 | IssuesEvent | 2018-08-20 01:53:03 | collinbarrett/FilterLists | https://api.github.com/repos/collinbarrett/FilterLists | opened | fix SnapshotService algorithm trying to insert duplicate raw rules | agent bug high priority | Since adding a unique index on rules.Raw in #353 , I can now see exceptions like below from the [SnapshotService](https://github.com/collinbarrett/FilterLists/tree/master/src/FilterLists.Services/Snapshot) that represent a flaw in the algorithm. It should not try to insert duplicate raw rules, but it is fairly often for some reason.
Example:
> FilterListId: 297 Exception: An error occurred while updating the entries. See the inner exception for details. at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.ReaderModificationCommandBatch.ExecuteAsync(IRelationalConnection connection, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.Internal.BatchExecutor.ExecuteAsync(DbContext _, ValueTuple`2 parameters, CancellationToken cancellationToken) at Pomelo.EntityFrameworkCore.MySql.Storage.Internal.MySqlExecutionStrategy.ExecuteAsync[TState,TResult](TState state, Func`4 operation, Func`4 verifySucceeded, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(IReadOnlyList`1 entriesToSave, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.DbContext.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at FilterLists.Services.Snapshot.SnapshotBatch.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/SnapshotBatch.cs:line 32 at FilterLists.Services.Snapshot.Snapshot.SaveBatches(IEnumerable`1 batches) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 122 at FilterLists.Services.Snapshot.Snapshot.SaveInBatches(IEnumerable`1 lines) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 114 at FilterLists.Services.Snapshot.Snapshot.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 65 at FilterLists.Services.Snapshot.Snapshot.TrySaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 43 Duplicate entry '###Meebo\:AdElement\.Root' for key 'IX_rules_Raw' at MySql.Data.MySqlClient.MySqlDataReader.ActivateResultSet(ResultSet resultSet) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 93 at MySql.Data.MySqlClient.MySqlDataReader.NextResultAsync(IOBehavior ioBehavior, CancellationToken cancellationToken) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 62 at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) | 1.0 | fix SnapshotService algorithm trying to insert duplicate raw rules - Since adding a unique index on rules.Raw in #353 , I can now see exceptions like below from the [SnapshotService](https://github.com/collinbarrett/FilterLists/tree/master/src/FilterLists.Services/Snapshot) that represent a flaw in the algorithm. It should not try to insert duplicate raw rules, but it is fairly often for some reason.
Example:
> FilterListId: 297 Exception: An error occurred while updating the entries. See the inner exception for details. at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.ReaderModificationCommandBatch.ExecuteAsync(IRelationalConnection connection, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.Internal.BatchExecutor.ExecuteAsync(DbContext _, ValueTuple`2 parameters, CancellationToken cancellationToken) at Pomelo.EntityFrameworkCore.MySql.Storage.Internal.MySqlExecutionStrategy.ExecuteAsync[TState,TResult](TState state, Func`4 operation, Func`4 verifySucceeded, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(IReadOnlyList`1 entriesToSave, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.DbContext.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at FilterLists.Services.Snapshot.SnapshotBatch.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/SnapshotBatch.cs:line 32 at FilterLists.Services.Snapshot.Snapshot.SaveBatches(IEnumerable`1 batches) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 122 at FilterLists.Services.Snapshot.Snapshot.SaveInBatches(IEnumerable`1 lines) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 114 at FilterLists.Services.Snapshot.Snapshot.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 65 at FilterLists.Services.Snapshot.Snapshot.TrySaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 43 Duplicate entry '###Meebo\:AdElement\.Root' for key 'IX_rules_Raw' at MySql.Data.MySqlClient.MySqlDataReader.ActivateResultSet(ResultSet resultSet) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 93 at MySql.Data.MySqlClient.MySqlDataReader.NextResultAsync(IOBehavior ioBehavior, CancellationToken cancellationToken) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 62 at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) | priority | fix snapshotservice algorithm trying to insert duplicate raw rules since adding a unique index on rules raw in i can now see exceptions like below from the that represent a flaw in the algorithm it should not try to insert duplicate raw rules but it is fairly often for some reason example filterlistid exception an error occurred while updating the entries see the inner exception for details at microsoft entityframeworkcore update affectedcountmodificationcommandbatch consumeasync relationaldatareader reader cancellationtoken cancellationtoken at microsoft entityframeworkcore update readermodificationcommandbatch executeasync irelationalconnection connection cancellationtoken cancellationtoken at microsoft entityframeworkcore update internal batchexecutor executeasync dbcontext valuetuple parameters cancellationtoken cancellationtoken at pomelo entityframeworkcore mysql storage internal mysqlexecutionstrategy executeasync tstate state func operation func verifysucceeded cancellationtoken cancellationtoken at microsoft entityframeworkcore changetracking internal statemanager savechangesasync ireadonlylist entriestosave cancellationtoken cancellationtoken at microsoft entityframeworkcore changetracking internal statemanager savechangesasync boolean acceptallchangesonsuccess cancellationtoken cancellationtoken at microsoft entityframeworkcore dbcontext savechangesasync boolean acceptallchangesonsuccess cancellationtoken cancellationtoken at filterlists services snapshot snapshotbatch saveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshotbatch cs line at filterlists services snapshot snapshot savebatches ienumerable batches in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot saveinbatches ienumerable lines in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot saveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot trysaveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line duplicate entry meebo adelement root for key ix rules raw at mysql data mysqlclient mysqldatareader activateresultset resultset resultset in c projects mysqlconnector src mysqlconnector mysql data mysqlclient mysqldatareader cs line at mysql data mysqlclient mysqldatareader nextresultasync iobehavior iobehavior cancellationtoken cancellationtoken in c projects mysqlconnector src mysqlconnector mysql data mysqlclient mysqldatareader cs line at microsoft entityframeworkcore update affectedcountmodificationcommandbatch consumeasync relationaldatareader reader cancellationtoken cancellationtoken | 1 |
283,504 | 8,719,731,675 | IssuesEvent | 2018-12-08 03:45:21 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | build_visit2_7_0 fails building python with icc. | bug likelihood medium priority reviewed severity high | Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \
console nothirdparty novisit \
mesa cmake python vtk qt icet pyside \
hdf5 silo szip \
makeflags -j4
Here is the Python failure:
icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC
Ibuild/temp.linuxx86_64-2.7/libffi/include
Ibuild/temp.linuxx86_64-2.7/libffi
I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src
I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include
I. IInclude I./Include I/usr/local/include
I/nfs/tmp2/brugger/build_python/Python2.7.5/Include
I/nfs/tmp2/brugger/build_python/Python2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
Wall fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing
fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259
DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT
DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE
DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES
DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include
I../../../include/QtCore I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o
.obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o
.obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o
.obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o
.obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o
.obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o
.obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o
.obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o
.obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o
.obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o
.obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o
.obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o
.obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o
.obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o
.obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o
.obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o
.obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o
.obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o
.obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o
.obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o
.obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o
.obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: ***** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap'
make: ***** [subtoolsbootstrapmake_defaultordered] Error 2
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1651
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: build_visit2_7_0 fails building python with icc.
Assigned to: Kathleen Biagas
Category:
Target version: 2.7.1
Author: Eric Brugger
Start: 11/08/2013
Due date:
% Done: 0
Estimated time:
Created: 11/08/2013 05:32 pm
Updated: 12/11/2013 05:33 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \
--console --no-thirdparty --no-visit \
--mesa --cmake --python --vtk --qt --icet --pyside \
--hdf5 --silo --szip \
--makeflags -j4
Here is the Python failure:
icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC
-Ibuild/temp.linux-x86_64-2.7/libffi/include
-Ibuild/temp.linux-x86_64-2.7/libffi
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src
-I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include
-I. -IInclude -I./Include -I/usr/local/include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
-Wall -fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing
-fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259
-DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
-DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT
-DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE
-DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES
-DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
-D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include
-I../../../include/QtCore -I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o
.obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o
.obj/release-static/qutfcodec.o .obj/release-static/qglobal.o
.obj/release-static/qmalloc.o .obj/release-static/qnumeric.o
.obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o
.obj/release-static/qdatastream.o .obj/release-static/qdir.o
.obj/release-static/qdiriterator.o .obj/release-static/qfile.o
.obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o
.obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o
.obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o
.obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o
.obj/release-static/qmetatype.o .obj/release-static/qvariant.o
.obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o
.obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o
.obj/release-static/qdatetime.o .obj/release-static/qhash.o
.obj/release-static/qlist.o .obj/release-static/qlocale.o
.obj/release-static/qlocale_tools.o .obj/release-static/qmap.o
.obj/release-static/qregexp.o .obj/release-static/qstring.o
.obj/release-static/qstringlist.o .obj/release-static/qvector.o
.obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o
.obj/release-static/qxmlstream.o .obj/release-static/qdom.o
.obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: *** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap'
make: *** [sub-tools-bootstrap-make_default-ordered] Error 2
Comments:
Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari
The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT).
2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
| 1.0 | build_visit2_7_0 fails building python with icc. - Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \
console nothirdparty novisit \
mesa cmake python vtk qt icet pyside \
hdf5 silo szip \
makeflags -j4
Here is the Python failure:
icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC
Ibuild/temp.linuxx86_64-2.7/libffi/include
Ibuild/temp.linuxx86_64-2.7/libffi
I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src
I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include
I. IInclude I./Include I/usr/local/include
I/nfs/tmp2/brugger/build_python/Python2.7.5/Include
I/nfs/tmp2/brugger/build_python/Python2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
Wall fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing
fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259
DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT
DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE
DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES
DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include
I../../../include/QtCore I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o
.obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o
.obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o
.obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o
.obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o
.obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o
.obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o
.obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o
.obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o
.obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o
.obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o
.obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o
.obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o
.obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o
.obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o
.obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o
.obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o
.obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o
.obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o
.obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o
.obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o
.obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: ***** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap'
make: ***** [subtoolsbootstrapmake_defaultordered] Error 2
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1651
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: build_visit2_7_0 fails building python with icc.
Assigned to: Kathleen Biagas
Category:
Target version: 2.7.1
Author: Eric Brugger
Start: 11/08/2013
Due date:
% Done: 0
Estimated time:
Created: 11/08/2013 05:32 pm
Updated: 12/11/2013 05:33 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \
--console --no-thirdparty --no-visit \
--mesa --cmake --python --vtk --qt --icet --pyside \
--hdf5 --silo --szip \
--makeflags -j4
Here is the Python failure:
icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC
-Ibuild/temp.linux-x86_64-2.7/libffi/include
-Ibuild/temp.linux-x86_64-2.7/libffi
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src
-I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include
-I. -IInclude -I./Include -I/usr/local/include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
-Wall -fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing
-fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259
-DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
-DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT
-DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE
-DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES
-DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
-D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include
-I../../../include/QtCore -I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o
.obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o
.obj/release-static/qutfcodec.o .obj/release-static/qglobal.o
.obj/release-static/qmalloc.o .obj/release-static/qnumeric.o
.obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o
.obj/release-static/qdatastream.o .obj/release-static/qdir.o
.obj/release-static/qdiriterator.o .obj/release-static/qfile.o
.obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o
.obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o
.obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o
.obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o
.obj/release-static/qmetatype.o .obj/release-static/qvariant.o
.obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o
.obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o
.obj/release-static/qdatetime.o .obj/release-static/qhash.o
.obj/release-static/qlist.o .obj/release-static/qlocale.o
.obj/release-static/qlocale_tools.o .obj/release-static/qmap.o
.obj/release-static/qregexp.o .obj/release-static/qstring.o
.obj/release-static/qstringlist.o .obj/release-static/qvector.o
.obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o
.obj/release-static/qxmlstream.o .obj/release-static/qdom.o
.obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: *** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap'
make: *** [sub-tools-bootstrap-make_default-ordered] Error 2
Comments:
Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari
The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT).
2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
| priority | build fails building python with icc rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapichgnu include build console nothirdparty novisit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fnostrictaliasing g dndebug fpic fpic ibuild temp libffi include ibuild temp libffi i nfs brugger build python modules ctypes libffi src i nfs brugger build python visit python icc include i iinclude i include i usr local include i nfs brugger build python include i nfs brugger build python c nfs brugger build python python modules ctypes libffi src c o build temp nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falignfunctions ansialias fstrict aliasing fpic fnojumptables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj releasestatic qlatincodec o obj releasestatic qsimplecodec o obj releasestatic qtextcodec o obj releasestatic qtsciicodec o obj releasestatic qutfcodec o obj releasestatic qglobal o obj releasestatic qmalloc o obj releasestatic qnumeric o obj releasestatic qabstractfileengine o obj releasestatic qbuffer o obj releasestatic qdatastream o obj releasestatic qdir o obj releasestatic qdiriterator o obj releasestatic qfile o obj releasestatic qfileinfo o obj releasestatic qfilesystementry o obj releasestatic qfilesystemengine o obj releasestatic qfsfileengine o obj releasestatic qfsfileengine iterator o obj releasestatic qiodevice o obj releasestatic qtemporaryfile o obj releasestatic qtextstream o obj releasestatic qmetatype o obj releasestatic qvariant o obj releasestatic qsystemerror o obj releasestatic qbitarray o obj releasestatic qbytearray o obj releasestatic qbytearraymatcher o obj releasestatic qdatetime o obj releasestatic qhash o obj releasestatic qlist o obj releasestatic qlocale o obj releasestatic qlocale tools o obj releasestatic qmap o obj releasestatic qregexp o obj releasestatic qstring o obj releasestatic qstringlist o obj releasestatic qvector o obj releasestatic qvsnprintf o obj releasestatic qxmlutils o obj releasestatic qxmlstream o obj releasestatic qdom o obj releasestatic qxml o obj releasestatic qfilesystemengine unix o obj release static qfilesystemiterator unix o obj releasestatic qfsfileengine unix o obj releasestatic qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python src tools bootstrap make error redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject build fails building python with icc assigned to kathleen biagas category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version impact expected use os all support group any description rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapich gnu include build console no thirdparty no visit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fno strict aliasing g dndebug fpic fpic ibuild temp linux libffi include ibuild temp linux libffi i nfs brugger build python python modules ctypes libffi src i nfs brugger build python visit python linux icc include i iinclude i include i usr local include i nfs brugger build python python include i nfs brugger build python python c nfs brugger build python python modules ctypes libffi src c o build temp linux nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falign functions ansi alias fstrict aliasing fpic fno jump tables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs linux icc i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj release static qlatincodec o obj release static qsimplecodec o obj release static qtextcodec o obj release static qtsciicodec o obj release static qutfcodec o obj release static qglobal o obj release static qmalloc o obj release static qnumeric o obj release static qabstractfileengine o obj release static qbuffer o obj release static qdatastream o obj release static qdir o obj release static qdiriterator o obj release static qfile o obj release static qfileinfo o obj release static qfilesystementry o obj release static qfilesystemengine o obj release static qfsfileengine o obj release static qfsfileengine iterator o obj release static qiodevice o obj release static qtemporaryfile o obj release static qtextstream o obj release static qmetatype o obj release static qvariant o obj release static qsystemerror o obj release static qbitarray o obj release static qbytearray o obj release static qbytearraymatcher o obj release static qdatetime o obj release static qhash o obj release static qlist o obj release static qlocale o obj release static qlocale tools o obj release static qmap o obj release static qregexp o obj release static qstring o obj release static qstringlist o obj release static qvector o obj release static qvsnprintf o obj release static qxmlutils o obj release static qxmlstream o obj release static qdom o obj release static qxml o obj release static qfilesystemengine unix o obj release static qfilesystemiterator unix o obj release static qfsfileengine unix o obj release static qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python qt everywhere opensource src src tools bootstrap make error comments hi eric the python error may not be critical at least for now looks like qt is configuring the ar command to be xiar mkspecs linux icc qmake conf qmake ar xiar cqsa quick fix might be to set a simple alias xiar ar or modify the qmake conf file to what ever might be appropriate call to create archives for your system best regards hari the qt build issue eric encountered is due to the lc setup we need to issue a use icc before attempting to build as doing so makes available the full suite of intel compiler tools including xiar used by qt update merge to trunk minor python patch to bv python fixing issue when compiling with icc the patch is from a newer version of library | 1 |
152,423 | 5,846,832,704 | IssuesEvent | 2017-05-10 17:03:57 | hecoding/Pac-Man | https://api.github.com/repos/hecoding/Pac-Man | closed | Fantasmas siempre "starterghosts" en la partida que usa la GUI para mostrar ejecuciones del mejor individuo encontrado | bug GUI high priority | Fix asap | 1.0 | Fantasmas siempre "starterghosts" en la partida que usa la GUI para mostrar ejecuciones del mejor individuo encontrado - Fix asap | priority | fantasmas siempre starterghosts en la partida que usa la gui para mostrar ejecuciones del mejor individuo encontrado fix asap | 1 |
381,248 | 11,275,454,480 | IssuesEvent | 2020-01-14 20:48:38 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode "wavelets" cause darktable to crash | bug: wip priority: high reproduce: confirmed understood: clear | <!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode set to "wavelets" cause darktable to crash with "An unhandled exception occured" box showing up. The same behaviour observed when zoom is set to 1600% or 800% and "denoise (profiled)" module with the same mode is activated.
**To Reproduce**
<!-- Provide detailed steps that can reproduce the behavior, such as:-->
1. Go to 'Correction group'
2. Click on 'Denoise (profiled)'
3. Set 'mode' to 'wavelets'
5. Change zoom to 1600% either by clicking 1600% or scrolling mouse wheel with Ctr
4. See error
Alternatively:
1. Go to 'Correction group'
2. Click on 'Denoise (profiled)'
3. Set 'mode' to 'wavelets'
4. Deactivate 'Denoise (profiled)' module
5. Change zoon to 1600%
6. Activate 'Denoise (profiled)' module
7. See error
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
Should not crash
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->

**Platform (please complete the following information):**
- Darktable Version: <!-- [e.g. 2.6.0] --> darktable 3.0.0
- OS: <!-- [e.g. Windows 8.1, Gentoo Linux] --> Windows 10
- CPU: Intel(R) Xeon(R) CPU E5-1620 v3 @ 3.50GHz
**Additional context**
<!-- Add any other context about the problem here, for example:
- Can you reproduce with another Darktable version?
- Can you reproduce with a RAW or Jpeg or both?
- Are the steps above reproduce with a fresh edit (removing history)?
- Attach an XMP if this is necessary
- Did you compile Darktable yourself? If so which compiler was used, with what options?
- Is the issue still present using an empty/new config-dir
-->
This is observable either with previously edited and with new raws. Backtrace file is usually almost empty, I also add another bactrace file generated after the same error but with content.
[darktable_bt_68B0D0.txt](https://github.com/darktable-org/darktable/files/4049412/darktable_bt_68B0D0.txt)
[darktable_bt_6FB0D0.txt](https://github.com/darktable-org/darktable/files/4049413/darktable_bt_6FB0D0.txt)
| 1.0 | Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode "wavelets" cause darktable to crash - <!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode set to "wavelets" cause darktable to crash with "An unhandled exception occured" box showing up. The same behaviour observed when zoom is set to 1600% or 800% and "denoise (profiled)" module with the same mode is activated.
**To Reproduce**
<!-- Provide detailed steps that can reproduce the behavior, such as:-->
1. Go to 'Correction group'
2. Click on 'Denoise (profiled)'
3. Set 'mode' to 'wavelets'
5. Change zoom to 1600% either by clicking 1600% or scrolling mouse wheel with Ctr
4. See error
Alternatively:
1. Go to 'Correction group'
2. Click on 'Denoise (profiled)'
3. Set 'mode' to 'wavelets'
4. Deactivate 'Denoise (profiled)' module
5. Change zoon to 1600%
6. Activate 'Denoise (profiled)' module
7. See error
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
Should not crash
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->

**Platform (please complete the following information):**
- Darktable Version: <!-- [e.g. 2.6.0] --> darktable 3.0.0
- OS: <!-- [e.g. Windows 8.1, Gentoo Linux] --> Windows 10
- CPU: Intel(R) Xeon(R) CPU E5-1620 v3 @ 3.50GHz
**Additional context**
<!-- Add any other context about the problem here, for example:
- Can you reproduce with another Darktable version?
- Can you reproduce with a RAW or Jpeg or both?
- Are the steps above reproduce with a fresh edit (removing history)?
- Attach an XMP if this is necessary
- Did you compile Darktable yourself? If so which compiler was used, with what options?
- Is the issue still present using an empty/new config-dir
-->
This is observable either with previously edited and with new raws. Backtrace file is usually almost empty, I also add another bactrace file generated after the same error but with content.
[darktable_bt_68B0D0.txt](https://github.com/darktable-org/darktable/files/4049412/darktable_bt_68B0D0.txt)
[darktable_bt_6FB0D0.txt](https://github.com/darktable-org/darktable/files/4049413/darktable_bt_6FB0D0.txt)
| priority | zooming in to or when denoise profiled module is on with mode wavelets cause darktable to crash important bug reports that do not make an effort to help the developers will be closed without notice make sure that this bug has not already been opened and or closed by searching the issues on github as duplicate bug reports will be closed a bug report simply stating that darktable crashes is unhelpful so please fill in most of the items below and provide detailed information describe the bug zooming in to or when denoise profiled module is on with mode set to wavelets cause darktable to crash with an unhandled exception occured box showing up the same behaviour observed when zoom is set to or and denoise profiled module with the same mode is activated to reproduce go to correction group click on denoise profiled set mode to wavelets change zoom to either by clicking or scrolling mouse wheel with ctr see error alternatively go to correction group click on denoise profiled set mode to wavelets deactivate denoise profiled module change zoon to activate denoise profiled module see error expected behavior should not crash screenshots platform please complete the following information darktable version darktable os windows cpu intel r xeon r cpu additional context add any other context about the problem here for example can you reproduce with another darktable version can you reproduce with a raw or jpeg or both are the steps above reproduce with a fresh edit removing history attach an xmp if this is necessary did you compile darktable yourself if so which compiler was used with what options is the issue still present using an empty new config dir this is observable either with previously edited and with new raws backtrace file is usually almost empty i also add another bactrace file generated after the same error but with content | 1 |
444,415 | 12,811,911,588 | IssuesEvent | 2020-07-04 02:15:41 | ctm/mb2-doc | https://api.github.com/repos/ctm/mb2-doc | opened | JADC got the big blind twice due to new player motion code | bug high priority | odbphatmackakabig got moved from a table with 7 players to our table where we only had 5. There logically was only one place where he could go, which was between JADC, who had just paid the big blind and me, who had just paid the small blind.
I should have gotten the button, which I did, odbphatmackakabig should have had to sit out (instead he paid the small blind) and JADC should have gotten the small blind.
I think this just shows that my implementation of #247 was flawed. I'll look at it when I'm awake tomorrow, but probably after I've exercised.
```
Dealing Pot-Limit Binglaha
deadhead blinds 2000
JADC blinds 4000
1 JADC 2 1850 4000 [ ] [ ] [ ] [ ]
2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ]
3 nnr13 0 24750 0 [ ] [ ] [ ] [ ]
4 B Meeeee2 2 11250 0 [ ] [ ] [ ] [ ]
5 deadhead 2 46000 2000 [ ] [ ] [ ] [ ]
My cards are: [Kd] [6c] [8c] [8s]
deadhead: gg
tombayz folds
nnr13: gg grizz
nnr13 folds
grizz: ty
Meeeee2 folds
It is 2000 to you
deadhead calls
JADC checks
1 JADC 2 1850 4000 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 > deadhead 2 44000 4000 [ ] [ ] [ ] [ ]
Board: 5s 8d Tc
My cards are: [Kd] [6c] [8c] [8s]
It is 0 to you
deadhead bets 4000
JADC asked for and gets 60 more seconds
odbphatmackakabig has come from Table 277
JADC will be turning in in 2 lammers before the next hand
JADC calls and is all-in
Meeeee2 rolled a 4
1 > JADC 2 0 5850 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ]
6 [odbphatmackakabig] 1 5750[ 0]
Board: 5s 8d Tc 7h
My cards are: [Kd] [6c] [8c] [8s]
1 > JADC 2 0 5850 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ]
6 [odbphatmackakabig] 1 5750[ 0]
Board: 5s 8d Tc 7h Th
My cards are: [Kd] [6c] [8c] [8s]
Uncalled 2150 returned to deadhead
deadhead Kd 6c 8c 8s
High Full House (Eights over Tens) 8c 8d 8s Th Tc
Low none
JADC 5d 7d 2h 9h
High Two Pair (Tens and Sevens) Tc Th 7h 7d 9h
Low none
deadhead wins 11700 for High
JADC redeemed a lammer for 5000
JADC redeemed a lammer for 5000
Dealing Pot-Limit Binglaha
odbphatmackakabig blinds 2000
JADC blinds 4000
1 JADC 0 6000 4000 [ ] [ ] [ ] [ ]
2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ]
3 nnr13 0 24750 0 [ ] [ ] [ ] [ ]
4 Meeeee2 2 11250 0 [ ] [ ] [ ] [ ]
5 B deadhead 2 53850 0 [ ] [ ] [ ] [ ]
6 odbphatmackakabig 1 3750 2000 [ ] [ ] [ ] [ ]
My cards are: [7s] [7h] [4d] [Ac]
``` | 1.0 | JADC got the big blind twice due to new player motion code - odbphatmackakabig got moved from a table with 7 players to our table where we only had 5. There logically was only one place where he could go, which was between JADC, who had just paid the big blind and me, who had just paid the small blind.
I should have gotten the button, which I did, odbphatmackakabig should have had to sit out (instead he paid the small blind) and JADC should have gotten the small blind.
I think this just shows that my implementation of #247 was flawed. I'll look at it when I'm awake tomorrow, but probably after I've exercised.
```
Dealing Pot-Limit Binglaha
deadhead blinds 2000
JADC blinds 4000
1 JADC 2 1850 4000 [ ] [ ] [ ] [ ]
2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ]
3 nnr13 0 24750 0 [ ] [ ] [ ] [ ]
4 B Meeeee2 2 11250 0 [ ] [ ] [ ] [ ]
5 deadhead 2 46000 2000 [ ] [ ] [ ] [ ]
My cards are: [Kd] [6c] [8c] [8s]
deadhead: gg
tombayz folds
nnr13: gg grizz
nnr13 folds
grizz: ty
Meeeee2 folds
It is 2000 to you
deadhead calls
JADC checks
1 JADC 2 1850 4000 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 > deadhead 2 44000 4000 [ ] [ ] [ ] [ ]
Board: 5s 8d Tc
My cards are: [Kd] [6c] [8c] [8s]
It is 0 to you
deadhead bets 4000
JADC asked for and gets 60 more seconds
odbphatmackakabig has come from Table 277
JADC will be turning in in 2 lammers before the next hand
JADC calls and is all-in
Meeeee2 rolled a 4
1 > JADC 2 0 5850 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ]
6 [odbphatmackakabig] 1 5750[ 0]
Board: 5s 8d Tc 7h
My cards are: [Kd] [6c] [8c] [8s]
1 > JADC 2 0 5850 [ ] [ ] [ ] [ ]
2 [tombayz] 0 20150[ 0]
3 [nnr13] 0 24750[ 0]
4 B [Meeeee2] 2 11250[ 0]
5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ]
6 [odbphatmackakabig] 1 5750[ 0]
Board: 5s 8d Tc 7h Th
My cards are: [Kd] [6c] [8c] [8s]
Uncalled 2150 returned to deadhead
deadhead Kd 6c 8c 8s
High Full House (Eights over Tens) 8c 8d 8s Th Tc
Low none
JADC 5d 7d 2h 9h
High Two Pair (Tens and Sevens) Tc Th 7h 7d 9h
Low none
deadhead wins 11700 for High
JADC redeemed a lammer for 5000
JADC redeemed a lammer for 5000
Dealing Pot-Limit Binglaha
odbphatmackakabig blinds 2000
JADC blinds 4000
1 JADC 0 6000 4000 [ ] [ ] [ ] [ ]
2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ]
3 nnr13 0 24750 0 [ ] [ ] [ ] [ ]
4 Meeeee2 2 11250 0 [ ] [ ] [ ] [ ]
5 B deadhead 2 53850 0 [ ] [ ] [ ] [ ]
6 odbphatmackakabig 1 3750 2000 [ ] [ ] [ ] [ ]
My cards are: [7s] [7h] [4d] [Ac]
``` | priority | jadc got the big blind twice due to new player motion code odbphatmackakabig got moved from a table with players to our table where we only had there logically was only one place where he could go which was between jadc who had just paid the big blind and me who had just paid the small blind i should have gotten the button which i did odbphatmackakabig should have had to sit out instead he paid the small blind and jadc should have gotten the small blind i think this just shows that my implementation of was flawed i ll look at it when i m awake tomorrow but probably after i ve exercised dealing pot limit binglaha deadhead blinds jadc blinds jadc tombayz b deadhead my cards are deadhead gg tombayz folds gg grizz folds grizz ty folds it is to you deadhead calls jadc checks jadc b deadhead board tc my cards are it is to you deadhead bets jadc asked for and gets more seconds odbphatmackakabig has come from table jadc will be turning in in lammers before the next hand jadc calls and is all in rolled a jadc b deadhead board tc my cards are jadc b deadhead board tc th my cards are uncalled returned to deadhead deadhead kd high full house eights over tens th tc low none jadc high two pair tens and sevens tc th low none deadhead wins for high jadc redeemed a lammer for jadc redeemed a lammer for dealing pot limit binglaha odbphatmackakabig blinds jadc blinds jadc tombayz b deadhead odbphatmackakabig my cards are | 1 |
73,286 | 3,410,538,691 | IssuesEvent | 2015-12-04 20:38:46 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Edit Dataset: Restricting a file on edit files tab results in TOA popup behind the white background. | Component: File Upload & Handling Component: UX & Upgrade Priority: High Status: QA Type: Bug |
The behavior varies, depending on whether you restrict the file on creating the dataset or just editing the file metadata. It works if you restrict from the view dataset mode.
1. On create dataset:
create a dataset, upload a file, restrict that file, save.
The TOA popup is completely behind the whiteout background and you cannot interact with the popup.
2. On edit file metadata:
On save, the TOA popup is half behind the whiteout background with the buttons visible and working and a > next to the cancel button.
| 1.0 | Edit Dataset: Restricting a file on edit files tab results in TOA popup behind the white background. -
The behavior varies, depending on whether you restrict the file on creating the dataset or just editing the file metadata. It works if you restrict from the view dataset mode.
1. On create dataset:
create a dataset, upload a file, restrict that file, save.
The TOA popup is completely behind the whiteout background and you cannot interact with the popup.
2. On edit file metadata:
On save, the TOA popup is half behind the whiteout background with the buttons visible and working and a > next to the cancel button.
| priority | edit dataset restricting a file on edit files tab results in toa popup behind the white background the behavior varies depending on whether you restrict the file on creating the dataset or just editing the file metadata it works if you restrict from the view dataset mode on create dataset create a dataset upload a file restrict that file save the toa popup is completely behind the whiteout background and you cannot interact with the popup on edit file metadata on save the toa popup is half behind the whiteout background with the buttons visible and working and a next to the cancel button | 1 |
33,659 | 2,770,742,634 | IssuesEvent | 2015-05-01 16:44:39 | biocore/qiita | https://api.github.com/repos/biocore/qiita | closed | Clean up instantiation of connection objects | priority: high | Once #1131 is merged, we should go through and clean up instances in the codebase of:
```python
conn_handler = SQLConnectionHandler() if conn_handler is None else conn_handler
```
And the variants. Either the `conn_handler` should be passed in, or a new object created. Since the database connections are not reopened on instantiation (unless necessary), the overhead of just doing `SQLConnectionHandler()` will be minor
cc @antgonza, who originally recommended opening this | 1.0 | Clean up instantiation of connection objects - Once #1131 is merged, we should go through and clean up instances in the codebase of:
```python
conn_handler = SQLConnectionHandler() if conn_handler is None else conn_handler
```
And the variants. Either the `conn_handler` should be passed in, or a new object created. Since the database connections are not reopened on instantiation (unless necessary), the overhead of just doing `SQLConnectionHandler()` will be minor
cc @antgonza, who originally recommended opening this | priority | clean up instantiation of connection objects once is merged we should go through and clean up instances in the codebase of python conn handler sqlconnectionhandler if conn handler is none else conn handler and the variants either the conn handler should be passed in or a new object created since the database connections are not reopened on instantiation unless necessary the overhead of just doing sqlconnectionhandler will be minor cc antgonza who originally recommended opening this | 1 |
261,496 | 8,234,254,772 | IssuesEvent | 2018-09-08 12:15:12 | Theophilix/event-table-edit | https://api.github.com/repos/Theophilix/event-table-edit | closed | Appointment tables: Bug with time / leading zero | bug high priority | When the user chooses a time with "10:00", like 08:00 - 10:00, it is displayed like this: "08:00 - 010:00". Outlook interprets the time like 01:00 (one o'clock!). This problem only appears, when 10:00 is the ending time.



| 1.0 | Appointment tables: Bug with time / leading zero - When the user chooses a time with "10:00", like 08:00 - 10:00, it is displayed like this: "08:00 - 010:00". Outlook interprets the time like 01:00 (one o'clock!). This problem only appears, when 10:00 is the ending time.



| priority | appointment tables bug with time leading zero when the user chooses a time with like it is displayed like this outlook interprets the time like one o clock this problem only appears when is the ending time | 1 |
695,279 | 23,850,793,852 | IssuesEvent | 2022-09-06 17:43:37 | guynir42/virtualobserver | https://api.github.com/repos/guynir42/virtualobserver | opened | Automatic naming for data files and keys | question high priority | When downloading data we need to save it inside files (often with internal keys, like HDF5 files).
The naming should probably be automatic, and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object's data is saved.
It should also conform to reasonable file size and number, e.g., stick 1000 lightcurves in each file. | 1.0 | Automatic naming for data files and keys - When downloading data we need to save it inside files (often with internal keys, like HDF5 files).
The naming should probably be automatic, and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object's data is saved.
It should also conform to reasonable file size and number, e.g., stick 1000 lightcurves in each file. | priority | automatic naming for data files and keys when downloading data we need to save it inside files often with internal keys like files the naming should probably be automatic and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object s data is saved it should also conform to reasonable file size and number e g stick lightcurves in each file | 1 |
185,966 | 6,732,350,377 | IssuesEvent | 2017-10-18 11:08:25 | DOAJ/doaj | https://api.github.com/repos/DOAJ/doaj | opened | DOAJ data download issue | bug high priority | Dear CL
I received the following email and I will need your help:
Hi Team,
We are not able to download DOAJ records. We are getting error while downloading records. PFB error.
Is there any server issue from your side? Please let us know. We are able to download some records but for most of records we are getting this error.
PFA for sample downloaded file.
Error:-
<html>
<head><title>503 Service Temporarily Unavailable</title></head>
<body bgcolor="white">
<center><h1>503 Service Temporarily Unavailable</h1></center>
<hr><center>nginx/1.4.6 (Ubuntu)</center>
</body>
</html>
;503
Steps to download data are as follows:
Step 1
https://doaj.org/oai.article?verb=ListSets
Step 2 FIND medicine
https://doaj.org/oai.article?verb=ListIdentifiers&metadataPrefix=oai_dc&set=RE9BSjpJbnRlcm5hbCBtZWRpY2luZQ~~
Step 3 - resumption token
https://doaj.org/oai.article?verb=ListIdentifiers&resumptionToken=eyJzIjogIlJFOUJTanBKYm5SbGNtNWhiQ0J0WldScFkybHVaUX5-IiwgIm0iOiAib2FpX2RjIiwgIm4iOiA2MDB9
Step 4
https://doaj.org/oai.article?verb=GetRecord&metadataPrefix=oai_dc&identifier=oai:doaj.org/article:fcb85b13cb3a44e6ab1f9dfa2fe605a6
| 1.0 | DOAJ data download issue - Dear CL
I received the following email and I will need your help:
Hi Team,
We are not able to download DOAJ records. We are getting error while downloading records. PFB error.
Is there any server issue from your side? Please let us know. We are able to download some records but for most of records we are getting this error.
PFA for sample downloaded file.
Error:-
<html>
<head><title>503 Service Temporarily Unavailable</title></head>
<body bgcolor="white">
<center><h1>503 Service Temporarily Unavailable</h1></center>
<hr><center>nginx/1.4.6 (Ubuntu)</center>
</body>
</html>
;503
Steps to download data are as follows:
Step 1
https://doaj.org/oai.article?verb=ListSets
Step 2 FIND medicine
https://doaj.org/oai.article?verb=ListIdentifiers&metadataPrefix=oai_dc&set=RE9BSjpJbnRlcm5hbCBtZWRpY2luZQ~~
Step 3 - resumption token
https://doaj.org/oai.article?verb=ListIdentifiers&resumptionToken=eyJzIjogIlJFOUJTanBKYm5SbGNtNWhiQ0J0WldScFkybHVaUX5-IiwgIm0iOiAib2FpX2RjIiwgIm4iOiA2MDB9
Step 4
https://doaj.org/oai.article?verb=GetRecord&metadataPrefix=oai_dc&identifier=oai:doaj.org/article:fcb85b13cb3a44e6ab1f9dfa2fe605a6
| priority | doaj data download issue dear cl i received the following email and i will need your help hi team we are not able to download doaj records we are getting error while downloading records pfb error is there any server issue from your side please let us know we are able to download some records but for most of records we are getting this error pfa for sample downloaded file error service temporarily unavailable service temporarily unavailable nginx ubuntu steps to download data are as follows step step find medicine step resumption token step | 1 |
201,451 | 7,031,613,941 | IssuesEvent | 2017-12-26 19:11:37 | qhacks/hacker-dashboard | https://api.github.com/repos/qhacks/hacker-dashboard | opened | Fix mongo issues (you know what they are ahah) | priority: blocker (high) | **Problem**
Fix the database issues
**Requirements**
- [ ] Resolve data inconsistencies
| 1.0 | Fix mongo issues (you know what they are ahah) - **Problem**
Fix the database issues
**Requirements**
- [ ] Resolve data inconsistencies
| priority | fix mongo issues you know what they are ahah problem fix the database issues requirements resolve data inconsistencies | 1 |
799,331 | 28,304,440,442 | IssuesEvent | 2023-04-10 09:33:10 | bounswe/bounswe2023group6 | https://api.github.com/repos/bounswe/bounswe2023group6 | closed | Filling the Project Plan Chart | priority: high type: task status: inprogress area: wiki area: milestone | ### Problem
Some of group members prepared well structured Project Plan Chart. It must be filled with related information, individually.
### Solution
I'll fill the chart by reviewing all the issues I've contributed to so far.
### Documentation
You can reach the solution [here]().
### Additional notes
_No response_
### Reviewers
Ömer Talip Akalın
### Deadline
10.04.2023 - 20.00 | 1.0 | Filling the Project Plan Chart - ### Problem
Some of group members prepared well structured Project Plan Chart. It must be filled with related information, individually.
### Solution
I'll fill the chart by reviewing all the issues I've contributed to so far.
### Documentation
You can reach the solution [here]().
### Additional notes
_No response_
### Reviewers
Ömer Talip Akalın
### Deadline
10.04.2023 - 20.00 | priority | filling the project plan chart problem some of group members prepared well structured project plan chart it must be filled with related information individually solution i ll fill the chart by reviewing all the issues i ve contributed to so far documentation you can reach the solution additional notes no response reviewers ömer talip akalın deadline | 1 |
365,903 | 10,799,514,235 | IssuesEvent | 2019-11-06 12:21:57 | ADFC-Hamburg/adfc-t30-paten-frontend | https://api.github.com/repos/ADFC-Hamburg/adfc-t30-paten-frontend | closed | Liveversion auf aktuellste Version | high-priority | Bitte die aktuelle Version, (am besten) nach Änderung des Datums zum Stichwort, in die Liveversion einspielen, damit die Kampagne los gehen kann. | 1.0 | Liveversion auf aktuellste Version - Bitte die aktuelle Version, (am besten) nach Änderung des Datums zum Stichwort, in die Liveversion einspielen, damit die Kampagne los gehen kann. | priority | liveversion auf aktuellste version bitte die aktuelle version am besten nach änderung des datums zum stichwort in die liveversion einspielen damit die kampagne los gehen kann | 1 |
402,434 | 11,809,716,916 | IssuesEvent | 2020-03-19 15:22:47 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | opened | PRs to fix patches in *php-buildpack* upstream project | Priority: High Type: Patch | ./api/php-buildpack/jobs/create_bosh_pre_start.sh
./api/php-buildpack/ig_resolver/patch_job_mf.sh | 1.0 | PRs to fix patches in *php-buildpack* upstream project - ./api/php-buildpack/jobs/create_bosh_pre_start.sh
./api/php-buildpack/ig_resolver/patch_job_mf.sh | priority | prs to fix patches in php buildpack upstream project api php buildpack jobs create bosh pre start sh api php buildpack ig resolver patch job mf sh | 1 |
9,934 | 2,608,968,736 | IssuesEvent | 2015-02-26 11:21:09 | GrannyCookies/scratchext2 | https://api.github.com/repos/GrannyCookies/scratchext2 | opened | Extension not loaded when using "Create" to make a new project | bug high priority JavaScript todo | Found by @Felizolinha:
The new loading of the extension does not work when using "Create" to make a new project until the user clicks "See the project page" or reloads it.
A workaround for it would be redefining JSredirectTo() to load ScratchExt when the url ends in "projects/editor".
Please can somebody fix this, probably in the gh-pages branch (scratchext.user.js) | 1.0 | Extension not loaded when using "Create" to make a new project - Found by @Felizolinha:
The new loading of the extension does not work when using "Create" to make a new project until the user clicks "See the project page" or reloads it.
A workaround for it would be redefining JSredirectTo() to load ScratchExt when the url ends in "projects/editor".
Please can somebody fix this, probably in the gh-pages branch (scratchext.user.js) | priority | extension not loaded when using create to make a new project found by felizolinha the new loading of the extension does not work when using create to make a new project until the user clicks see the project page or reloads it a workaround for it would be redefining jsredirectto to load scratchext when the url ends in projects editor please can somebody fix this probably in the gh pages branch scratchext user js | 1 |
499,177 | 14,442,342,523 | IssuesEvent | 2020-12-07 18:00:57 | Edgeryders-Participio/realities | https://api.github.com/repos/Edgeryders-Participio/realities | closed | Create orgs and users that are missing in the mongodb | Priority: High | Right now e.g. I don't exist in the prod plato-core db. Prod doesn't seem to crash though so that's fine for now. But we want orgs and users to exist there so we can properly use orgIds.
My thinking atm (middle of the night):
* Orgs: The single placeholder org we have now can be created in a db migration (built this for neo4j migrations but using it for mongo works fine as well I guess). Additional orgs we'll worry about when we add ui for creating those
* Users: Create them in the same place we're creating them for neo4j (i.e. in auth-callback. make sure to not run it if we don't have an orgId(silent refresh), we want the user to have an org) | 1.0 | Create orgs and users that are missing in the mongodb - Right now e.g. I don't exist in the prod plato-core db. Prod doesn't seem to crash though so that's fine for now. But we want orgs and users to exist there so we can properly use orgIds.
My thinking atm (middle of the night):
* Orgs: The single placeholder org we have now can be created in a db migration (built this for neo4j migrations but using it for mongo works fine as well I guess). Additional orgs we'll worry about when we add ui for creating those
* Users: Create them in the same place we're creating them for neo4j (i.e. in auth-callback. make sure to not run it if we don't have an orgId(silent refresh), we want the user to have an org) | priority | create orgs and users that are missing in the mongodb right now e g i don t exist in the prod plato core db prod doesn t seem to crash though so that s fine for now but we want orgs and users to exist there so we can properly use orgids my thinking atm middle of the night orgs the single placeholder org we have now can be created in a db migration built this for migrations but using it for mongo works fine as well i guess additional orgs we ll worry about when we add ui for creating those users create them in the same place we re creating them for i e in auth callback make sure to not run it if we don t have an orgid silent refresh we want the user to have an org | 1 |
706,647 | 24,280,360,143 | IssuesEvent | 2022-09-28 16:50:56 | AY2223S1-CS2103T-W15-3/tp | https://api.github.com/repos/AY2223S1-CS2103T-W15-3/tp | closed | Update Developer Guide (Use Cases) | priority.High type.Task | # TODO
Give use cases (textual form) for a few representative user stories that need multiple steps to complete.
_e.g. Adding a tag to a person (assume the user needs to find the person first)_
See sample [here](https://se-education.org/addressbook-level3/DeveloperGuide.html#use-cases).
## Things to note (and/or avoid)
- Important use cases missing (a use case is important if it involves a user interaction that is worthy of documenting e.g., it has multiple extensions -- this is not the same as the feature being important)
- Formatting/notational errors
- Incorrect step numbering
- Unnecessary UI details mentioned
- Missing/unnecessary steps
- Missing extensions | 1.0 | Update Developer Guide (Use Cases) - # TODO
Give use cases (textual form) for a few representative user stories that need multiple steps to complete.
_e.g. Adding a tag to a person (assume the user needs to find the person first)_
See sample [here](https://se-education.org/addressbook-level3/DeveloperGuide.html#use-cases).
## Things to note (and/or avoid)
- Important use cases missing (a use case is important if it involves a user interaction that is worthy of documenting e.g., it has multiple extensions -- this is not the same as the feature being important)
- Formatting/notational errors
- Incorrect step numbering
- Unnecessary UI details mentioned
- Missing/unnecessary steps
- Missing extensions | priority | update developer guide use cases todo give use cases textual form for a few representative user stories that need multiple steps to complete e g adding a tag to a person assume the user needs to find the person first see sample things to note and or avoid important use cases missing a use case is important if it involves a user interaction that is worthy of documenting e g it has multiple extensions this is not the same as the feature being important formatting notational errors incorrect step numbering unnecessary ui details mentioned missing unnecessary steps missing extensions | 1 |
798,545 | 28,289,461,848 | IssuesEvent | 2023-04-09 02:23:43 | KATO-Hiro/AtCoderClans | https://api.github.com/repos/KATO-Hiro/AtCoderClans | closed | 一部のページのタイトルが飛んでいるかも? | invalid priority high | ## WHAT
- サイトジェネレータをv9に変えたあたりから?
- 一部のページのみ該当している
- 本家の#4824が関連しているか?
## TODO
- [ ] 該当するページを列挙
- [ ] ヘッダを追加? | 1.0 | 一部のページのタイトルが飛んでいるかも? - ## WHAT
- サイトジェネレータをv9に変えたあたりから?
- 一部のページのみ該当している
- 本家の#4824が関連しているか?
## TODO
- [ ] 該当するページを列挙
- [ ] ヘッダを追加? | priority | 一部のページのタイトルが飛んでいるかも what 一部のページのみ該当している 本家の todo 該当するページを列挙 ヘッダを追加 | 1 |
659,689 | 21,937,739,193 | IssuesEvent | 2022-05-23 15:06:18 | ooni/ooni.org | https://api.github.com/repos/ooni/ooni.org | opened | Prepare for Internet without Borders events | priority/high workshop community | OONI is participating in the Internet without Borders events (https://internetborders.net/) for the Russian diaspora. As part of these events, we'll be giving OONI presentations and facilitating OONI hackathons.
In preparation for these events, we're working on the following:
- [ ] Create (and share) slides for OONI presentations
- [ ] Finalize challenges for OONI hackathons
- [ ] Create video for OONI hackathons
- [ ] Create OONI channel on the event Mattermost (for hackathon coordination)
- [ ] Update relevant spreadsheet (coordination with organizers) | 1.0 | Prepare for Internet without Borders events - OONI is participating in the Internet without Borders events (https://internetborders.net/) for the Russian diaspora. As part of these events, we'll be giving OONI presentations and facilitating OONI hackathons.
In preparation for these events, we're working on the following:
- [ ] Create (and share) slides for OONI presentations
- [ ] Finalize challenges for OONI hackathons
- [ ] Create video for OONI hackathons
- [ ] Create OONI channel on the event Mattermost (for hackathon coordination)
- [ ] Update relevant spreadsheet (coordination with organizers) | priority | prepare for internet without borders events ooni is participating in the internet without borders events for the russian diaspora as part of these events we ll be giving ooni presentations and facilitating ooni hackathons in preparation for these events we re working on the following create and share slides for ooni presentations finalize challenges for ooni hackathons create video for ooni hackathons create ooni channel on the event mattermost for hackathon coordination update relevant spreadsheet coordination with organizers | 1 |
74,049 | 3,427,685,633 | IssuesEvent | 2015-12-10 03:34:59 | phetsims/website | https://api.github.com/repos/phetsims/website | closed | Change directions for comments on activity pages | High Priority Teacher Resources pages | Currently, on any activity page, users can add a comment. The text next to the entry box says:
"What do you think about this activity? How did you use it or change it for your class? Professionally constructive comments welcome."
We would like to change it to say:
"How did you use this activity or change it for your class? Answer keys are often not available because the activities are inquiry based and lead to multiple answers. Professionally constructive comments welcome." | 1.0 | Change directions for comments on activity pages - Currently, on any activity page, users can add a comment. The text next to the entry box says:
"What do you think about this activity? How did you use it or change it for your class? Professionally constructive comments welcome."
We would like to change it to say:
"How did you use this activity or change it for your class? Answer keys are often not available because the activities are inquiry based and lead to multiple answers. Professionally constructive comments welcome." | priority | change directions for comments on activity pages currently on any activity page users can add a comment the text next to the entry box says what do you think about this activity how did you use it or change it for your class professionally constructive comments welcome we would like to change it to say how did you use this activity or change it for your class answer keys are often not available because the activities are inquiry based and lead to multiple answers professionally constructive comments welcome | 1 |
541,653 | 15,830,997,789 | IssuesEvent | 2021-04-06 13:10:09 | bryntum/support | https://api.github.com/repos/bryntum/support | opened | Popups misaligned | bug high-priority | Reproducible in blog inside a web component
<img width="906" alt="Screenshot 2021-04-06 at 15 09 17" src="https://user-images.githubusercontent.com/218570/113715687-15f7b100-96ea-11eb-8f9f-d45c63d95626.png">
| 1.0 | Popups misaligned - Reproducible in blog inside a web component
<img width="906" alt="Screenshot 2021-04-06 at 15 09 17" src="https://user-images.githubusercontent.com/218570/113715687-15f7b100-96ea-11eb-8f9f-d45c63d95626.png">
| priority | popups misaligned reproducible in blog inside a web component img width alt screenshot at src | 1 |
428,183 | 12,404,316,628 | IssuesEvent | 2020-05-21 15:20:17 | nthopinion/covid19 | https://api.github.com/repos/nthopinion/covid19 | closed | Physician Registration: read-only version of physician page for physicians with "level 0" access | High Priority Trello [Sync] | * they can rate answers
* they can't answer questions (disabled fields)
Context: https://docs.google.com/document/d/1y-26s8otakzhvMqKYpj_3xNFNdGsa1REpwlt4sNLL2I/edit
Related tasks in Trello attachments.
┆Issue is synchronized with this [Trello card](https://trello.com/c/PYZEjPin) by [Unito](https://www.unito.io/learn-more)
┆Attachments: <a href="https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access">https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access</a> | <a href="https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information">https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information</a>
| 1.0 | Physician Registration: read-only version of physician page for physicians with "level 0" access - * they can rate answers
* they can't answer questions (disabled fields)
Context: https://docs.google.com/document/d/1y-26s8otakzhvMqKYpj_3xNFNdGsa1REpwlt4sNLL2I/edit
Related tasks in Trello attachments.
┆Issue is synchronized with this [Trello card](https://trello.com/c/PYZEjPin) by [Unito](https://www.unito.io/learn-more)
┆Attachments: <a href="https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access">https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access</a> | <a href="https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information">https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information</a>
| priority | physician registration read only version of physician page for physicians with level access they can rate answers they can t answer questions disabled fields context related tasks in trello attachments ┆issue is synchronized with this by ┆attachments a href a href | 1 |
364,270 | 10,761,538,483 | IssuesEvent | 2019-10-31 21:00:54 | semperfiwebdesign/all-in-one-seo-pack | https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack | closed | Fatal error: Uncaught Error in PHP 7.1 | Priority | High | Reported here: https://wordpress.org/support/topic/php-7-1-problem/
I'm not exactly sure what's prompting this, but I don't think this has to do with PHP 7.1. My guess is that the edit array key simply doesn't exist already for some reason.
It may be as simple as checking for it first.
unset( $actions['edit'] );
| 1.0 | Fatal error: Uncaught Error in PHP 7.1 - Reported here: https://wordpress.org/support/topic/php-7-1-problem/
I'm not exactly sure what's prompting this, but I don't think this has to do with PHP 7.1. My guess is that the edit array key simply doesn't exist already for some reason.
It may be as simple as checking for it first.
unset( $actions['edit'] );
| priority | fatal error uncaught error in php reported here i m not exactly sure what s prompting this but i don t think this has to do with php my guess is that the edit array key simply doesn t exist already for some reason it may be as simple as checking for it first unset actions | 1 |
725,881 | 24,979,279,007 | IssuesEvent | 2022-11-02 10:23:48 | AY2223S1-CS2113-W13-1/tp | https://api.github.com/repos/AY2223S1-CS2113-W13-1/tp | closed | Be able to cancel login | type.Bug priority.High | Once you are in the process of logging in but do not want to log in you should be able to exit | 1.0 | Be able to cancel login - Once you are in the process of logging in but do not want to log in you should be able to exit | priority | be able to cancel login once you are in the process of logging in but do not want to log in you should be able to exit | 1 |
684,072 | 23,405,996,569 | IssuesEvent | 2022-08-12 12:54:15 | cassproject/cass-editor | https://api.github.com/repos/cassproject/cass-editor | closed | Missing a few competencies for DOL Frameworks and not able to export them with this error "TypeError: Cannot read property 'ceasn:isChildOf' of undefined" | blocker High Priority Credential Engine CaSS | @gloverkari There are the two competencies i am not able to export and seeing the same error
DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139
CASS-- https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139
DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
CASS-https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
| 1.0 | Missing a few competencies for DOL Frameworks and not able to export them with this error "TypeError: Cannot read property 'ceasn:isChildOf' of undefined" - @gloverkari There are the two competencies i am not able to export and seeing the same error
DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139
CASS-- https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139
DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
CASS-https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
| priority | missing a few competencies for dol frameworks and not able to export them with this error typeerror cannot read property ceasn ischildof of undefined gloverkari there are the two competencies i am not able to export and seeing the same error dol cass dol cass | 1 |
716,026 | 24,618,672,844 | IssuesEvent | 2022-10-15 16:40:51 | onesoft-sudo/sudobot | https://api.github.com/repos/onesoft-sudo/sudobot | closed | Profile filter interferes with muting users | bug help wanted priority:high important status:ready-for-fixing semver:minor | After a mute by profile filter, if a user gets muted by a moderator, they will be unmuted back after profile filter takeback. Add a check if the moderators has run mute command. | 1.0 | Profile filter interferes with muting users - After a mute by profile filter, if a user gets muted by a moderator, they will be unmuted back after profile filter takeback. Add a check if the moderators has run mute command. | priority | profile filter interferes with muting users after a mute by profile filter if a user gets muted by a moderator they will be unmuted back after profile filter takeback add a check if the moderators has run mute command | 1 |
555,705 | 16,463,730,436 | IssuesEvent | 2021-05-22 01:50:52 | ncopenpass/CampaignFinance | https://api.github.com/repos/ncopenpass/CampaignFinance | opened | design: create design for amount filter on tables | Class Prototype High Priority ui | ### Task
Update the figma prototype to include the amount filter. | 1.0 | design: create design for amount filter on tables - ### Task
Update the figma prototype to include the amount filter. | priority | design create design for amount filter on tables task update the figma prototype to include the amount filter | 1 |
263,626 | 8,299,271,563 | IssuesEvent | 2018-09-21 01:53:55 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] documentation button under help in context nav goes to invalid documentation link | bug priority: high | ### Expected behavior
Documentation site should load
### Actual behavior
Button loads URL: https://docs.craftercms.org/en/3.0/
URL is missing index.html
User gets S3's 404
### Steps to reproduce the problem
* log in to Studio
* click help
* click documentation
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
Studio Version Number: 3.0.17-SNAPSHOT-5e8acf
Build Number: 5e8acf27a697e7a6a52f4156a5c390a3538c6687
Build Date/Time: 09-19-2018 16:41:26 -0400
| 1.0 | [studio-ui] documentation button under help in context nav goes to invalid documentation link - ### Expected behavior
Documentation site should load
### Actual behavior
Button loads URL: https://docs.craftercms.org/en/3.0/
URL is missing index.html
User gets S3's 404
### Steps to reproduce the problem
* log in to Studio
* click help
* click documentation
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
Studio Version Number: 3.0.17-SNAPSHOT-5e8acf
Build Number: 5e8acf27a697e7a6a52f4156a5c390a3538c6687
Build Date/Time: 09-19-2018 16:41:26 -0400
| priority | documentation button under help in context nav goes to invalid documentation link expected behavior documentation site should load actual behavior button loads url url is missing index html user gets s steps to reproduce the problem log in to studio click help click documentation log stack trace use n a specs version studio version number snapshot build number build date time | 1 |
196,845 | 6,949,968,918 | IssuesEvent | 2017-12-06 09:05:37 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | remoteshell is very slow | component:coral priority:high sprint2 status:pending type:usability | `remoteshell` has a LOT of long-ish sleeps in it.
```
[root@mgmt1 ~]# time updatenode h41n01 -P remoteshell
h41n01: xcatdsklspost: downloaded postscripts successfully
h41n01: Fri Dec 1 01:06:17 EST 2017 Running postscript: remoteshell
h41n01:
h41n01: postscript: remoteshell exited with code 0
h41n01: Running of postscripts has completed.
real 2m17.663s
user 0m0.566s
sys 0m0.029s
```
I tried to enable xCAT flow support, but that actually took longer.
```
[root@mgmt1 ~]# time updatenode h41n01 -P remoteshell
h41n01: xcatdsklspost: downloaded postscripts successfully
h41n01: Fri Dec 1 01:10:07 EST 2017 Running postscript: remoteshell
h41n01:
h41n01: postscript: remoteshell exited with code 0
h41n01: Running of postscripts has completed.
real 3m0.601s
user 0m0.588s
sys 0m0.010s
```
I hopped on the node and saw 14 concurrent `xcatflowrequest` commands that were reparented to init. I didn't tune `xcatmaxconnections` or `xcatmaxbatchconnections`, but I only ran this on a single node, so the defaults should be ok.
It would be great if `remoteshell` could get all the keys in a single command instead of one sleep+RTT for each key. | 1.0 | remoteshell is very slow - `remoteshell` has a LOT of long-ish sleeps in it.
```
[root@mgmt1 ~]# time updatenode h41n01 -P remoteshell
h41n01: xcatdsklspost: downloaded postscripts successfully
h41n01: Fri Dec 1 01:06:17 EST 2017 Running postscript: remoteshell
h41n01:
h41n01: postscript: remoteshell exited with code 0
h41n01: Running of postscripts has completed.
real 2m17.663s
user 0m0.566s
sys 0m0.029s
```
I tried to enable xCAT flow support, but that actually took longer.
```
[root@mgmt1 ~]# time updatenode h41n01 -P remoteshell
h41n01: xcatdsklspost: downloaded postscripts successfully
h41n01: Fri Dec 1 01:10:07 EST 2017 Running postscript: remoteshell
h41n01:
h41n01: postscript: remoteshell exited with code 0
h41n01: Running of postscripts has completed.
real 3m0.601s
user 0m0.588s
sys 0m0.010s
```
I hopped on the node and saw 14 concurrent `xcatflowrequest` commands that were reparented to init. I didn't tune `xcatmaxconnections` or `xcatmaxbatchconnections`, but I only ran this on a single node, so the defaults should be ok.
It would be great if `remoteshell` could get all the keys in a single command instead of one sleep+RTT for each key. | priority | remoteshell is very slow remoteshell has a lot of long ish sleeps in it time updatenode p remoteshell xcatdsklspost downloaded postscripts successfully fri dec est running postscript remoteshell postscript remoteshell exited with code running of postscripts has completed real user sys i tried to enable xcat flow support but that actually took longer time updatenode p remoteshell xcatdsklspost downloaded postscripts successfully fri dec est running postscript remoteshell postscript remoteshell exited with code running of postscripts has completed real user sys i hopped on the node and saw concurrent xcatflowrequest commands that were reparented to init i didn t tune xcatmaxconnections or xcatmaxbatchconnections but i only ran this on a single node so the defaults should be ok it would be great if remoteshell could get all the keys in a single command instead of one sleep rtt for each key | 1 |
313,126 | 9,557,487,695 | IssuesEvent | 2019-05-03 11:43:19 | fritzing/fritzing-app | https://api.github.com/repos/fritzing/fritzing-app | closed | disappearing parts--when changing z order of a wire in bb view | Priority-High bug imported | _From [irasc...@gmail.com](https://code.google.com/u/104729248032245122687/) on June 11, 2013 12:52:04_
Attached sketch no longer exhibits the behavior. Something is broken in the program state, but saving and loading cleans up the problem. When the striped wire was raised in z-order, the resistor disappeared in bb view. The part was still visible in the other views.
**Attachment:** [DC_motor_control---disapearing parts.fzz](http://code.google.com/p/fritzing/issues/detail?id=2623)
_Original issue: http://code.google.com/p/fritzing/issues/detail?id=2623_
| 1.0 | disappearing parts--when changing z order of a wire in bb view - _From [irasc...@gmail.com](https://code.google.com/u/104729248032245122687/) on June 11, 2013 12:52:04_
Attached sketch no longer exhibits the behavior. Something is broken in the program state, but saving and loading cleans up the problem. When the striped wire was raised in z-order, the resistor disappeared in bb view. The part was still visible in the other views.
**Attachment:** [DC_motor_control---disapearing parts.fzz](http://code.google.com/p/fritzing/issues/detail?id=2623)
_Original issue: http://code.google.com/p/fritzing/issues/detail?id=2623_
| priority | disappearing parts when changing z order of a wire in bb view from on june attached sketch no longer exhibits the behavior something is broken in the program state but saving and loading cleans up the problem when the striped wire was raised in z order the resistor disappeared in bb view the part was still visible in the other views attachment original issue | 1 |
660,878 | 22,034,515,033 | IssuesEvent | 2022-05-28 10:58:06 | naturalcrit/homebrewery | https://api.github.com/repos/naturalcrit/homebrewery | closed | Columns not displaying correctly on Chrome 102.0.5005.63 | bug solution found P1 - high priority | ## Environment
**Browser(s):** Chrome Version 102.0.5005.63 (Official Build) (64-bit)
**Operating System:** Windows 10 (x64)
**Legacy or v3 Renderer:** v3 Renderer
## Description
Manual column breaks are rendered with too much spacing and are pushed off the page on Chrome 102.0.5005.63 but not on earlier versions.
## Images
[Chrome 101.0.4951.67](https://i.imgur.com/gKY8d44.png)
[Chrome 102.0.5005.63](https://i.imgur.com/q21LTWZ.png)
**Share Link** :
https://homebrewery.naturalcrit.com/share/19O72H7NrXlGJ17rXmBXZ13QkDBSNGyw0fd5PofH4aTmO | 1.0 | Columns not displaying correctly on Chrome 102.0.5005.63 - ## Environment
**Browser(s):** Chrome Version 102.0.5005.63 (Official Build) (64-bit)
**Operating System:** Windows 10 (x64)
**Legacy or v3 Renderer:** v3 Renderer
## Description
Manual column breaks are rendered with too much spacing and are pushed off the page on Chrome 102.0.5005.63 but not on earlier versions.
## Images
[Chrome 101.0.4951.67](https://i.imgur.com/gKY8d44.png)
[Chrome 102.0.5005.63](https://i.imgur.com/q21LTWZ.png)
**Share Link** :
https://homebrewery.naturalcrit.com/share/19O72H7NrXlGJ17rXmBXZ13QkDBSNGyw0fd5PofH4aTmO | priority | columns not displaying correctly on chrome environment browser s chrome version official build bit operating system windows legacy or renderer renderer description manual column breaks are rendered with too much spacing and are pushed off the page on chrome but not on earlier versions images share link | 1 |
636,913 | 20,612,987,606 | IssuesEvent | 2022-03-07 10:27:23 | Soulcialize/souldragonknight | https://api.github.com/repos/Soulcialize/souldragonknight | opened | Implement AI for Alpha stage enemies | type.Enhancement priority.High | There are four types of enemies planned for Alpha stage. Two of them will target the Knight, while the other two will target the Dragon.
**Dragon:**
- [ ] Flying enemy, floats towards the Dragon before charging up a melee attack
- [ ] Ground enemy, protected/shielded such that the Dragon is unable to damage it. The Dragon must empower the knight's attack so that the knight can destroy this enemy. This enemy will fire projectiles at the Dragon.
**Knight:**
- [ ] Ground enemy, runs towards the Knight before charging up a melee attack
- [ ] Flying enemy, fire projectiles at the Knight. The enemy will be flying out of the Knight's reach. The Knight must ride the Dragon so that the Knight can attack and destroy this enemy. | 1.0 | Implement AI for Alpha stage enemies - There are four types of enemies planned for Alpha stage. Two of them will target the Knight, while the other two will target the Dragon.
**Dragon:**
- [ ] Flying enemy, floats towards the Dragon before charging up a melee attack
- [ ] Ground enemy, protected/shielded such that the Dragon is unable to damage it. The Dragon must empower the knight's attack so that the knight can destroy this enemy. This enemy will fire projectiles at the Dragon.
**Knight:**
- [ ] Ground enemy, runs towards the Knight before charging up a melee attack
- [ ] Flying enemy, fire projectiles at the Knight. The enemy will be flying out of the Knight's reach. The Knight must ride the Dragon so that the Knight can attack and destroy this enemy. | priority | implement ai for alpha stage enemies there are four types of enemies planned for alpha stage two of them will target the knight while the other two will target the dragon dragon flying enemy floats towards the dragon before charging up a melee attack ground enemy protected shielded such that the dragon is unable to damage it the dragon must empower the knight s attack so that the knight can destroy this enemy this enemy will fire projectiles at the dragon knight ground enemy runs towards the knight before charging up a melee attack flying enemy fire projectiles at the knight the enemy will be flying out of the knight s reach the knight must ride the dragon so that the knight can attack and destroy this enemy | 1 |
390,271 | 11,541,155,897 | IssuesEvent | 2020-02-18 02:56:37 | ritsec/cluster-duck | https://api.github.com/repos/ritsec/cluster-duck | closed | Change Wan network IP range | competition configuration high-priority management | Change Wan network IP range
==============
Change Wan network IP range to the upper 2/3rds of our current /24 network
| 1.0 | Change Wan network IP range - Change Wan network IP range
==============
Change Wan network IP range to the upper 2/3rds of our current /24 network
| priority | change wan network ip range change wan network ip range change wan network ip range to the upper of our current network | 1 |
117,903 | 4,728,890,290 | IssuesEvent | 2016-10-18 17:10:00 | MRN-Code/penny-collector | https://api.github.com/repos/MRN-Code/penny-collector | closed | Rework file transfer | enhancement high priority | File in question: https://github.com/MRN-Code/penny-collector/blob/develop/src/transferFile.js.
It’s pretty difficult to reason about. Turns out, the [Node.js tus client supports streams](https://github.com/tus/tus-js-client/blob/master/demo/node.js). We should be able to do something like:
```js
const promisify = require('bluebird').promisify;
const tar = require('tar-fs');
const tus = require('tus');
module.exports = transferFile(filePath) {
const file = tar.pack(filePath);
return promisify(fs.stat)(file).then(stat => {
return new Promise((resolve, reject) => {
const upload = new tus.Client(file, {
endpoint: '', // Get form configuration
metadata: {
file: filePath,
},
onError: reject,
onSuccess: resolve,
resume: true,
uploadSize: stat.size,
});
upload.start();
});
});
};
``` | 1.0 | Rework file transfer - File in question: https://github.com/MRN-Code/penny-collector/blob/develop/src/transferFile.js.
It’s pretty difficult to reason about. Turns out, the [Node.js tus client supports streams](https://github.com/tus/tus-js-client/blob/master/demo/node.js). We should be able to do something like:
```js
const promisify = require('bluebird').promisify;
const tar = require('tar-fs');
const tus = require('tus');
module.exports = transferFile(filePath) {
const file = tar.pack(filePath);
return promisify(fs.stat)(file).then(stat => {
return new Promise((resolve, reject) => {
const upload = new tus.Client(file, {
endpoint: '', // Get form configuration
metadata: {
file: filePath,
},
onError: reject,
onSuccess: resolve,
resume: true,
uploadSize: stat.size,
});
upload.start();
});
});
};
``` | priority | rework file transfer file in question it’s pretty difficult to reason about turns out the we should be able to do something like js const promisify require bluebird promisify const tar require tar fs const tus require tus module exports transferfile filepath const file tar pack filepath return promisify fs stat file then stat return new promise resolve reject const upload new tus client file endpoint get form configuration metadata file filepath onerror reject onsuccess resolve resume true uploadsize stat size upload start | 1 |
180,922 | 6,654,438,823 | IssuesEvent | 2017-09-29 12:50:28 | CoderDojo/community-platform | https://api.github.com/repos/CoderDojo/community-platform | closed | Minify all static files | high priority technical | We are including separately all of the javascript and css files we need.
We could make the site a lot faster and it would help our SEO to minify and combine the bower files.
e.g. http://stackoverflow.com/questions/25322326/combine-and-minify-all-bower-libraries-with-gruntjs
| 1.0 | Minify all static files - We are including separately all of the javascript and css files we need.
We could make the site a lot faster and it would help our SEO to minify and combine the bower files.
e.g. http://stackoverflow.com/questions/25322326/combine-and-minify-all-bower-libraries-with-gruntjs
| priority | minify all static files we are including separately all of the javascript and css files we need we could make the site a lot faster and it would help our seo to minify and combine the bower files e g | 1 |
796,999 | 28,134,890,142 | IssuesEvent | 2023-04-01 08:58:36 | telstra/open-kilda | https://api.github.com/repos/telstra/open-kilda | opened | Add ability to validate HA flow | priority/2-high feature | Ha flow validation must use rile manager to build expected rules
Part of #5061 | 1.0 | Add ability to validate HA flow - Ha flow validation must use rile manager to build expected rules
Part of #5061 | priority | add ability to validate ha flow ha flow validation must use rile manager to build expected rules part of | 1 |
160,790 | 6,102,586,658 | IssuesEvent | 2017-06-20 16:48:53 | crowdAI/crowdai | https://api.github.com/repos/crowdAI/crowdai | closed | CrowdAI logo on mobile different | high priority v2 | For some reason, the crowdAI samurAI loses his eyes on mobile...
<img width="402" alt="screen shot 2017-06-17 at 1 56 20 pm" src="https://user-images.githubusercontent.com/215057/27252691-e14b90e2-5364-11e7-94b8-61f0a3bd6339.png">
| 1.0 | CrowdAI logo on mobile different - For some reason, the crowdAI samurAI loses his eyes on mobile...
<img width="402" alt="screen shot 2017-06-17 at 1 56 20 pm" src="https://user-images.githubusercontent.com/215057/27252691-e14b90e2-5364-11e7-94b8-61f0a3bd6339.png">
| priority | crowdai logo on mobile different for some reason the crowdai samurai loses his eyes on mobile img width alt screen shot at pm src | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.