Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
338,627
10,232,453,098
IssuesEvent
2019-08-18 17:37:04
futurismo-org/titan
https://api.github.com/repos/futurismo-org/titan
closed
りんごリジェクト対応(5回目)
high priority
ちゃんと見ていないので、解説する。 1. 1 Safety: Objectionable Content 1. 2 Safety: User Generated Content
1.0
りんごリジェクト対応(5回目) - ちゃんと見ていないので、解説する。 1. 1 Safety: Objectionable Content 1. 2 Safety: User Generated Content
priority
りんごリジェクト対応 ちゃんと見ていないので、解説する。 safety objectionable content safety user generated content
1
43,879
2,893,715,337
IssuesEvent
2015-06-15 19:26:02
SCIInstitute/shapeworks
https://api.github.com/repos/SCIInstitute/shapeworks
reopened
Final Release
High Priority IBBM
No more features to add at this point. Only bugs that are important for IBBM allowed after this.
1.0
Final Release - No more features to add at this point. Only bugs that are important for IBBM allowed after this.
priority
final release no more features to add at this point only bugs that are important for ibbm allowed after this
1
186,230
6,734,519,625
IssuesEvent
2017-10-18 18:20:16
resin-io/resin-cli
https://api.github.com/repos/resin-io/resin-cli
closed
resin device register only does the old long UUIDs
priority:high type:bug
With resinOS 2.0, the device UUIDs became half the length (~31~ 32 chars down from 62 characters). Running `resin device register AppName` only able to do the original 62 character version. Not sure if it's a problem, just highlight it, because this way images freshly downloaded from the Dashboard and images configured with the CLI will have different behaviour for 2.0. What would be the right default behaviour? Or this just doesn't have any real consequences? <img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_20dt)
1.0
resin device register only does the old long UUIDs - With resinOS 2.0, the device UUIDs became half the length (~31~ 32 chars down from 62 characters). Running `resin device register AppName` only able to do the original 62 character version. Not sure if it's a problem, just highlight it, because this way images freshly downloaded from the Dashboard and images configured with the CLI will have different behaviour for 2.0. What would be the right default behaviour? Or this just doesn't have any real consequences? <img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_20dt)
priority
resin device register only does the old long uuids with resinos the device uuids became half the length chars down from characters running resin device register appname only able to do the original character version not sure if it s a problem just highlight it because this way images freshly downloaded from the dashboard and images configured with the cli will have different behaviour for what would be the right default behaviour or this just doesn t have any real consequences
1
3,912
2,542,061,046
IssuesEvent
2015-01-28 14:06:33
bethlakshmi/GBE2
https://api.github.com/repos/bethlakshmi/GBE2
closed
Update Your Bio
High Priority
Your bio, which is now live on the site, reads: Betty has been doing this too long. She also ported the data. You might want to change that...
1.0
Update Your Bio - Your bio, which is now live on the site, reads: Betty has been doing this too long. She also ported the data. You might want to change that...
priority
update your bio your bio which is now live on the site reads betty has been doing this too long she also ported the data you might want to change that
1
628,072
19,974,919,871
IssuesEvent
2022-01-29 00:52:10
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
Option to set locale globally
Difficulty: [2] Intermediate Effort: [3] High Priority: ♨︎ Critical Type: ★ Enhancement
### Setting locale for fmt_\* globally Im using gt increasingly in html reports in Europe, therefore I would like to set the locale globally, for example via an option that switches every standard gt fmt_\* to use the locale specified in the option. That would make (working with) gt code in non-US locales less verbose and reduces the need to use purrr::partial
1.0
Option to set locale globally - ### Setting locale for fmt_\* globally Im using gt increasingly in html reports in Europe, therefore I would like to set the locale globally, for example via an option that switches every standard gt fmt_\* to use the locale specified in the option. That would make (working with) gt code in non-US locales less verbose and reduces the need to use purrr::partial
priority
option to set locale globally setting locale for fmt globally im using gt increasingly in html reports in europe therefore i would like to set the locale globally for example via an option that switches every standard gt fmt to use the locale specified in the option that would make working with gt code in non us locales less verbose and reduces the need to use purrr partial
1
528,599
15,370,530,770
IssuesEvent
2021-03-02 08:55:35
Mobsya/aseba
https://api.github.com/repos/Mobsya/aseba
opened
ThymioSuite on Big Sur do not seen robots, problem of Discovery
Mac OS specific Thymio Device Manager bug high priority
Some user reported that on Big Sur (some on 10.15.7), robot are not seen. Lot of investigation was made with users and apple around a correct packaging and Notarization. That's resolved the problem for some user because TDM was blocked by gatekeeper. Finally bug remains with some Big Sur user where the package was not the problem. Finally we discovered that TDM cannot show itself with the Discovery process (Bonjour) a error came back from the library with "bad parameters". [https://github.com/Mobsya/aseba/blob/6a557425761f494f2536ba725889acad54840215/aseba/thymio-device-manager/aseba_node_registery.cpp#L138](url). This behaviour has been seen on two user computer but cannot be reproduced.
1.0
ThymioSuite on Big Sur do not seen robots, problem of Discovery - Some user reported that on Big Sur (some on 10.15.7), robot are not seen. Lot of investigation was made with users and apple around a correct packaging and Notarization. That's resolved the problem for some user because TDM was blocked by gatekeeper. Finally bug remains with some Big Sur user where the package was not the problem. Finally we discovered that TDM cannot show itself with the Discovery process (Bonjour) a error came back from the library with "bad parameters". [https://github.com/Mobsya/aseba/blob/6a557425761f494f2536ba725889acad54840215/aseba/thymio-device-manager/aseba_node_registery.cpp#L138](url). This behaviour has been seen on two user computer but cannot be reproduced.
priority
thymiosuite on big sur do not seen robots problem of discovery some user reported that on big sur some on robot are not seen lot of investigation was made with users and apple around a correct packaging and notarization that s resolved the problem for some user because tdm was blocked by gatekeeper finally bug remains with some big sur user where the package was not the problem finally we discovered that tdm cannot show itself with the discovery process bonjour a error came back from the library with bad parameters url this behaviour has been seen on two user computer but cannot be reproduced
1
533,233
15,586,878,098
IssuesEvent
2021-03-18 02:59:41
AY2021S2-CS2103T-T12-3/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-3/tp
opened
Alias command is not found in autocomplete panel
priority.High
A possible reason is that **getAutoCompleteCommands()** is not pulling that command yet. **Alias command as found in UG.** ![image](https://user-images.githubusercontent.com/77429783/111566839-dc154800-87d8-11eb-9996-88388bfdc716.png) **Alias command not found.** ![image](https://user-images.githubusercontent.com/77429783/111566869-e9323700-87d8-11eb-89e9-7e1c76f7f817.png)
1.0
Alias command is not found in autocomplete panel - A possible reason is that **getAutoCompleteCommands()** is not pulling that command yet. **Alias command as found in UG.** ![image](https://user-images.githubusercontent.com/77429783/111566839-dc154800-87d8-11eb-9996-88388bfdc716.png) **Alias command not found.** ![image](https://user-images.githubusercontent.com/77429783/111566869-e9323700-87d8-11eb-89e9-7e1c76f7f817.png)
priority
alias command is not found in autocomplete panel a possible reason is that getautocompletecommands is not pulling that command yet alias command as found in ug alias command not found
1
478,479
13,780,041,736
IssuesEvent
2020-10-08 14:28:05
carbon-design-system/ibm-dotcom-library
https://api.github.com/repos/carbon-design-system/ibm-dotcom-library
closed
Web Component: Develop Table of Contents of the React version - Group 2
Airtable Done dev package: web components priority: high
#### User Story <!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} --> > As a `[user role below]`: IBM.com Library developer > I need to: create the `Table of Contents` > so that I can: provide ibm.com adopter developers a web component version for every react version available in the ibm.com Library #### Additional information <!-- {{Please provide any additional information or resources for reference}} --> - Story within Storybook with corresponding knobs - Utilize Carbon - Create with Shadow DOM and Custom Elements standards - **See the Epic for the Design and Functional specs information** - [React canary environment](https://ibmdotcom-react-canary.mybluemix.net/?path=/docs/overview-getting-started--page) - Prod QA testing issue (#3631) #### Acceptance criteria - [ ] Include README for the web component and corresponding styles - [ ] Create Web Components styles in styles package - [ ] No custom styles in web-components package - [ ] Do not create knobs in Storybook that include JSON objects - [ ] Break out Storybook stories into multiple variation stories, if applicable - [ ] Create codesandbox example under `/packages/web-components/examples/codesandbox` and include in README - [ ] Minimum 80% unit test coverage - [ ] A comment is posted in the Prod QA issue, tagging Praveen when development is finished
1.0
Web Component: Develop Table of Contents of the React version - Group 2 - #### User Story <!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} --> > As a `[user role below]`: IBM.com Library developer > I need to: create the `Table of Contents` > so that I can: provide ibm.com adopter developers a web component version for every react version available in the ibm.com Library #### Additional information <!-- {{Please provide any additional information or resources for reference}} --> - Story within Storybook with corresponding knobs - Utilize Carbon - Create with Shadow DOM and Custom Elements standards - **See the Epic for the Design and Functional specs information** - [React canary environment](https://ibmdotcom-react-canary.mybluemix.net/?path=/docs/overview-getting-started--page) - Prod QA testing issue (#3631) #### Acceptance criteria - [ ] Include README for the web component and corresponding styles - [ ] Create Web Components styles in styles package - [ ] No custom styles in web-components package - [ ] Do not create knobs in Storybook that include JSON objects - [ ] Break out Storybook stories into multiple variation stories, if applicable - [ ] Create codesandbox example under `/packages/web-components/examples/codesandbox` and include in README - [ ] Minimum 80% unit test coverage - [ ] A comment is posted in the Prod QA issue, tagging Praveen when development is finished
priority
web component develop table of contents of the react version group user story as a ibm com library developer i need to create the table of contents so that i can provide ibm com adopter developers a web component version for every react version available in the ibm com library additional information story within storybook with corresponding knobs utilize carbon create with shadow dom and custom elements standards see the epic for the design and functional specs information prod qa testing issue acceptance criteria include readme for the web component and corresponding styles create web components styles in styles package no custom styles in web components package do not create knobs in storybook that include json objects break out storybook stories into multiple variation stories if applicable create codesandbox example under packages web components examples codesandbox and include in readme minimum unit test coverage a comment is posted in the prod qa issue tagging praveen when development is finished
1
759,362
26,591,525,305
IssuesEvent
2023-01-23 09:06:52
codelab-app/builder
https://api.github.com/repos/codelab-app/builder
closed
Proposal for component slots
priority: high
## The Problem The current component system works for only basic templating. For example, you can't create a useful layout component right now. ## The solution We need what's the equivalent of slots in templating tools. In Vue they are called slots, in Rails this is done through partials, in Laravel you have component slots. And in React this functionality is filled mostly by render props or by passing components as props/in context. ## Implementation I can imagine 2 ways to do this ### 1. The explicit way Users explicitly define an API for their components, similar to how we have an api for Atoms props. For example: <img width="962" alt="image" src="https://user-images.githubusercontent.com/57956282/187932904-ee36b004-bbb9-4671-9119-4faf59705218.png"> For slots we can use existing types, like RenderPropsType, ReactNodeType, ElementType. This api serves as the place of truth for defining the inputs that a component takes. The benefit of this is that that's not only applicable for slots, but we can also assign other props to the component, like strings, numbers, etc. We use this interface to generate a form for the component, just like we do for atoms. The next part is to be able to assign this slot to a particular element One way to do that is to bind it to props. Say that we have a Div atom with this API <img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187933534-e081da3a-0039-45b5-99b1-8e831d7adfdf.png"> Now we only need to connect `heroContent` from the Layout's API to the `children` of the Divs API. The easiest way I imagine is to bind it as we bind global state variables. <img width="1908" alt="image" src="https://user-images.githubusercontent.com/57956282/187933790-70b0db34-6773-4a60-9dda-604fc74fb4db.png"> This would require modifying the prop evaluating code to take into account the current component that the element is in and its props. ### The implicit way We create a new Atom Type, for example named `Slot`. The user creates a new element as usual and assigns it an Atom with type Slot: <img width="1919" alt="image" src="https://user-images.githubusercontent.com/57956282/187934585-3e701b16-624b-48ab-aef7-989c6c449acc.png"> Then on the component instance, we render a form that has all of the elements inside it with atom type Slot and we allow the user to pick a Component to render for them. <img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187935140-50fc67cc-ab9d-425f-8c5b-dd23bd7f4494.png"> The data from this form is stored on the component instance either as a separate field or as a special prop. It has the shape of a key-value object where the key is the id of the Slot-atomed element and the value is the component id to render. This is then used when evaluating the props to render the specific component instead of the slot-atomed element. This approach seems simpler, but it's less flexible since the user can't define other component props other than slots. #### Note In both implementations, we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it. Any thoughts or other ideas?
1.0
Proposal for component slots - ## The Problem The current component system works for only basic templating. For example, you can't create a useful layout component right now. ## The solution We need what's the equivalent of slots in templating tools. In Vue they are called slots, in Rails this is done through partials, in Laravel you have component slots. And in React this functionality is filled mostly by render props or by passing components as props/in context. ## Implementation I can imagine 2 ways to do this ### 1. The explicit way Users explicitly define an API for their components, similar to how we have an api for Atoms props. For example: <img width="962" alt="image" src="https://user-images.githubusercontent.com/57956282/187932904-ee36b004-bbb9-4671-9119-4faf59705218.png"> For slots we can use existing types, like RenderPropsType, ReactNodeType, ElementType. This api serves as the place of truth for defining the inputs that a component takes. The benefit of this is that that's not only applicable for slots, but we can also assign other props to the component, like strings, numbers, etc. We use this interface to generate a form for the component, just like we do for atoms. The next part is to be able to assign this slot to a particular element One way to do that is to bind it to props. Say that we have a Div atom with this API <img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187933534-e081da3a-0039-45b5-99b1-8e831d7adfdf.png"> Now we only need to connect `heroContent` from the Layout's API to the `children` of the Divs API. The easiest way I imagine is to bind it as we bind global state variables. <img width="1908" alt="image" src="https://user-images.githubusercontent.com/57956282/187933790-70b0db34-6773-4a60-9dda-604fc74fb4db.png"> This would require modifying the prop evaluating code to take into account the current component that the element is in and its props. ### The implicit way We create a new Atom Type, for example named `Slot`. The user creates a new element as usual and assigns it an Atom with type Slot: <img width="1919" alt="image" src="https://user-images.githubusercontent.com/57956282/187934585-3e701b16-624b-48ab-aef7-989c6c449acc.png"> Then on the component instance, we render a form that has all of the elements inside it with atom type Slot and we allow the user to pick a Component to render for them. <img width="959" alt="image" src="https://user-images.githubusercontent.com/57956282/187935140-50fc67cc-ab9d-425f-8c5b-dd23bd7f4494.png"> The data from this form is stored on the component instance either as a separate field or as a special prop. It has the shape of a key-value object where the key is the id of the Slot-atomed element and the value is the component id to render. This is then used when evaluating the props to render the specific component instead of the slot-atomed element. This approach seems simpler, but it's less flexible since the user can't define other component props other than slots. #### Note In both implementations, we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it. Any thoughts or other ideas?
priority
proposal for component slots the problem the current component system works for only basic templating for example you can t create a useful layout component right now the solution we need what s the equivalent of slots in templating tools in vue they are called slots in rails this is done through partials in laravel you have component slots and in react this functionality is filled mostly by render props or by passing components as props in context implementation i can imagine ways to do this the explicit way users explicitly define an api for their components similar to how we have an api for atoms props for example img width alt image src for slots we can use existing types like renderpropstype reactnodetype elementtype this api serves as the place of truth for defining the inputs that a component takes the benefit of this is that that s not only applicable for slots but we can also assign other props to the component like strings numbers etc we use this interface to generate a form for the component just like we do for atoms the next part is to be able to assign this slot to a particular element one way to do that is to bind it to props say that we have a div atom with this api img width alt image src now we only need to connect herocontent from the layout s api to the children of the divs api the easiest way i imagine is to bind it as we bind global state variables img width alt image src this would require modifying the prop evaluating code to take into account the current component that the element is in and its props the implicit way we create a new atom type for example named slot the user creates a new element as usual and assigns it an atom with type slot img width alt image src then on the component instance we render a form that has all of the elements inside it with atom type slot and we allow the user to pick a component to render for them img width alt image src the data from this form is stored on the component instance either as a separate field or as a special prop it has the shape of a key value object where the key is the id of the slot atomed element and the value is the component id to render this is then used when evaluating the props to render the specific component instead of the slot atomed element this approach seems simpler but it s less flexible since the user can t define other component props other than slots note in both implementations we can additionally add the ability to directly drag and drop an element to the slot to avoid creating a component for it any thoughts or other ideas
1
687,341
23,522,443,739
IssuesEvent
2022-08-19 07:34:34
roq-trading/roq-issues
https://api.github.com/repos/roq-trading/roq-issues
closed
[roq-server] Using an invalid account causes crash
bug high priority support
Validation is done correctly and an `OrderAck` with the reject is prepared for sending. When sending the `OrderAck`, the gateway needs to find the `account_id` and that's where it fails. There's a low-level optimization that allows clients to only process updates where `account_id`'s (as known to the gateway) are used for filtering. The client filtering is based on the subscription configuration and managed inside the `roq-client` library, i.e. not in the "user" code. This is therefore a problem -- an `account_id` is needed for the filtering. Somehow we need to allow for missing `account_id` as well.
1.0
[roq-server] Using an invalid account causes crash - Validation is done correctly and an `OrderAck` with the reject is prepared for sending. When sending the `OrderAck`, the gateway needs to find the `account_id` and that's where it fails. There's a low-level optimization that allows clients to only process updates where `account_id`'s (as known to the gateway) are used for filtering. The client filtering is based on the subscription configuration and managed inside the `roq-client` library, i.e. not in the "user" code. This is therefore a problem -- an `account_id` is needed for the filtering. Somehow we need to allow for missing `account_id` as well.
priority
using an invalid account causes crash validation is done correctly and an orderack with the reject is prepared for sending when sending the orderack the gateway needs to find the account id and that s where it fails there s a low level optimization that allows clients to only process updates where account id s as known to the gateway are used for filtering the client filtering is based on the subscription configuration and managed inside the roq client library i e not in the user code this is therefore a problem an account id is needed for the filtering somehow we need to allow for missing account id as well
1
393,638
11,622,676,081
IssuesEvent
2020-02-27 07:10:26
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Cannot commit edits to spatialite layers in QGIS 3.12
Bug Data Provider High Priority Regression
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** When I try to commit edits to a spatialite layer I get a yellow warning saying the following: `Could not commit changes to layer Facilities` `Errors: ERROR: 5 feature(s) not added.` `Provider errors:` `SQLite error: unknown cause` `SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)` and a red warning saying the following: `Layer Facilities: SQLite error: unknown cause SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)` This prevents me from making any edits to spatialite layers. **How to Reproduce** 1. Create a new spatialite database and layer 2. Enable editing, add features, click `save edits` 3. See error(s) **QGIS and OS versions** <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> QGIS version 3.12.0-București QGIS code revision cd141490ec Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.0.4 Running against GDAL/OGR 3.0.4 Compiled against GEOS 3.8.0-CAPI-1.13.1 Running against GEOS 3.8.0-CAPI-1.13.1 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.1 Running against PROJ Rel. 6.3.1, February 10th, 2020 OS Version Windows 10 (10.0) **Test Project and Database** [test_project_and_db.zip](https://github.com/qgis/QGIS/files/4256866/test_project_and_db.zip)
1.0
Cannot commit edits to spatialite layers in QGIS 3.12 - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** When I try to commit edits to a spatialite layer I get a yellow warning saying the following: `Could not commit changes to layer Facilities` `Errors: ERROR: 5 feature(s) not added.` `Provider errors:` `SQLite error: unknown cause` `SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)` and a red warning saying the following: `Layer Facilities: SQLite error: unknown cause SQL: INSERT INTO "facilities"("geometry",,"type","comment") VALUES (GeomFromWKB(?, 2157),,?,?)` This prevents me from making any edits to spatialite layers. **How to Reproduce** 1. Create a new spatialite database and layer 2. Enable editing, add features, click `save edits` 3. See error(s) **QGIS and OS versions** <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> QGIS version 3.12.0-București QGIS code revision cd141490ec Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.0.4 Running against GDAL/OGR 3.0.4 Compiled against GEOS 3.8.0-CAPI-1.13.1 Running against GEOS 3.8.0-CAPI-1.13.1 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.1 Running against PROJ Rel. 6.3.1, February 10th, 2020 OS Version Windows 10 (10.0) **Test Project and Database** [test_project_and_db.zip](https://github.com/qgis/QGIS/files/4256866/test_project_and_db.zip)
priority
cannot commit edits to spatialite layers in qgis bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when i try to commit edits to a spatialite layer i get a yellow warning saying the following could not commit changes to layer facilities errors error feature s not added provider errors sqlite error unknown cause sql insert into facilities geometry type comment values geomfromwkb and a red warning saying the following layer facilities sqlite error unknown cause sql insert into facilities geometry type comment values geomfromwkb this prevents me from making any edits to spatialite layers how to reproduce create a new spatialite database and layer enable editing add features click save edits see error s qgis and os versions qgis version bucurești qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel february os version windows test project and database
1
444,298
12,809,386,262
IssuesEvent
2020-07-03 15:32:04
cds-snc/covid-shield-mobile
https://api.github.com/repos/cds-snc/covid-shield-mobile
reopened
"Share your random codes" notification - Screen in background on infinite load state
bug high priority
Environment : Pixel 3XL, Android v9 (BrowserStack), version v14 Scenario : - Go to Enter you code screen from the menu - Enter a 8-digit code and submit - Click on "Agree" on the upload code screen Expected : A notification about "share you random codes" is displayed on top of the upload code screen Issue : The screen in the background of the notification is on an infinite load state ![bs_realdroid_Mobile_Google Pixel 3 XL-9 0-1440x2960 (3)](https://user-images.githubusercontent.com/47008295/85966981-2ff3ae80-b976-11ea-99fd-d505b802cb61.jpg)
1.0
"Share your random codes" notification - Screen in background on infinite load state - Environment : Pixel 3XL, Android v9 (BrowserStack), version v14 Scenario : - Go to Enter you code screen from the menu - Enter a 8-digit code and submit - Click on "Agree" on the upload code screen Expected : A notification about "share you random codes" is displayed on top of the upload code screen Issue : The screen in the background of the notification is on an infinite load state ![bs_realdroid_Mobile_Google Pixel 3 XL-9 0-1440x2960 (3)](https://user-images.githubusercontent.com/47008295/85966981-2ff3ae80-b976-11ea-99fd-d505b802cb61.jpg)
priority
share your random codes notification screen in background on infinite load state environment pixel android browserstack version scenario go to enter you code screen from the menu enter a digit code and submit click on agree on the upload code screen expected a notification about share you random codes is displayed on top of the upload code screen issue the screen in the background of the notification is on an infinite load state
1
354,862
10,573,840,280
IssuesEvent
2019-10-07 12:53:28
AY1920S1-CS2103T-W11-1/main
https://api.github.com/repos/AY1920S1-CS2103T-W11-1/main
opened
As a user, I want to mark tasks as done/undone
priority.High status.Ongoing type.Story
... so I can manage my progress in the training plans.
1.0
As a user, I want to mark tasks as done/undone - ... so I can manage my progress in the training plans.
priority
as a user i want to mark tasks as done undone so i can manage my progress in the training plans
1
394,718
11,647,939,654
IssuesEvent
2020-03-01 17:53:29
Rammelkast/AntiCheatReloaded
https://api.github.com/repos/Rammelkast/AntiCheatReloaded
closed
Speed bypass
bypass help wanted high priority
Video: https://youtu.be/H9u6GS0jNes Code: ``` package AppleClient.modules.movement; import org.lwjgl.input.Keyboard; import AppleClient.events.EventTarget; import AppleClient.events.events.EventMove; import AppleClient.events.events.EventTick; import AppleClient.modules.Module; import net.minecraft.util.MovementInput; public class MemeSpeed extends Module { public MemeSpeed() { super("ACRSpeed", Keyboard.KEY_Z, 7733063, Category.MOVEMENT, "memes", new String[] {"2fast4uboi"}, true); } @Override public void onEnable() { if(mc.thePlayer != null) { } super.onEnable(); } public void setSpeed(double speed) { final MovementInput movementInput = mc.thePlayer.movementInput; float forward = movementInput.moveForward; float strafe = movementInput.moveStrafe; float yaw = mc.thePlayer.rotationYaw; if (forward == 0.0f && strafe == 0.0f) { mc.thePlayer.motionX = 0.0; mc.thePlayer.motionZ = 0.0; } else if (forward != 0.0f) { if (strafe >= 1.0f) { yaw += ((forward > 0.0f) ? -45 : 45); strafe = 0.0f; } else if (strafe <= -1.0f) { yaw += ((forward > 0.0f) ? 45 : -45); strafe = 0.0f; } if (forward > 0.0f) { forward = 1.0f; } else if (forward < 0.0f) { forward = -1.0f; } } final double mx = Math.cos(Math.toRadians(yaw + 90.0f)); final double mz = Math.sin(Math.toRadians(yaw + 90.0f)); mc.thePlayer.motionX = forward * speed * mx + strafe * speed * mz; mc.thePlayer.motionZ = forward * speed * mz - strafe * speed * mx; if (forward == 0.0f && strafe == 0.0f) { mc.thePlayer.motionX = 0.0; mc.thePlayer.motionZ = 0.0; } } @EventTarget private void onUpdate(EventTick event) { if((mc.thePlayer.moveForward != 0.0D || mc.thePlayer.moveStrafing != 0.0D) && mc.thePlayer.onGround) { mc.thePlayer.motionY = 0.4D; } } @EventTarget public void onMove(EventMove event) { //heres how it changes between speeds to bypass. boolean hack = mc.thePlayer.ticksExisted % 2 == 0; MemeSpeed.this.setSpeed(hack ? 0.06D : 1.6D); } } ``` What it does is change between a slow speed, then goes fast for a short amount of time, before the anticheat can recognize it.
1.0
Speed bypass - Video: https://youtu.be/H9u6GS0jNes Code: ``` package AppleClient.modules.movement; import org.lwjgl.input.Keyboard; import AppleClient.events.EventTarget; import AppleClient.events.events.EventMove; import AppleClient.events.events.EventTick; import AppleClient.modules.Module; import net.minecraft.util.MovementInput; public class MemeSpeed extends Module { public MemeSpeed() { super("ACRSpeed", Keyboard.KEY_Z, 7733063, Category.MOVEMENT, "memes", new String[] {"2fast4uboi"}, true); } @Override public void onEnable() { if(mc.thePlayer != null) { } super.onEnable(); } public void setSpeed(double speed) { final MovementInput movementInput = mc.thePlayer.movementInput; float forward = movementInput.moveForward; float strafe = movementInput.moveStrafe; float yaw = mc.thePlayer.rotationYaw; if (forward == 0.0f && strafe == 0.0f) { mc.thePlayer.motionX = 0.0; mc.thePlayer.motionZ = 0.0; } else if (forward != 0.0f) { if (strafe >= 1.0f) { yaw += ((forward > 0.0f) ? -45 : 45); strafe = 0.0f; } else if (strafe <= -1.0f) { yaw += ((forward > 0.0f) ? 45 : -45); strafe = 0.0f; } if (forward > 0.0f) { forward = 1.0f; } else if (forward < 0.0f) { forward = -1.0f; } } final double mx = Math.cos(Math.toRadians(yaw + 90.0f)); final double mz = Math.sin(Math.toRadians(yaw + 90.0f)); mc.thePlayer.motionX = forward * speed * mx + strafe * speed * mz; mc.thePlayer.motionZ = forward * speed * mz - strafe * speed * mx; if (forward == 0.0f && strafe == 0.0f) { mc.thePlayer.motionX = 0.0; mc.thePlayer.motionZ = 0.0; } } @EventTarget private void onUpdate(EventTick event) { if((mc.thePlayer.moveForward != 0.0D || mc.thePlayer.moveStrafing != 0.0D) && mc.thePlayer.onGround) { mc.thePlayer.motionY = 0.4D; } } @EventTarget public void onMove(EventMove event) { //heres how it changes between speeds to bypass. boolean hack = mc.thePlayer.ticksExisted % 2 == 0; MemeSpeed.this.setSpeed(hack ? 0.06D : 1.6D); } } ``` What it does is change between a slow speed, then goes fast for a short amount of time, before the anticheat can recognize it.
priority
speed bypass video code package appleclient modules movement import org lwjgl input keyboard import appleclient events eventtarget import appleclient events events eventmove import appleclient events events eventtick import appleclient modules module import net minecraft util movementinput public class memespeed extends module public memespeed super acrspeed keyboard key z category movement memes new string true override public void onenable if mc theplayer null super onenable public void setspeed double speed final movementinput movementinput mc theplayer movementinput float forward movementinput moveforward float strafe movementinput movestrafe float yaw mc theplayer rotationyaw if forward strafe mc theplayer motionx mc theplayer motionz else if forward if strafe yaw forward strafe else if strafe yaw forward strafe if forward forward else if forward forward final double mx math cos math toradians yaw final double mz math sin math toradians yaw mc theplayer motionx forward speed mx strafe speed mz mc theplayer motionz forward speed mz strafe speed mx if forward strafe mc theplayer motionx mc theplayer motionz eventtarget private void onupdate eventtick event if mc theplayer moveforward mc theplayer movestrafing mc theplayer onground mc theplayer motiony eventtarget public void onmove eventmove event heres how it changes between speeds to bypass boolean hack mc theplayer ticksexisted memespeed this setspeed hack what it does is change between a slow speed then goes fast for a short amount of time before the anticheat can recognize it
1
153,971
5,906,750,331
IssuesEvent
2017-05-19 15:52:52
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
closed
[Request] Add jmespath
High Priority in progress Library - Request to Add/Update
**Library name:** jmespath **Git repository url:** https://github.com/jmespath/jmespath.js **npm package name or url** (if there is one): https://www.npmjs.com/package/jmespath **License (List them all if it's multiple):** Apache License, Version 2.0 **Official homepage:** http://jmespath.org/ **Wanna say something? Leave message here:** ===================== Notes from cdnjs maintainer: Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first. We encourage you to add a library via sending pull request, it'll be faster than just opening a request issue, since there are tons of issues, please wait with patience, and please don't forget to read the guidelines for contributing, thanks!!
1.0
[Request] Add jmespath - **Library name:** jmespath **Git repository url:** https://github.com/jmespath/jmespath.js **npm package name or url** (if there is one): https://www.npmjs.com/package/jmespath **License (List them all if it's multiple):** Apache License, Version 2.0 **Official homepage:** http://jmespath.org/ **Wanna say something? Leave message here:** ===================== Notes from cdnjs maintainer: Please read the [README.md](https://github.com/cdnjs/cdnjs#cdnjs-library-repository) and [CONTRIBUTING.md](https://github.com/cdnjs/cdnjs/blob/master/CONTRIBUTING.md) document first. We encourage you to add a library via sending pull request, it'll be faster than just opening a request issue, since there are tons of issues, please wait with patience, and please don't forget to read the guidelines for contributing, thanks!!
priority
add jmespath library name jmespath git repository url npm package name or url if there is one license list them all if it s multiple apache license version official homepage wanna say something leave message here notes from cdnjs maintainer please read the and document first we encourage you to add a library via sending pull request it ll be faster than just opening a request issue since there are tons of issues please wait with patience and please don t forget to read the guidelines for contributing thanks
1
161,676
6,132,993,337
IssuesEvent
2017-06-25 09:21:29
play2-maven-plugin/play2-maven-plugin
https://api.github.com/repos/play2-maven-plugin/play2-maven-plugin
closed
Upgrade Play! version from 2.6.0-RC2 to 2.6.0
Component-Maven-Plugin Component-Provider26 Priority-High Type-Task
Upgrade: - `play` version from `2.6.0-RC2` to `2.6.0` - `twirl` version from `1.3.0` to `1.3.2` - `ebean-agent` version from `10.1.7` to `10.3.1` (used in `play-ebean` version `4.0.2`) Upgrade in documentation and test projects: - `akka-*` dependencies versions to `2.5.3` - `play-ebean` version to `4.0.2` - `play-json` version to `2.6.0` - `play-slick` version to `3.0.0` - `hibernate-entitymanager` version to `5.2.10.Final` - `scalatestplus-play` version to `3.0.0`
1.0
Upgrade Play! version from 2.6.0-RC2 to 2.6.0 - Upgrade: - `play` version from `2.6.0-RC2` to `2.6.0` - `twirl` version from `1.3.0` to `1.3.2` - `ebean-agent` version from `10.1.7` to `10.3.1` (used in `play-ebean` version `4.0.2`) Upgrade in documentation and test projects: - `akka-*` dependencies versions to `2.5.3` - `play-ebean` version to `4.0.2` - `play-json` version to `2.6.0` - `play-slick` version to `3.0.0` - `hibernate-entitymanager` version to `5.2.10.Final` - `scalatestplus-play` version to `3.0.0`
priority
upgrade play version from to upgrade play version from to twirl version from to ebean agent version from to used in play ebean version upgrade in documentation and test projects akka dependencies versions to play ebean version to play json version to play slick version to hibernate entitymanager version to final scalatestplus play version to
1
681,150
23,298,773,054
IssuesEvent
2022-08-07 01:45:32
zot4plan/Zot4Plan
https://api.github.com/repos/zot4plan/Zot4Plan
opened
INS-15 Tracking taken GE courses
Priority: high Type: feature request
**Story** Users want to know how many and which GE courses have been taken. **Requirement** 1. Each GE category has a badge which display the number of taken courses 2. Show list of taken courses when hovering the badge
1.0
INS-15 Tracking taken GE courses - **Story** Users want to know how many and which GE courses have been taken. **Requirement** 1. Each GE category has a badge which display the number of taken courses 2. Show list of taken courses when hovering the badge
priority
ins tracking taken ge courses story users want to know how many and which ge courses have been taken requirement each ge category has a badge which display the number of taken courses show list of taken courses when hovering the badge
1
449,301
12,966,629,447
IssuesEvent
2020-07-21 01:08:52
dhowe/Website
https://api.github.com/repos/dhowe/Website
closed
Ready new design to go live
priority: high
lets get the new website design pushed live: - [x] check resolution of all images and replace any pixelated/low-res - [x] check display in all mobile
1.0
Ready new design to go live - lets get the new website design pushed live: - [x] check resolution of all images and replace any pixelated/low-res - [x] check display in all mobile
priority
ready new design to go live lets get the new website design pushed live check resolution of all images and replace any pixelated low res check display in all mobile
1
374,480
11,091,183,010
IssuesEvent
2019-12-15 10:28:58
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
excise.wb.gov.in - see bug description
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Firefox 72.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx **Browser / Version**: Firefox 72.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: the page whichi was working dssapeared and new page started **Steps to Reproduce**: i lost the page which iwas working [![Screenshot Description](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87-thumb.jpeg)](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191210230245</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> <p>Console Messages:</p> <pre> [{'level': 'error', 'log': ["SyntaxError: expected expression, got '}'"], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/CSS/menu.js', 'pos': '216:10'}, {'level': 'warn', 'log': ['This page uses the non standard property zoom. Consider using calc() in the relevant property values, or using transform along with transform-origin: 0 0.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '9:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '10:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '11:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '12:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/js/demos.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '21:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/themeroller/themeswitchertool/.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '22:1'}, {'level': 'error', 'log': ['TypeError: $(...).simplyScroll is not a function'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '153:28'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'error', 'log': ['ReferenceError: theme_path is not defined'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/nic9e1a.js?oo10gf', 'pos': '221:5'}, {'level': 'error', 'log': ['TypeError: this.mqo is null'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Javascript/marquee.js', 'pos': '14:202'}] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
excise.wb.gov.in - see bug description - <!-- @browser: Firefox 72.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx **Browser / Version**: Firefox 72.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: the page whichi was working dssapeared and new page started **Steps to Reproduce**: i lost the page which iwas working [![Screenshot Description](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87-thumb.jpeg)](https://webcompat.com/uploads/2019/12/151a4e01-abc0-4731-a07e-41417c726d87.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191210230245</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> <p>Console Messages:</p> <pre> [{'level': 'error', 'log': ["SyntaxError: expected expression, got '}'"], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/CSS/menu.js', 'pos': '216:10'}, {'level': 'warn', 'log': ['This page uses the non standard property zoom. Consider using calc() in the relevant property values, or using transform along with transform-origin: 0 0.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '9:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.simplyscroll2.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '10:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.jtweetsanywhere.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '11:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/WBSBCL/Bevco/NIC/js/jquery.bxSlider.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '12:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/js/demos.js was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/js/demos.js.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '21:1'}, {'level': 'warn', 'log': ['The script from https://excise.wb.gov.in/themeroller/themeswitchertool/ was loaded even though its MIME type () is not a valid JavaScript MIME type.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '0:0'}, {'level': 'warn', 'log': ['Loading failed for the <script> with source https://excise.wb.gov.in/themeroller/themeswitchertool/.'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '22:1'}, {'level': 'error', 'log': ['TypeError: $(...).simplyScroll is not a function'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Common/NIC_Home.aspx', 'pos': '153:28'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'log', 'log': ['[cycle] terminating; zero elements found by selector'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/jquery.cycle.all.2.749e1a.js?oo10gf', 'pos': '19:18'}, {'level': 'error', 'log': ['ReferenceError: theme_path is not defined'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Rectangle_Box/nic9e1a.js?oo10gf', 'pos': '221:5'}, {'level': 'error', 'log': ['TypeError: this.mqo is null'], 'uri': 'https://excise.wb.gov.in/WBSBCL/Bevco/NIC/Javascript/marquee.js', 'pos': '14:202'}] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
excise wb gov in see bug description url browser version firefox operating system windows tested another browser unknown problem type something else description the page whichi was working dssapeared and new page started steps to reproduce i lost the page which iwas working browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false console messages uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos level error log uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level log log terminating zero elements found by selector uri pos level error log uri pos level error log uri pos from with ❤️
1
545,354
15,948,800,697
IssuesEvent
2021-04-15 06:29:08
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.3 release-184] Can't migrate White tiger cycle 10
Category: Tech Priority: High Squad: Wild Turkey Status: Fixed Type: Bug
https://drive.google.com/file/d/1vBiHPylX02eiHy53IfosJ8ZJSu-62x9p/view?usp=sharing [log_210408120622.log](https://github.com/StrangeLoopGames/EcoIssues/files/6277913/log_210408120622.log) ``` Failed to start the server. Exception was Exception: NullReferenceException Message:Object reference not set to an instance of an object. Source:Eco.Gameplay System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Gameplay.Items.WorkOrder.get_Product() at Eco.Gameplay.Items.WorkOrder.UILinkContent() at Eco.Gameplay.Systems.TextLinks.UILinkExtensions.UILink(ILinkable linkable) at Eco.Gameplay.Utils.SimpleEntry.get_MarkedUpName() at Eco.Core.Utils.PropertyScanning.PropertyScanner.SetupValidity(ScanScope scope, ScanSettings settings, ScanResults results) at Eco.Core.Utils.PropertyScanning.PropertyScanner.ScanObj(Object root, ScanSettings settings) at Eco.Core.Utils.PropertyScanning.PropertyScanner.Scan(ScanSettings settings) at Eco.Gameplay.Utils.SimpleEntry.Initialize() at Eco.Gameplay.Economy.WorkParties.WorkParty.Initialize() at Eco.Core.Systems.Registrar.Initialize() at Eco.Core.Systems.Registrars.Init() at Eco.Core.Utils.Initializer.Initialize() at Eco.Server.PluginManager.InitializeAsync(StartupInfo startupInfo) at Eco.Server.Startup.StartAsync(StartupInfo startupInfo) ```
1.0
[0.9.3 release-184] Can't migrate White tiger cycle 10 - https://drive.google.com/file/d/1vBiHPylX02eiHy53IfosJ8ZJSu-62x9p/view?usp=sharing [log_210408120622.log](https://github.com/StrangeLoopGames/EcoIssues/files/6277913/log_210408120622.log) ``` Failed to start the server. Exception was Exception: NullReferenceException Message:Object reference not set to an instance of an object. Source:Eco.Gameplay System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Gameplay.Items.WorkOrder.get_Product() at Eco.Gameplay.Items.WorkOrder.UILinkContent() at Eco.Gameplay.Systems.TextLinks.UILinkExtensions.UILink(ILinkable linkable) at Eco.Gameplay.Utils.SimpleEntry.get_MarkedUpName() at Eco.Core.Utils.PropertyScanning.PropertyScanner.SetupValidity(ScanScope scope, ScanSettings settings, ScanResults results) at Eco.Core.Utils.PropertyScanning.PropertyScanner.ScanObj(Object root, ScanSettings settings) at Eco.Core.Utils.PropertyScanning.PropertyScanner.Scan(ScanSettings settings) at Eco.Gameplay.Utils.SimpleEntry.Initialize() at Eco.Gameplay.Economy.WorkParties.WorkParty.Initialize() at Eco.Core.Systems.Registrar.Initialize() at Eco.Core.Systems.Registrars.Init() at Eco.Core.Utils.Initializer.Initialize() at Eco.Server.PluginManager.InitializeAsync(StartupInfo startupInfo) at Eco.Server.Startup.StartAsync(StartupInfo startupInfo) ```
priority
can t migrate white tiger cycle failed to start the server exception was exception nullreferenceexception message object reference not set to an instance of an object source eco gameplay system nullreferenceexception object reference not set to an instance of an object at eco gameplay items workorder get product at eco gameplay items workorder uilinkcontent at eco gameplay systems textlinks uilinkextensions uilink ilinkable linkable at eco gameplay utils simpleentry get markedupname at eco core utils propertyscanning propertyscanner setupvalidity scanscope scope scansettings settings scanresults results at eco core utils propertyscanning propertyscanner scanobj object root scansettings settings at eco core utils propertyscanning propertyscanner scan scansettings settings at eco gameplay utils simpleentry initialize at eco gameplay economy workparties workparty initialize at eco core systems registrar initialize at eco core systems registrars init at eco core utils initializer initialize at eco server pluginmanager initializeasync startupinfo startupinfo at eco server startup startasync startupinfo startupinfo
1
585,670
17,514,187,860
IssuesEvent
2021-08-11 03:42:12
encorelab/ck-board
https://api.github.com/repos/encorelab/ck-board
closed
Move post modification buttons on post
bug high priority
Move edit and delete buttons on the post objects themselves, instead of having them on the side toolbar.
1.0
Move post modification buttons on post - Move edit and delete buttons on the post objects themselves, instead of having them on the side toolbar.
priority
move post modification buttons on post move edit and delete buttons on the post objects themselves instead of having them on the side toolbar
1
582,870
17,372,779,126
IssuesEvent
2021-07-30 16:07:18
MaibornWolff/codecharta
https://api.github.com/repos/MaibornWolff/codecharta
opened
Legend improvements
difficulty:low feature javascript pr-visualization priority:high
# Feature request ## Description As a user, I want the legend to be as useful as possible so that I have all necessary information without any information that is not needed or being distracted by anything. ## Acceptance criteria - Change the background color of the legend to be identical to the background of the map (it should be 100% transparent) - Remove the bounding box of the legend so that the content is visible without anything distracting the user - Remove "Outgoing" and "Incoming" Edge from the legend in case there is no such metric available - Add all used metrics to the legend and add a description in case we know what the abbreviation stands for ![image](https://user-images.githubusercontent.com/8822573/127678325-76957b55-37bc-4075-aa8b-991dd1dc7eac.png) ![image](https://user-images.githubusercontent.com/8822573/127679683-8ea9c488-83f0-496a-8b15-3829f380ca8a.png) ## Development notes (optional Task Breakdown) - [ ] Remove the bounding box and changing the background color - [ ] Remove "outgoing" and "incoming" edges from the legend, if not applicable - [ ] Add all used metrics to the legend - [ ] Add a list of known metric descriptions to the frontend and show a translation next to the entry
1.0
Legend improvements - # Feature request ## Description As a user, I want the legend to be as useful as possible so that I have all necessary information without any information that is not needed or being distracted by anything. ## Acceptance criteria - Change the background color of the legend to be identical to the background of the map (it should be 100% transparent) - Remove the bounding box of the legend so that the content is visible without anything distracting the user - Remove "Outgoing" and "Incoming" Edge from the legend in case there is no such metric available - Add all used metrics to the legend and add a description in case we know what the abbreviation stands for ![image](https://user-images.githubusercontent.com/8822573/127678325-76957b55-37bc-4075-aa8b-991dd1dc7eac.png) ![image](https://user-images.githubusercontent.com/8822573/127679683-8ea9c488-83f0-496a-8b15-3829f380ca8a.png) ## Development notes (optional Task Breakdown) - [ ] Remove the bounding box and changing the background color - [ ] Remove "outgoing" and "incoming" edges from the legend, if not applicable - [ ] Add all used metrics to the legend - [ ] Add a list of known metric descriptions to the frontend and show a translation next to the entry
priority
legend improvements feature request description as a user i want the legend to be as useful as possible so that i have all necessary information without any information that is not needed or being distracted by anything acceptance criteria change the background color of the legend to be identical to the background of the map it should be transparent remove the bounding box of the legend so that the content is visible without anything distracting the user remove outgoing and incoming edge from the legend in case there is no such metric available add all used metrics to the legend and add a description in case we know what the abbreviation stands for development notes optional task breakdown remove the bounding box and changing the background color remove outgoing and incoming edges from the legend if not applicable add all used metrics to the legend add a list of known metric descriptions to the frontend and show a translation next to the entry
1
563,454
16,685,275,210
IssuesEvent
2021-06-08 07:22:38
nlpsandbox/nlpsandbox.io
https://api.github.com/repos/nlpsandbox/nlpsandbox.io
closed
Create multi-site compatible leaderboard
Priority: High
- [x] Select the columns in the leaderboard (@tschaffter ) - [ ] Implement the leaderboard (@thomasyu888 )
1.0
Create multi-site compatible leaderboard - - [x] Select the columns in the leaderboard (@tschaffter ) - [ ] Implement the leaderboard (@thomasyu888 )
priority
create multi site compatible leaderboard select the columns in the leaderboard tschaffter implement the leaderboard
1
794,481
28,037,790,952
IssuesEvent
2023-03-28 16:11:17
asastats/channel
https://api.github.com/repos/asastats/channel
opened
ASA Stats is displaying a different swap price for Tinyman
bug high priority
Typing the same ASA amount in Tinyman gives a different swap value.
1.0
ASA Stats is displaying a different swap price for Tinyman - Typing the same ASA amount in Tinyman gives a different swap value.
priority
asa stats is displaying a different swap price for tinyman typing the same asa amount in tinyman gives a different swap value
1
524,799
15,223,629,826
IssuesEvent
2021-02-18 03:09:21
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Data copy from CPU to GPU use default stream in nightly version.
high priority module: cuda oncall: distributed
## 🐛 Bug After https://github.com/pytorch/pytorch/pull/46304, data copy per iteration from cpu to gpu can't overlap with last iteration's computation/communication. There is my test result which shows different behaviors by literally same code. nightly: ![image](https://user-images.githubusercontent.com/32845984/103058754-1908de80-45de-11eb-8111-94a8d1ae81b4.png) 1.5.1 release: ![image](https://user-images.githubusercontent.com/32845984/103058787-2e7e0880-45de-11eb-9a71-48b74d2f6dbd.png) ## To Reproduce Steps to reproduce the behavior: 1. Run the same code in nightly build and 1.5.1 release build. Use nsys profiling it. My test code is listed as below, you can use it do some tests. But I think the reason that incurs different behavior is clear. Is it a new feature? If so, I might not think it's a good one. ``` import os import time import sys import random import traceback import numpy as np import torch import torch.nn as nn import torch.nn.parallel import torch.distributed as dist import torch.optim import torch.utils.data import torch.utils.data.distributed import torchvision.transforms as transforms import torchvision.datasets as datasets import torchvision.models as models from torch.multiprocessing import Pool, Process import argparse class AverageMeter(object): """Computes and stores the average and current value""" def __init__(self): self.reset() def reset(self): self.val = 0 self.avg = 0 self.sum = 0 self.count = 0 def update(self, val, n=1): self.val = val self.sum += val * n self.count += n self.avg = self.sum / self.count def accuracy(output, target, topk=(1,)): """Computes the precision@k for the specified values of k""" with torch.no_grad(): maxk = max(topk) batch_size = target.size(0) _, pred = output.topk(maxk, 1, True, True) pred = pred.t() correct = pred.eq(target.view(1, -1).expand_as(pred)) res = [] for k in topk: correct_k = correct[:k].contiguous().view(-1).float().sum(0, keepdim=True) res.append(correct_k.mul_(100.0 / batch_size)) return res def train(train_loader, model, criterion, optimizer, epoch, batch_size): batch_time = AverageMeter() data_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() local_rank = os.environ['LOCAL_RANK'] # switch to train mode model.train() end = time.time() for i, (input, target) in enumerate(train_loader): # measure data loading time data_time.update(time.time() - end) # Create non_blocking tensors for distributed training # input = input.cuda(non_blocking=True) target = target.cuda(non_blocking=True) # compute output output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output, target, topk=(1, 5)) losses.update(loss.item(), input.size(0)) top1.update(prec1[0], input.size(0)) top5.update(prec5[0], input.size(0)) # compute gradients in a backward pass optimizer.zero_grad() loss.backward() # Call step of optimizer to update model params optimizer.step() # measure elapsed time if i >= 10: batch_time.update(time.time() - end) end = time.time() if local_rank == '0' and i % 10 == 0 and i > 10: # if local_rank == '0' and i > 0: print('Epoch: [{0}][{1}/{2}]\t' 'Speed {speed_now:.3f} ({speed_avg:.3f})\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Data {data_time.val:.3f} ({data_time.avg:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( epoch, i, len(train_loader), batch_time=batch_time, speed_now=batch_size/batch_time.val, speed_avg=batch_size/batch_time.avg, data_time=data_time, loss=losses, top1=top1, top5=top5)) def adjust_learning_rate(initial_lr, optimizer, epoch): """Sets the learning rate to the initial LR decayed by 10 every 30 epochs""" lr = initial_lr * (0.1 ** (epoch // 30)) for param_group in optimizer.param_groups: param_group['lr'] = lr def validate(val_loader, model, criterion): batch_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to evaluate mode model.eval() with torch.no_grad(): end = time.time() for i, (input, target) in enumerate(val_loader): input = input.cuda(non_blocking=True) target = target.cuda(non_blocking=True) # compute output output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output, target, topk=(1, 5)) losses.update(loss.item(), input.size(0)) top1.update(prec1[0], input.size(0)) top5.update(prec5[0], input.size(0)) # measure elapsed time torch.cuda.synchronize() batch_time.update(time.time() - end) end = time.time() if i % 100 == 0: print('Test: [{0}/{1}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( i, len(val_loader), batch_time=batch_time, loss=losses, top1=top1, top5=top5)) print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}' .format(top1=top1, top5=top5)) return top1.avg def start(backend): print("Collect Inputs...") # Batch Size for training and testing batch_size = 64 # Number of additional worker processes for dataloading workers = 8 # Number of epochs to train for num_epochs = 1 # Starting Learning Rate starting_lr = 0.01 # Distributed backend type # dist_backend = 'nccl' # dist_backend = 'esl' dist_backend = backend # Url used to setup distributed training dist.init_process_group(dist_backend, init_method='env://') print("Wordsize: ", os.environ['WORLD_SIZE']) local_rank = int(os.environ['LOCAL_RANK']) print("Local_Rank: ", local_rank) dp_device_ids = [local_rank] torch.cuda.set_device(local_rank) device = torch.device("cuda", local_rank) print("Initialize Model...") # Construct Model model = models.vgg16(pretrained=False).to(device) # Make model DistributedDataParallel model = torch.nn.parallel.DistributedDataParallel( model, device_ids=dp_device_ids, output_device=local_rank) # define loss function (criterion) and optimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = torch.optim.SGD( model.parameters(), starting_lr, momentum=0.9, weight_decay=1e-4) print("Initialize Dataloaders...") # Define the transform for the data. Notice, we must resize to 224x224 with this dataset and model. transform = transforms.Compose( [transforms.Resize(224), transforms.ToTensor(), transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))]) # Initialize Datasets. STL10 will automatically download if not present # trainset = datasets.STL10( # root='/workspace/dataset', split='train', download=True, transform=transform) # valset = datasets.STL10(root='/workspace/dataset', # split='test', download=True, transform=transform) trainset = datasets.FakeData(size=batch_size*int(os.environ['WORLD_SIZE'])*100, num_classes=1000, ,transform=transform) valset = datasets.FakeData(size=1000, transform=transform) # Create DistributedSampler to handle distributing the dataset across nodes when training # This can only be called after torch.distributed.init_process_group is called train_sampler = torch.utils.data.distributed.DistributedSampler(trainset) # Create the Dataloaders to feed data to the training and validation steps train_loader = torch.utils.data.DataLoader(trainset, batch_size=batch_size, shuffle=(train_sampler is None), num_workers=workers, pin_memory=True, sampler=train_sampler) val_loader = torch.utils.data.DataLoader(valset, batch_size=batch_size, shuffle=False, num_workers=workers, pin_memory=True) best_prec1 = 0 for epoch in range(num_epochs): # Set epoch count for DistributedSampler train_sampler.set_epoch(epoch) # Adjust learning rate according to schedule adjust_learning_rate(starting_lr, optimizer, epoch) # train for one epoch print("\nBegin Training Epoch {}".format(epoch + 1)) train(train_loader, model, criterion, optimizer, epoch, batch_size*int(os.environ['WORLD_SIZE'])) # # evaluate on validation set # print("Begin Validation @ Epoch {}".format(epoch + 1)) # prec1 = validate(val_loader, model, criterion) # # remember best prec@1 and save checkpoint if desired # # is_best = prec1 > best_prec1 # best_prec1 = max(prec1, best_prec1) # print("Epoch Summary: ") # print("\tEpoch Accuracy: {}".format(prec1)) # print("\tBest Accuracy: {}".format(best_prec1)) def set_seed(self, seed=0): random.seed(seed) np.random.seed(seed) torch.manual_seed(seed) #为CPU设置种子用于生成随机数,以使得结果是确定的 torch.cuda.manual_seed(seed) #为当前GPU设置随机种子; torch.cuda.manual_seed_all(seed) #如果使用多个GPU,为所有的GPU设置种子。 torch.backends.cudnn.deterministic = True # CPU和GPU结果一致 torch.backends.cudnn.benchmark = False if __name__ == '__main__': # set_seed(42) parser = argparse.ArgumentParser() parser.add_argument("--local_rank", type=int) args = parser.parse_args() os.environ['LOCAL_RANK'] = str(args.local_rank) backend = nccl start(backend) # usage: python -m torch.distributed.launch --nproc_per_node=NUM_GPUS_YOU_HAVE --nnodes=2 --node_rank=1 --master_addr="127.0.0.1" --master_port=1997 YOUR_TRAINING_SCRIPT.py (--arg1 --arg2 --arg3 and all other arguments of your training script) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment ``` PyTorch version: 1.5.1 │12:02:37 eth0 600B/s 0B/s Is debug build: No │12:02:37 eth0 600B/s 0B/s CUDA used to build PyTorch: 10.0 │12:02:37 eth0 1.17KB/s 14.02KB/s │12:02:38 eth0 600B/s 0B/s OS: Ubuntu 16.04.6 LTS │12:02:38 eth0 600B/s 0B/s GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.12) 5.4.0 20160609 │12:02:38 eth0 600B/s 0B/s CMake version: version 3.5.1 │12:02:38 eth0 600B/s 0B/s │12:02:38 eth0 600B/s 0B/s Python version: 3.7 │12:02:38 eth0 600B/s 0B/s Is CUDA available: Yes │12:02:38 eth0 600B/s 0B/s CUDA runtime version: 10.0.130 │12:02:38 eth0 600B/s 0B/s GPU models and configuration: │12:02:39 eth0 600B/s 0B/s GPU 0: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 1: Tesla V100-SXM2-16GB │time Interface RX TX GPU 2: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 3: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s GPU 4: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 5: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s GPU 6: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 7: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s │12:02:39 eth0 600B/s 0B/s Nvidia driver version: 418.74 │12:02:40 eth0 600B/s 0B/s cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 │12:02:40 eth0 600B/s 0B/s │12:02:40 eth0 600B/s 0B/s Versions of relevant libraries: │12:02:40 eth0 600B/s 0B/s [pip3] numpy==1.19.2 │12:02:40 eth0 600B/s 0B/s [pip3] torch==1.5.1 │12:02:40 eth0 600B/s 0B/s [pip3] torchsummary==1.5.1 │12:02:40 eth0 600B/s 0B/s [pip3] torchvision==0.6.1 │12:02:40 eth0 600B/s 0B/s [conda] blas 1.0 mkl defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl 2020.2 256 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl-service 2.3.0 py37he904b0f_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl_fft 1.2.0 py37h23d657b_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl_random 1.1.1 py37h0573a6f_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] torch 1.5.1 pypi_0 pypi │time Interface RX TX [conda] torchsummary 1.5.1 pypi_0 pypi │12:02:41 eth0 600B/s 0B/s [conda] torchvision 0.6.1 pypi_0 pypi ``` cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @ngimel @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @agolynski @SciPioneer @H-Huang @mrzzd
1.0
Data copy from CPU to GPU use default stream in nightly version. - ## 🐛 Bug After https://github.com/pytorch/pytorch/pull/46304, data copy per iteration from cpu to gpu can't overlap with last iteration's computation/communication. There is my test result which shows different behaviors by literally same code. nightly: ![image](https://user-images.githubusercontent.com/32845984/103058754-1908de80-45de-11eb-8111-94a8d1ae81b4.png) 1.5.1 release: ![image](https://user-images.githubusercontent.com/32845984/103058787-2e7e0880-45de-11eb-9a71-48b74d2f6dbd.png) ## To Reproduce Steps to reproduce the behavior: 1. Run the same code in nightly build and 1.5.1 release build. Use nsys profiling it. My test code is listed as below, you can use it do some tests. But I think the reason that incurs different behavior is clear. Is it a new feature? If so, I might not think it's a good one. ``` import os import time import sys import random import traceback import numpy as np import torch import torch.nn as nn import torch.nn.parallel import torch.distributed as dist import torch.optim import torch.utils.data import torch.utils.data.distributed import torchvision.transforms as transforms import torchvision.datasets as datasets import torchvision.models as models from torch.multiprocessing import Pool, Process import argparse class AverageMeter(object): """Computes and stores the average and current value""" def __init__(self): self.reset() def reset(self): self.val = 0 self.avg = 0 self.sum = 0 self.count = 0 def update(self, val, n=1): self.val = val self.sum += val * n self.count += n self.avg = self.sum / self.count def accuracy(output, target, topk=(1,)): """Computes the precision@k for the specified values of k""" with torch.no_grad(): maxk = max(topk) batch_size = target.size(0) _, pred = output.topk(maxk, 1, True, True) pred = pred.t() correct = pred.eq(target.view(1, -1).expand_as(pred)) res = [] for k in topk: correct_k = correct[:k].contiguous().view(-1).float().sum(0, keepdim=True) res.append(correct_k.mul_(100.0 / batch_size)) return res def train(train_loader, model, criterion, optimizer, epoch, batch_size): batch_time = AverageMeter() data_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() local_rank = os.environ['LOCAL_RANK'] # switch to train mode model.train() end = time.time() for i, (input, target) in enumerate(train_loader): # measure data loading time data_time.update(time.time() - end) # Create non_blocking tensors for distributed training # input = input.cuda(non_blocking=True) target = target.cuda(non_blocking=True) # compute output output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output, target, topk=(1, 5)) losses.update(loss.item(), input.size(0)) top1.update(prec1[0], input.size(0)) top5.update(prec5[0], input.size(0)) # compute gradients in a backward pass optimizer.zero_grad() loss.backward() # Call step of optimizer to update model params optimizer.step() # measure elapsed time if i >= 10: batch_time.update(time.time() - end) end = time.time() if local_rank == '0' and i % 10 == 0 and i > 10: # if local_rank == '0' and i > 0: print('Epoch: [{0}][{1}/{2}]\t' 'Speed {speed_now:.3f} ({speed_avg:.3f})\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Data {data_time.val:.3f} ({data_time.avg:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( epoch, i, len(train_loader), batch_time=batch_time, speed_now=batch_size/batch_time.val, speed_avg=batch_size/batch_time.avg, data_time=data_time, loss=losses, top1=top1, top5=top5)) def adjust_learning_rate(initial_lr, optimizer, epoch): """Sets the learning rate to the initial LR decayed by 10 every 30 epochs""" lr = initial_lr * (0.1 ** (epoch // 30)) for param_group in optimizer.param_groups: param_group['lr'] = lr def validate(val_loader, model, criterion): batch_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to evaluate mode model.eval() with torch.no_grad(): end = time.time() for i, (input, target) in enumerate(val_loader): input = input.cuda(non_blocking=True) target = target.cuda(non_blocking=True) # compute output output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output, target, topk=(1, 5)) losses.update(loss.item(), input.size(0)) top1.update(prec1[0], input.size(0)) top5.update(prec5[0], input.size(0)) # measure elapsed time torch.cuda.synchronize() batch_time.update(time.time() - end) end = time.time() if i % 100 == 0: print('Test: [{0}/{1}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( i, len(val_loader), batch_time=batch_time, loss=losses, top1=top1, top5=top5)) print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}' .format(top1=top1, top5=top5)) return top1.avg def start(backend): print("Collect Inputs...") # Batch Size for training and testing batch_size = 64 # Number of additional worker processes for dataloading workers = 8 # Number of epochs to train for num_epochs = 1 # Starting Learning Rate starting_lr = 0.01 # Distributed backend type # dist_backend = 'nccl' # dist_backend = 'esl' dist_backend = backend # Url used to setup distributed training dist.init_process_group(dist_backend, init_method='env://') print("Wordsize: ", os.environ['WORLD_SIZE']) local_rank = int(os.environ['LOCAL_RANK']) print("Local_Rank: ", local_rank) dp_device_ids = [local_rank] torch.cuda.set_device(local_rank) device = torch.device("cuda", local_rank) print("Initialize Model...") # Construct Model model = models.vgg16(pretrained=False).to(device) # Make model DistributedDataParallel model = torch.nn.parallel.DistributedDataParallel( model, device_ids=dp_device_ids, output_device=local_rank) # define loss function (criterion) and optimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = torch.optim.SGD( model.parameters(), starting_lr, momentum=0.9, weight_decay=1e-4) print("Initialize Dataloaders...") # Define the transform for the data. Notice, we must resize to 224x224 with this dataset and model. transform = transforms.Compose( [transforms.Resize(224), transforms.ToTensor(), transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))]) # Initialize Datasets. STL10 will automatically download if not present # trainset = datasets.STL10( # root='/workspace/dataset', split='train', download=True, transform=transform) # valset = datasets.STL10(root='/workspace/dataset', # split='test', download=True, transform=transform) trainset = datasets.FakeData(size=batch_size*int(os.environ['WORLD_SIZE'])*100, num_classes=1000, ,transform=transform) valset = datasets.FakeData(size=1000, transform=transform) # Create DistributedSampler to handle distributing the dataset across nodes when training # This can only be called after torch.distributed.init_process_group is called train_sampler = torch.utils.data.distributed.DistributedSampler(trainset) # Create the Dataloaders to feed data to the training and validation steps train_loader = torch.utils.data.DataLoader(trainset, batch_size=batch_size, shuffle=(train_sampler is None), num_workers=workers, pin_memory=True, sampler=train_sampler) val_loader = torch.utils.data.DataLoader(valset, batch_size=batch_size, shuffle=False, num_workers=workers, pin_memory=True) best_prec1 = 0 for epoch in range(num_epochs): # Set epoch count for DistributedSampler train_sampler.set_epoch(epoch) # Adjust learning rate according to schedule adjust_learning_rate(starting_lr, optimizer, epoch) # train for one epoch print("\nBegin Training Epoch {}".format(epoch + 1)) train(train_loader, model, criterion, optimizer, epoch, batch_size*int(os.environ['WORLD_SIZE'])) # # evaluate on validation set # print("Begin Validation @ Epoch {}".format(epoch + 1)) # prec1 = validate(val_loader, model, criterion) # # remember best prec@1 and save checkpoint if desired # # is_best = prec1 > best_prec1 # best_prec1 = max(prec1, best_prec1) # print("Epoch Summary: ") # print("\tEpoch Accuracy: {}".format(prec1)) # print("\tBest Accuracy: {}".format(best_prec1)) def set_seed(self, seed=0): random.seed(seed) np.random.seed(seed) torch.manual_seed(seed) #为CPU设置种子用于生成随机数,以使得结果是确定的 torch.cuda.manual_seed(seed) #为当前GPU设置随机种子; torch.cuda.manual_seed_all(seed) #如果使用多个GPU,为所有的GPU设置种子。 torch.backends.cudnn.deterministic = True # CPU和GPU结果一致 torch.backends.cudnn.benchmark = False if __name__ == '__main__': # set_seed(42) parser = argparse.ArgumentParser() parser.add_argument("--local_rank", type=int) args = parser.parse_args() os.environ['LOCAL_RANK'] = str(args.local_rank) backend = nccl start(backend) # usage: python -m torch.distributed.launch --nproc_per_node=NUM_GPUS_YOU_HAVE --nnodes=2 --node_rank=1 --master_addr="127.0.0.1" --master_port=1997 YOUR_TRAINING_SCRIPT.py (--arg1 --arg2 --arg3 and all other arguments of your training script) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment ``` PyTorch version: 1.5.1 │12:02:37 eth0 600B/s 0B/s Is debug build: No │12:02:37 eth0 600B/s 0B/s CUDA used to build PyTorch: 10.0 │12:02:37 eth0 1.17KB/s 14.02KB/s │12:02:38 eth0 600B/s 0B/s OS: Ubuntu 16.04.6 LTS │12:02:38 eth0 600B/s 0B/s GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.12) 5.4.0 20160609 │12:02:38 eth0 600B/s 0B/s CMake version: version 3.5.1 │12:02:38 eth0 600B/s 0B/s │12:02:38 eth0 600B/s 0B/s Python version: 3.7 │12:02:38 eth0 600B/s 0B/s Is CUDA available: Yes │12:02:38 eth0 600B/s 0B/s CUDA runtime version: 10.0.130 │12:02:38 eth0 600B/s 0B/s GPU models and configuration: │12:02:39 eth0 600B/s 0B/s GPU 0: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 1: Tesla V100-SXM2-16GB │time Interface RX TX GPU 2: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 3: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s GPU 4: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 5: Tesla V100-SXM2-16GB │12:02:39 eth0 1.17KB/s 0B/s GPU 6: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s GPU 7: Tesla V100-SXM2-16GB │12:02:39 eth0 600B/s 0B/s │12:02:39 eth0 600B/s 0B/s Nvidia driver version: 418.74 │12:02:40 eth0 600B/s 0B/s cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 │12:02:40 eth0 600B/s 0B/s │12:02:40 eth0 600B/s 0B/s Versions of relevant libraries: │12:02:40 eth0 600B/s 0B/s [pip3] numpy==1.19.2 │12:02:40 eth0 600B/s 0B/s [pip3] torch==1.5.1 │12:02:40 eth0 600B/s 0B/s [pip3] torchsummary==1.5.1 │12:02:40 eth0 600B/s 0B/s [pip3] torchvision==0.6.1 │12:02:40 eth0 600B/s 0B/s [conda] blas 1.0 mkl defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl 2020.2 256 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl-service 2.3.0 py37he904b0f_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl_fft 1.2.0 py37h23d657b_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] mkl_random 1.1.1 py37h0573a6f_0 defaults │12:02:41 eth0 600B/s 0B/s [conda] torch 1.5.1 pypi_0 pypi │time Interface RX TX [conda] torchsummary 1.5.1 pypi_0 pypi │12:02:41 eth0 600B/s 0B/s [conda] torchvision 0.6.1 pypi_0 pypi ``` cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @ngimel @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @agolynski @SciPioneer @H-Huang @mrzzd
priority
data copy from cpu to gpu use default stream in nightly version 🐛 bug after data copy per iteration from cpu to gpu can t overlap with last iteration s computation communication there is my test result which shows different behaviors by literally same code nightly release to reproduce steps to reproduce the behavior run the same code in nightly build and release build use nsys profiling it my test code is listed as below you can use it do some tests but i think the reason that incurs different behavior is clear is it a new feature if so i might not think it s a good one import os import time import sys import random import traceback import numpy as np import torch import torch nn as nn import torch nn parallel import torch distributed as dist import torch optim import torch utils data import torch utils data distributed import torchvision transforms as transforms import torchvision datasets as datasets import torchvision models as models from torch multiprocessing import pool process import argparse class averagemeter object computes and stores the average and current value def init self self reset def reset self self val self avg self sum self count def update self val n self val val self sum val n self count n self avg self sum self count def accuracy output target topk computes the precision k for the specified values of k with torch no grad maxk max topk batch size target size pred output topk maxk true true pred pred t correct pred eq target view expand as pred res for k in topk correct k correct contiguous view float sum keepdim true res append correct k mul batch size return res def train train loader model criterion optimizer epoch batch size batch time averagemeter data time averagemeter losses averagemeter averagemeter averagemeter local rank os environ switch to train mode model train end time time for i input target in enumerate train loader measure data loading time data time update time time end create non blocking tensors for distributed training input input cuda non blocking true target target cuda non blocking true compute output output model input loss criterion output target measure accuracy and record loss accuracy output target topk losses update loss item input size update input size update input size compute gradients in a backward pass optimizer zero grad loss backward call step of optimizer to update model params optimizer step measure elapsed time if i batch time update time time end end time time if local rank and i and i if local rank and i print epoch t speed speed now speed avg t time batch time val batch time avg t data data time val data time avg t loss loss val loss avg t prec val avg t prec val avg format epoch i len train loader batch time batch time speed now batch size batch time val speed avg batch size batch time avg data time data time loss losses def adjust learning rate initial lr optimizer epoch sets the learning rate to the initial lr decayed by every epochs lr initial lr epoch for param group in optimizer param groups param group lr def validate val loader model criterion batch time averagemeter losses averagemeter averagemeter averagemeter switch to evaluate mode model eval with torch no grad end time time for i input target in enumerate val loader input input cuda non blocking true target target cuda non blocking true compute output output model input loss criterion output target measure accuracy and record loss accuracy output target topk losses update loss item input size update input size update input size measure elapsed time torch cuda synchronize batch time update time time end end time time if i print test t time batch time val batch time avg t loss loss val loss avg t prec val avg t prec val avg format i len val loader batch time batch time loss losses print prec avg prec avg format return avg def start backend print collect inputs batch size for training and testing batch size number of additional worker processes for dataloading workers number of epochs to train for num epochs starting learning rate starting lr distributed backend type dist backend nccl dist backend esl dist backend backend url used to setup distributed training dist init process group dist backend init method env print wordsize os environ local rank int os environ print local rank local rank dp device ids torch cuda set device local rank device torch device cuda local rank print initialize model construct model model models pretrained false to device make model distributeddataparallel model torch nn parallel distributeddataparallel model device ids dp device ids output device local rank define loss function criterion and optimizer criterion nn crossentropyloss cuda optimizer torch optim sgd model parameters starting lr momentum weight decay print initialize dataloaders define the transform for the data notice we must resize to with this dataset and model transform transforms compose transforms resize transforms totensor transforms normalize initialize datasets will automatically download if not present trainset datasets root workspace dataset split train download true transform transform valset datasets root workspace dataset split test download true transform transform trainset datasets fakedata size batch size int os environ num classes transform transform valset datasets fakedata size transform transform create distributedsampler to handle distributing the dataset across nodes when training this can only be called after torch distributed init process group is called train sampler torch utils data distributed distributedsampler trainset create the dataloaders to feed data to the training and validation steps train loader torch utils data dataloader trainset batch size batch size shuffle train sampler is none num workers workers pin memory true sampler train sampler val loader torch utils data dataloader valset batch size batch size shuffle false num workers workers pin memory true best for epoch in range num epochs set epoch count for distributedsampler train sampler set epoch epoch adjust learning rate according to schedule adjust learning rate starting lr optimizer epoch train for one epoch print nbegin training epoch format epoch train train loader model criterion optimizer epoch batch size int os environ evaluate on validation set print begin validation epoch format epoch validate val loader model criterion remember best prec and save checkpoint if desired is best best best max best print epoch summary print tepoch accuracy format print tbest accuracy format best def set seed self seed random seed seed np random seed seed torch manual seed seed 为cpu设置种子用于生成随机数,以使得结果是确定的 torch cuda manual seed seed 为当前gpu设置随机种子; torch cuda manual seed all seed 如果使用多个gpu,为所有的gpu设置种子。 torch backends cudnn deterministic true cpu和gpu结果一致 torch backends cudnn benchmark false if name main set seed parser argparse argumentparser parser add argument local rank type int args parser parse args os environ str args local rank backend nccl start backend usage python m torch distributed launch nproc per node num gpus you have nnodes node rank master addr master port your training script py and all other arguments of your training script expected behavior environment pytorch version │ s s is debug build no │ s s cuda used to build pytorch │ s s │ s s os ubuntu lts │ s s gcc version ubuntu │ s s cmake version version │ s s │ s s python version │ s s is cuda available yes │ s s cuda runtime version │ s s gpu models and configuration │ s s gpu tesla │ s s gpu tesla │time interface rx tx gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s gpu tesla │ s s │ s s nvidia driver version │ s s cudnn version usr lib linux gnu libcudnn so │ s s │ s s versions of relevant libraries │ s s numpy │ s s torch │ s s torchsummary │ s s torchvision │ s s blas mkl defaults │ s s mkl defaults │ s s mkl service defaults │ s s mkl fft defaults │ s s mkl random defaults │ s s torch pypi pypi │time interface rx tx torchsummary pypi pypi │ s s torchvision pypi pypi cc ezyang gchanan bdhirsh jbschlosser ngimel pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jiayisuse agolynski scipioneer h huang mrzzd
1
454,599
13,104,449,841
IssuesEvent
2020-08-04 10:17:48
dnd-mentee-3rd/dnd-mentee-3rd-1-planto
https://api.github.com/repos/dnd-mentee-3rd/dnd-mentee-3rd-1-planto
closed
로그인 View 구현
Android Priority: High iOS
## 필요한 요소 - Email - [x] iOS - [x] Android - 비밀번호 - [x] iOS - [x] Android - 자동 로그인 체크박스 - [x] iOS - [x] Android - 로그인 버튼 - [x] iOS - [x] Android - 비밀번호 찾기 버튼 - [x] iOS - [x] Android ## 필요한 기능 - 이메일 형식 검사 - [x] iOS - [x] Android - 이메일과 비밀번호가 등록된 정보와 맞는지 검사 - [x] iOS - [x] Android - 자동 로그인 버튼 토글 - [x] iOS - [x] Android - 로그인 버튼 액션 - 정보 일치 시 인증 정보 저장 - [x] iOS - [x] Android - 정보 불일치 시 정보 확인 요청 Alert - [x] iOS - [x] Android - 비밀번호 찾기 버튼 액션 (추후 진행)
1.0
로그인 View 구현 - ## 필요한 요소 - Email - [x] iOS - [x] Android - 비밀번호 - [x] iOS - [x] Android - 자동 로그인 체크박스 - [x] iOS - [x] Android - 로그인 버튼 - [x] iOS - [x] Android - 비밀번호 찾기 버튼 - [x] iOS - [x] Android ## 필요한 기능 - 이메일 형식 검사 - [x] iOS - [x] Android - 이메일과 비밀번호가 등록된 정보와 맞는지 검사 - [x] iOS - [x] Android - 자동 로그인 버튼 토글 - [x] iOS - [x] Android - 로그인 버튼 액션 - 정보 일치 시 인증 정보 저장 - [x] iOS - [x] Android - 정보 불일치 시 정보 확인 요청 Alert - [x] iOS - [x] Android - 비밀번호 찾기 버튼 액션 (추후 진행)
priority
로그인 view 구현 필요한 요소 email ios android 비밀번호 ios android 자동 로그인 체크박스 ios android 로그인 버튼 ios android 비밀번호 찾기 버튼 ios android 필요한 기능 이메일 형식 검사 ios android 이메일과 비밀번호가 등록된 정보와 맞는지 검사 ios android 자동 로그인 버튼 토글 ios android 로그인 버튼 액션 정보 일치 시 인증 정보 저장 ios android 정보 불일치 시 정보 확인 요청 alert ios android 비밀번호 찾기 버튼 액션 추후 진행
1
746,796
26,046,512,838
IssuesEvent
2022-12-22 14:50:07
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
opened
[PTR][Dungeon] Zul Aman
Dungeon Item Loot NPC Spell Priority: High Status: Confirmed
**How to reproduce:** 1. akil'zon gauntlet event chest got no loot, tested it a lot even with all bosses dead i didnt see any loots. (beside gold satchel) ![image](https://user-images.githubusercontent.com/36106057/209156216-4662574f-e327-4f2a-a836-c1d9480bc40c.png)   2. Jan alai if you stay in a corner out of range of him he wont engage you in melee combat also his firebomb spell is not working as intended he is not dropping any bombs ![image](https://user-images.githubusercontent.com/36106057/209138065-b013b268-1682-48b4-8d0d-7be8e0751ab0.png) ![image](https://user-images.githubusercontent.com/36106057/209138129-a712946d-5836-4e80-bf09-91496a9454e8.png)   3. Zandalari hierophant trash pack before hex lord, hierophant simply wont react after pulling his trash pack until you attack him directly ![image](https://user-images.githubusercontent.com/36106057/209140458-c7f5b717-c3bb-4777-aa06-3791b6a671e5.png)   4. halazzi water totem got too much hp it should be 100k hp based on this https://cata-twinhead.twinstar.cz/?npc=52755 and this video https://youtu.be/w80XUSB3ls4?t=234 right now it has more than 1mil hp ![image](https://user-images.githubusercontent.com/36106057/209150297-9ef165c3-3457-40b7-ba25-c9d94598043b.png) 4. on hexlord malacrass shackling the darkheart wont pull the boss ![image](https://user-images.githubusercontent.com/36106057/209140645-c4e7de22-7651-44f7-9718-a0355f7dc236.png)
1.0
[PTR][Dungeon] Zul Aman - **How to reproduce:** 1. akil'zon gauntlet event chest got no loot, tested it a lot even with all bosses dead i didnt see any loots. (beside gold satchel) ![image](https://user-images.githubusercontent.com/36106057/209156216-4662574f-e327-4f2a-a836-c1d9480bc40c.png)   2. Jan alai if you stay in a corner out of range of him he wont engage you in melee combat also his firebomb spell is not working as intended he is not dropping any bombs ![image](https://user-images.githubusercontent.com/36106057/209138065-b013b268-1682-48b4-8d0d-7be8e0751ab0.png) ![image](https://user-images.githubusercontent.com/36106057/209138129-a712946d-5836-4e80-bf09-91496a9454e8.png)   3. Zandalari hierophant trash pack before hex lord, hierophant simply wont react after pulling his trash pack until you attack him directly ![image](https://user-images.githubusercontent.com/36106057/209140458-c7f5b717-c3bb-4777-aa06-3791b6a671e5.png)   4. halazzi water totem got too much hp it should be 100k hp based on this https://cata-twinhead.twinstar.cz/?npc=52755 and this video https://youtu.be/w80XUSB3ls4?t=234 right now it has more than 1mil hp ![image](https://user-images.githubusercontent.com/36106057/209150297-9ef165c3-3457-40b7-ba25-c9d94598043b.png) 4. on hexlord malacrass shackling the darkheart wont pull the boss ![image](https://user-images.githubusercontent.com/36106057/209140645-c4e7de22-7651-44f7-9718-a0355f7dc236.png)
priority
zul aman how to reproduce akil zon gauntlet event chest got no loot tested it a lot even with all bosses dead i didnt see any loots beside gold satchel   jan alai if you stay in a corner out of range of him he wont engage you in melee combat also his firebomb spell is not working as intended he is not dropping any bombs   zandalari hierophant trash pack before hex lord hierophant simply wont react after pulling his trash pack until you attack him directly   halazzi water totem got too much hp it should be hp based on this and this video right now it has more than hp on hexlord malacrass shackling the darkheart wont pull the boss
1
757,883
26,533,517,563
IssuesEvent
2023-01-19 14:11:15
fractal-analytics-platform/fractal-server
https://api.github.com/repos/fractal-analytics-platform/fractal-server
closed
JobExecutionError without relevant error message
High Priority
I'm trying to run a new example and my workflow fails during the `Convert Yokogawa to OME-Zarr` with a `JobExecutionError`. The weird thing is that the .err file of the job says the job finished successfully: ``` Matplotlib created a temporary config/cache directory at /tmp/matplotlib-wrbssjh0 because the default path (/data/homes/fractal/.cache/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing. 2023-01-16 13:49:38,878; INFO; START yokogawa_to_ome_zarr task 2023-01-16 13:49:39,263; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C01*.png 2023-01-16 13:49:55,731; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C02*.png 2023-01-16 13:50:09,745; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A02*C03*.png 2023-01-16 13:50:39,890; INFO; END yokogawa_to_ome_zarr task ``` And the slurm .err file only contains this: ``` Content of /net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.9445845.err: chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/0.slurm.submit.sbatch': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.submit.sbatch': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/cfut.in.uoaLSyH2xsDzvmmasowgIC7ahBxNWwbF.pickle': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/workflow.log': Operation not permitted ``` But that also appears in jobs that don't fail. Thus, I don't understand why this job is failing (currently looking into this) and it's worrying that the job fails without a clear error message... @tcompa the chmod failing can't actually fail the job, right? Otherwise, I wouldn't understand why this reproducibly happens in the workflow here, but not in other examples... I'll check with a different executor with more memory (shouldn't actually be an issue, but who knows) and a different example script.
1.0
JobExecutionError without relevant error message - I'm trying to run a new example and my workflow fails during the `Convert Yokogawa to OME-Zarr` with a `JobExecutionError`. The weird thing is that the .err file of the job says the job finished successfully: ``` Matplotlib created a temporary config/cache directory at /tmp/matplotlib-wrbssjh0 because the default path (/data/homes/fractal/.cache/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing. 2023-01-16 13:49:38,878; INFO; START yokogawa_to_ome_zarr task 2023-01-16 13:49:39,263; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C01*.png 2023-01-16 13:49:55,731; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A01*C02*.png 2023-01-16 13:50:09,745; INFO; glob path: /data/homes/jluethi/v1_fractal/fractal-demos/examples/02_cardio_small/../images/10.5281_zenodo.7057076/*_B03_*A02*C03*.png 2023-01-16 13:50:39,890; INFO; END yokogawa_to_ome_zarr task ``` And the slurm .err file only contains this: ``` Content of /net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.9445845.err: chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/0.slurm.submit.sbatch': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/1_par_20200812-CardiomyocyteDifferentiation14-Cycle1_zarr_B_03_0_.slurm.submit.sbatch': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/cfut.in.uoaLSyH2xsDzvmmasowgIC7ahBxNWwbF.pickle': Operation not permitted chmod: changing permissions of '/net/nfs4/pelkmanslab-fileserver-common/data/homes/fractal/joel/fractal_v1/fractal-demos/examples/server/artifacts/workflow_000006_job_000006/workflow.log': Operation not permitted ``` But that also appears in jobs that don't fail. Thus, I don't understand why this job is failing (currently looking into this) and it's worrying that the job fails without a clear error message... @tcompa the chmod failing can't actually fail the job, right? Otherwise, I wouldn't understand why this reproducibly happens in the workflow here, but not in other examples... I'll check with a different executor with more memory (shouldn't actually be an issue, but who knows) and a different example script.
priority
jobexecutionerror without relevant error message i m trying to run a new example and my workflow fails during the convert yokogawa to ome zarr with a jobexecutionerror the weird thing is that the err file of the job says the job finished successfully matplotlib created a temporary config cache directory at tmp matplotlib because the default path data homes fractal cache matplotlib is not a writable directory it is highly recommended to set the mplconfigdir environment variable to a writable directory in particular to speed up the import of matplotlib and to better support multiprocessing info start yokogawa to ome zarr task info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info glob path data homes jluethi fractal fractal demos examples cardio small images zenodo png info end yokogawa to ome zarr task and the slurm err file only contains this content of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job par zarr b slurm err chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job slurm submit sbatch operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job par zarr b slurm submit sbatch operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job cfut in pickle operation not permitted chmod changing permissions of net pelkmanslab fileserver common data homes fractal joel fractal fractal demos examples server artifacts workflow job workflow log operation not permitted but that also appears in jobs that don t fail thus i don t understand why this job is failing currently looking into this and it s worrying that the job fails without a clear error message tcompa the chmod failing can t actually fail the job right otherwise i wouldn t understand why this reproducibly happens in the workflow here but not in other examples i ll check with a different executor with more memory shouldn t actually be an issue but who knows and a different example script
1
429,635
12,426,663,328
IssuesEvent
2020-05-24 22:25:42
stevenwaterman/musetree
https://api.github.com/repos/stevenwaterman/musetree
closed
Load starter samples into the app
High Priority enhancement good first issue help wanted
In the [official musenet app](https://openai.com/blog/musenet/), you can specify what you want the generator to start with. Add the 12 samples in the official app to musetree. It's fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song. Loading the MP3 for those samples may be difficult - 2 options: * Hit the `/audio` musenet endpoint to get the mp3 for the encoding * Hard-code the mp3 base64 into the app alongside the encoding. (Temporarily) Custom uploads are more complex and therefore in a separate issue.
1.0
Load starter samples into the app - In the [official musenet app](https://openai.com/blog/musenet/), you can specify what you want the generator to start with. Add the 12 samples in the official app to musetree. It's fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song. Loading the MP3 for those samples may be difficult - 2 options: * Hit the `/audio` musenet endpoint to get the mp3 for the encoding * Hard-code the mp3 base64 into the app alongside the encoding. (Temporarily) Custom uploads are more complex and therefore in a separate issue.
priority
load starter samples into the app in the you can specify what you want the generator to start with add the samples in the official app to musetree it s fairly simple to do by just using the official app and inspecting the network requests and extracting the encoding for the song loading the for those samples may be difficult options hit the audio musenet endpoint to get the for the encoding hard code the into the app alongside the encoding temporarily custom uploads are more complex and therefore in a separate issue
1
486,436
14,009,063,038
IssuesEvent
2020-10-29 01:21:29
AFlyingCar/WarringStates
https://api.github.com/repos/AFlyingCar/WarringStates
closed
Declaiming a Chunk in the Capitol removes Claimer
High Priority bug
Despite a message "You cannot rescind you claim on your capitol"
1.0
Declaiming a Chunk in the Capitol removes Claimer - Despite a message "You cannot rescind you claim on your capitol"
priority
declaiming a chunk in the capitol removes claimer despite a message you cannot rescind you claim on your capitol
1
156,091
5,964,210,149
IssuesEvent
2017-05-30 08:14:44
mkdo/kapow-theme
https://api.github.com/repos/mkdo/kapow-theme
closed
FE assets re-build
Priority: High Status: Completed Type: Maintenance
The compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules. Can be completed when the full-build testing is finished at the end of Phase 2.0.
1.0
FE assets re-build - The compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules. Can be completed when the full-build testing is finished at the end of Phase 2.0.
priority
fe assets re build the compiled assets that come as part of the theme need to be regenerated to take into account the changes made in the other modules can be completed when the full build testing is finished at the end of phase
1
6,899
2,595,479,709
IssuesEvent
2015-02-20 14:36:23
centre-for-educational-technology/edidaktikum
https://api.github.com/repos/centre-for-educational-technology/edidaktikum
opened
Kogumik (täiendamisel)
High Priority New feature
Kogumik on õppematerjalide kollektsioon. Kõik õppejõu rollis kasutajad saavad luua kogumikke. Kogumiku loomiseks on kaks võimalust: ...
1.0
Kogumik (täiendamisel) - Kogumik on õppematerjalide kollektsioon. Kõik õppejõu rollis kasutajad saavad luua kogumikke. Kogumiku loomiseks on kaks võimalust: ...
priority
kogumik täiendamisel kogumik on õppematerjalide kollektsioon kõik õppejõu rollis kasutajad saavad luua kogumikke kogumiku loomiseks on kaks võimalust
1
331,998
10,083,062,497
IssuesEvent
2019-07-25 12:51:43
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
`with torch.enable_grad` also works outside a `no_grad` context
high priority module: autograd module: docs small triaged
## 📚 Documentation The [`torch.autograd.enable_grad` documentation](https://pytorch.org/docs/stable/autograd.html#torch.autograd.enable_grad) says: > Enables gradient calculation inside a `no_grad` context. This has no effect outside of `no_grad`. This implies: ``` torch.set_grad_enabled(False) with torch.enable_grad: # Gradient tracking will NOT be enabled here. torch.set_grad_enabled(True) vs: with torch.no_grad(): with torch.enable_grad: # Gradient tracking IS enabled here. ``` However the observed behaviour (`1.0.1.post2`) is: ``` x = torch.tensor([.1], requires_grad=True) with torch.no_grad(): with torch.enable_grad(): y = x * 2 print(y.requires_grad) # True (as expected) with torch.set_grad_enabled(False): y = x * 2 print(y.requires_grad) # False (as expected) with torch.set_grad_enabled(False): with torch.enable_grad(): y = x * 2 print(y.requires_grad) # True, but False expected from doc quote ``` Note the last example is not "inside a `no_grad` context", but it still works. Other prior art: [PyTorch set_grad_enabled(False) vs with no_grad():](https://stackoverflow.com/a/53447634/5353461) ----------- I'm assuming the documentation is incorrect, and it should say simply: > Enables gradient calculation. Would you accept a PR?
1.0
`with torch.enable_grad` also works outside a `no_grad` context - ## 📚 Documentation The [`torch.autograd.enable_grad` documentation](https://pytorch.org/docs/stable/autograd.html#torch.autograd.enable_grad) says: > Enables gradient calculation inside a `no_grad` context. This has no effect outside of `no_grad`. This implies: ``` torch.set_grad_enabled(False) with torch.enable_grad: # Gradient tracking will NOT be enabled here. torch.set_grad_enabled(True) vs: with torch.no_grad(): with torch.enable_grad: # Gradient tracking IS enabled here. ``` However the observed behaviour (`1.0.1.post2`) is: ``` x = torch.tensor([.1], requires_grad=True) with torch.no_grad(): with torch.enable_grad(): y = x * 2 print(y.requires_grad) # True (as expected) with torch.set_grad_enabled(False): y = x * 2 print(y.requires_grad) # False (as expected) with torch.set_grad_enabled(False): with torch.enable_grad(): y = x * 2 print(y.requires_grad) # True, but False expected from doc quote ``` Note the last example is not "inside a `no_grad` context", but it still works. Other prior art: [PyTorch set_grad_enabled(False) vs with no_grad():](https://stackoverflow.com/a/53447634/5353461) ----------- I'm assuming the documentation is incorrect, and it should say simply: > Enables gradient calculation. Would you accept a PR?
priority
with torch enable grad also works outside a no grad context 📚 documentation the says enables gradient calculation inside a no grad context this has no effect outside of no grad this implies torch set grad enabled false with torch enable grad gradient tracking will not be enabled here torch set grad enabled true vs with torch no grad with torch enable grad gradient tracking is enabled here however the observed behaviour is x torch tensor requires grad true with torch no grad with torch enable grad y x print y requires grad true as expected with torch set grad enabled false y x print y requires grad false as expected with torch set grad enabled false with torch enable grad y x print y requires grad true but false expected from doc quote note the last example is not inside a no grad context but it still works other prior art i m assuming the documentation is incorrect and it should say simply enables gradient calculation would you accept a pr
1
632,465
20,198,013,114
IssuesEvent
2022-02-11 12:32:06
AlpsBTE/Plot-System
https://api.github.com/repos/AlpsBTE/Plot-System
opened
[Feature] - Make the plot outline usable in code
enhancement high priority
**Is your feature request related to a problem? Please describe.** Currently, there is no way to get the polygon outline of the plot in the PlotSystem plugin. I first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region, not a poly region. For the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot. This would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for /tpll etc. **Describe the solution you'd like** Add a column "outline" to the table plotsystem_plots which contains a list of points of the polygon from the plot. When the plot is created the PlotSystemTerra PlotCreator.java adds the outline data additionally to the other data that is inserted. **Additional context** The "one-world" branch is on hold until this is released
1.0
[Feature] - Make the plot outline usable in code - **Is your feature request related to a problem? Please describe.** Currently, there is no way to get the polygon outline of the plot in the PlotSystem plugin. I first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region, not a poly region. For the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot. This would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for /tpll etc. **Describe the solution you'd like** Add a column "outline" to the table plotsystem_plots which contains a list of points of the polygon from the plot. When the plot is created the PlotSystemTerra PlotCreator.java adds the outline data additionally to the other data that is inserted. **Additional context** The "one-world" branch is on hold until this is released
priority
make the plot outline usable in code is your feature request related to a problem please describe currently there is no way to get the polygon outline of the plot in the plotsystem plugin i first tought that you could get the outline by reading the schematic file and then using the region file from it but it turns out that this region is a cuboid region not a poly region for the current plot system this made no difference since the world around it was air but for the oneworld project it is important to have access to the poly outline of the plot this would also remove the need to load the schematic file everytime when the plot outline needs to be accessed for example for tpll etc describe the solution you d like add a column outline to the table plotsystem plots which contains a list of points of the polygon from the plot when the plot is created the plotsystemterra plotcreator java adds the outline data additionally to the other data that is inserted additional context the one world branch is on hold until this is released
1
552,677
16,246,774,495
IssuesEvent
2021-05-07 15:20:41
DCS-LCSR/SignStream3
https://api.github.com/repos/DCS-LCSR/SignStream3
closed
Overlap that should not be allowed - gloss items - est. 8 hours
bug cannot reproduce priority HIGH
I'm working with file 112, which I emailed you very recently -- utterance 82. The endpoint of one sign (GROUP/TOGETHER) overlaps with the initial hold of the following sign. This should not be allowed... <img width="713" alt="screen shot 2018-06-06 at 2 10 10 pm" src="https://user-images.githubusercontent.com/13629122/41056894-9146a3ac-6993-11e8-9f5f-f316643d70a9.png"> I just deleted that initial hold, and I am (correctly) no longer allowed to add a hold, because it would involve disallowed overlap. So perhaps this is a holdover from an older version of the program that did not enforce things properly. Just letting you know, in any case.
1.0
Overlap that should not be allowed - gloss items - est. 8 hours - I'm working with file 112, which I emailed you very recently -- utterance 82. The endpoint of one sign (GROUP/TOGETHER) overlaps with the initial hold of the following sign. This should not be allowed... <img width="713" alt="screen shot 2018-06-06 at 2 10 10 pm" src="https://user-images.githubusercontent.com/13629122/41056894-9146a3ac-6993-11e8-9f5f-f316643d70a9.png"> I just deleted that initial hold, and I am (correctly) no longer allowed to add a hold, because it would involve disallowed overlap. So perhaps this is a holdover from an older version of the program that did not enforce things properly. Just letting you know, in any case.
priority
overlap that should not be allowed gloss items est hours i m working with file which i emailed you very recently utterance the endpoint of one sign group together overlaps with the initial hold of the following sign this should not be allowed img width alt screen shot at pm src i just deleted that initial hold and i am correctly no longer allowed to add a hold because it would involve disallowed overlap so perhaps this is a holdover from an older version of the program that did not enforce things properly just letting you know in any case
1
178,034
6,598,077,340
IssuesEvent
2017-09-16 00:02:32
how-is/how_is
https://api.github.com/repos/how-is/how_is
closed
move away from slop bc of conflicts
high priority
this is like the 3rd time i've had slop interfere with something, one of which prompted a new release only to encounter a different conflict the next time I tried to use it(!). so, as much as I like it, I think it'd be best to use something else. probably just Ruby's built in OptionParser, even if it's not as nice imo. this is blocking work on https://github.com/rubytogether/rubytogether.org/pull/91
1.0
move away from slop bc of conflicts - this is like the 3rd time i've had slop interfere with something, one of which prompted a new release only to encounter a different conflict the next time I tried to use it(!). so, as much as I like it, I think it'd be best to use something else. probably just Ruby's built in OptionParser, even if it's not as nice imo. this is blocking work on https://github.com/rubytogether/rubytogether.org/pull/91
priority
move away from slop bc of conflicts this is like the time i ve had slop interfere with something one of which prompted a new release only to encounter a different conflict the next time i tried to use it so as much as i like it i think it d be best to use something else probably just ruby s built in optionparser even if it s not as nice imo this is blocking work on
1
519,775
15,057,030,623
IssuesEvent
2021-02-03 21:03:20
protofire/omen-exchange
https://api.github.com/repos/protofire/omen-exchange
opened
Market list item in Market overview is not showing the correct "current prediction"
bug priority:high
Market: https://xdai.dxtest.eth.link/#/0xd3f61725f4a1905223160abec29b98e81cd569ba if you see the market details: <img width="610" alt="Bildschirmfoto 2021-02-03 um 22 02 02" src="https://user-images.githubusercontent.com/5337809/106809020-76775d00-666b-11eb-9d3f-8dc673645294.png"> Current Prediction shows: `1,920.94 USD` If you go back to the market details view, we see the current prediction is `1,657.76 USD` <img width="613" alt="Bildschirmfoto 2021-02-03 um 22 02 39" src="https://user-images.githubusercontent.com/5337809/106809150-9e66c080-666b-11eb-94e6-984af0b129f3.png">
1.0
Market list item in Market overview is not showing the correct "current prediction" - Market: https://xdai.dxtest.eth.link/#/0xd3f61725f4a1905223160abec29b98e81cd569ba if you see the market details: <img width="610" alt="Bildschirmfoto 2021-02-03 um 22 02 02" src="https://user-images.githubusercontent.com/5337809/106809020-76775d00-666b-11eb-9d3f-8dc673645294.png"> Current Prediction shows: `1,920.94 USD` If you go back to the market details view, we see the current prediction is `1,657.76 USD` <img width="613" alt="Bildschirmfoto 2021-02-03 um 22 02 39" src="https://user-images.githubusercontent.com/5337809/106809150-9e66c080-666b-11eb-94e6-984af0b129f3.png">
priority
market list item in market overview is not showing the correct current prediction market if you see the market details img width alt bildschirmfoto um src current prediction shows usd if you go back to the market details view we see the current prediction is usd img width alt bildschirmfoto um src
1
157,835
6,016,754,062
IssuesEvent
2017-06-07 07:56:29
buggins/dlangui
https://api.github.com/repos/buggins/dlangui
closed
Folder and file icons disappeared from FileDialog
bug high_priority
Since this commit https://github.com/buggins/dlangui/commit/dfe707322254e75e362853aac3226ed337a33714 no folder or file icon shown in file dialog
1.0
Folder and file icons disappeared from FileDialog - Since this commit https://github.com/buggins/dlangui/commit/dfe707322254e75e362853aac3226ed337a33714 no folder or file icon shown in file dialog
priority
folder and file icons disappeared from filedialog since this commit no folder or file icon shown in file dialog
1
343,378
10,329,021,746
IssuesEvent
2019-09-02 11:03:10
python/mypy
https://api.github.com/repos/python/mypy
closed
Error codes in # type: ignore don't suppress related notes
bug priority-0-high topic-usability
For example: ```python import nostubs # type: ignore[import] ``` Causes ``` mypy/report.py:30: note: (Stub files are from https://github.com/python/typeshed) ``` and error code 1. I didn't check carefully, but I think this affects all error messages that also emit additional notes (there are quite many of them).
1.0
Error codes in # type: ignore don't suppress related notes - For example: ```python import nostubs # type: ignore[import] ``` Causes ``` mypy/report.py:30: note: (Stub files are from https://github.com/python/typeshed) ``` and error code 1. I didn't check carefully, but I think this affects all error messages that also emit additional notes (there are quite many of them).
priority
error codes in type ignore don t suppress related notes for example python import nostubs type ignore causes mypy report py note stub files are from and error code i didn t check carefully but i think this affects all error messages that also emit additional notes there are quite many of them
1
415,685
12,133,028,555
IssuesEvent
2020-04-23 08:20:54
arkhn/pyrog
https://api.github.com/repos/arkhn/pyrog
closed
Slicing hierarchy error
Bug High Priority
When importing a profile with extensions (ex : observation-bp https://www.hl7.org/fhir/bp.html or variant http://hl7.org/fhir/uv/genomics-reporting/variant.html), the slice structure is not as expected. Example of observation-bp: * Actual state ![image](https://user-images.githubusercontent.com/48218773/78557021-5c34de80-7810-11ea-814a-cd317bef70b5.png) * Expected see https://www.hl7.org/fhir/bp.html#profile ![image](https://user-images.githubusercontent.com/48218773/78557371-f39a3180-7810-11ea-9d58-e51b1caf3e34.png)
1.0
Slicing hierarchy error - When importing a profile with extensions (ex : observation-bp https://www.hl7.org/fhir/bp.html or variant http://hl7.org/fhir/uv/genomics-reporting/variant.html), the slice structure is not as expected. Example of observation-bp: * Actual state ![image](https://user-images.githubusercontent.com/48218773/78557021-5c34de80-7810-11ea-814a-cd317bef70b5.png) * Expected see https://www.hl7.org/fhir/bp.html#profile ![image](https://user-images.githubusercontent.com/48218773/78557371-f39a3180-7810-11ea-9d58-e51b1caf3e34.png)
priority
slicing hierarchy error when importing a profile with extensions ex observation bp or variant the slice structure is not as expected example of observation bp actual state expected see
1
189,065
6,793,598,703
IssuesEvent
2017-11-01 08:23:20
dalaranwow/dalaran-wow
https://api.github.com/repos/dalaranwow/dalaran-wow
closed
Warrior Talent: Sweeping Strikes -- Defensive Stance cancels SS.
Class - Warrior Fixed - Waiting for PTR update Priority - High
**Current behaviour**: If you have Sweeping Strikes up, switching to Defensive Stance cancels it. **Expected behaviour**: SS shouldn't be cancelled by switching into DS. **Steps to reproduce the problem**: 1. Use Sweeping Strikes; 2. Go into Defensive Stance -- SS is gone. **Include proofs for this behaviour** Sweeping Strikes ID: https://wotlk-twinhead.twinstar.cz/?spell=12328 Retail proof: https://youtu.be/nWUkwISLPPI?t=290 -- watch the activation of Sweeping Strikes in correspondence to the Stance and keep an eye on both. See that upon switching to Defensive Stance, SS remains.
1.0
Warrior Talent: Sweeping Strikes -- Defensive Stance cancels SS. - **Current behaviour**: If you have Sweeping Strikes up, switching to Defensive Stance cancels it. **Expected behaviour**: SS shouldn't be cancelled by switching into DS. **Steps to reproduce the problem**: 1. Use Sweeping Strikes; 2. Go into Defensive Stance -- SS is gone. **Include proofs for this behaviour** Sweeping Strikes ID: https://wotlk-twinhead.twinstar.cz/?spell=12328 Retail proof: https://youtu.be/nWUkwISLPPI?t=290 -- watch the activation of Sweeping Strikes in correspondence to the Stance and keep an eye on both. See that upon switching to Defensive Stance, SS remains.
priority
warrior talent sweeping strikes defensive stance cancels ss current behaviour if you have sweeping strikes up switching to defensive stance cancels it expected behaviour ss shouldn t be cancelled by switching into ds steps to reproduce the problem use sweeping strikes go into defensive stance ss is gone include proofs for this behaviour sweeping strikes id retail proof watch the activation of sweeping strikes in correspondence to the stance and keep an eye on both see that upon switching to defensive stance ss remains
1
650,404
21,390,822,545
IssuesEvent
2022-04-21 06:54:06
eclipse/dirigible
https://api.github.com/repos/eclipse/dirigible
opened
[EDM] Put all the generated artefacts under a 'gen' folder
web-ide usability priority-high efforts-medium component-template
Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned.
1.0
[EDM] Put all the generated artefacts under a 'gen' folder - Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned.
priority
put all the generated artefacts under a gen folder put all the generated artefacts under a gen folder it will be easy and definitive when regeneration is needed what to be cleaned
1
474,503
13,671,187,719
IssuesEvent
2020-09-29 06:31:51
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
closed
Event Wizard publishing with deleted tickets
Priority: High bug
Sometimes the ticket display shows “there are no tickets” even though a free ticket is defined
1.0
Event Wizard publishing with deleted tickets - Sometimes the ticket display shows “there are no tickets” even though a free ticket is defined
priority
event wizard publishing with deleted tickets sometimes the ticket display shows “there are no tickets” even though a free ticket is defined
1
63,626
3,197,112,323
IssuesEvent
2015-10-01 01:18:13
MarcNo/lifelines
https://api.github.com/repos/MarcNo/lifelines
closed
docs directory missing from dist
auto-migrated high priority sourceforge
distribution lifelines-3.0.11.tar.gz file does not include docs directory Reported by: puddle Original Ticket: "bugs/71":https://sourceforge.net/p/bugs/71
1.0
docs directory missing from dist - distribution lifelines-3.0.11.tar.gz file does not include docs directory Reported by: puddle Original Ticket: "bugs/71":https://sourceforge.net/p/bugs/71
priority
docs directory missing from dist distribution lifelines tar gz file does not include docs directory reported by puddle original ticket bugs
1
630,597
20,114,377,407
IssuesEvent
2022-02-07 17:55:55
webcat12345/ngx-ui-switch
https://api.github.com/repos/webcat12345/ngx-ui-switch
closed
reverse attribute question
Priority: High Status: Available
Hello! What is the intended way of "reverse"? I currently have a boolean coming from a backend with "false" as the value, which causes the switch to be in "off" mode. I want to reverse this, so that the switch is "on" when its value is false. Is this what "reverse" should do? If that's the case, then I think it's broken.
1.0
reverse attribute question - Hello! What is the intended way of "reverse"? I currently have a boolean coming from a backend with "false" as the value, which causes the switch to be in "off" mode. I want to reverse this, so that the switch is "on" when its value is false. Is this what "reverse" should do? If that's the case, then I think it's broken.
priority
reverse attribute question hello what is the intended way of reverse i currently have a boolean coming from a backend with false as the value which causes the switch to be in off mode i want to reverse this so that the switch is on when its value is false is this what reverse should do if that s the case then i think it s broken
1
752,703
26,321,357,735
IssuesEvent
2023-01-10 00:12:24
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
closed
Input: `contentBefore` and `contentAfter` don't work with `bundleIcons`
Type: Bug :bug: Priority 1: High Partner Ask Status: In PR Component: Input
**Repro:** https://codesandbox.io/s/mystifying-austin-02jpd4?file=/example.tsx ![image](https://user-images.githubusercontent.com/20744592/208868776-ffe89ce8-ff44-4b43-ac1f-525ea3cd33e6.png) The specificity of the content styles in `Input` are more specific than the styles applied in bundled icons, the result is that both filled and regular icons are always visible https://github.com/microsoft/fluentui/blob/66bf89f634cad4a275e957d7a2214c7e73ff8c2e/packages/react-components/react-input/src/components/Input/useInputStyles.ts#L215-L235
1.0
Input: `contentBefore` and `contentAfter` don't work with `bundleIcons` - **Repro:** https://codesandbox.io/s/mystifying-austin-02jpd4?file=/example.tsx ![image](https://user-images.githubusercontent.com/20744592/208868776-ffe89ce8-ff44-4b43-ac1f-525ea3cd33e6.png) The specificity of the content styles in `Input` are more specific than the styles applied in bundled icons, the result is that both filled and regular icons are always visible https://github.com/microsoft/fluentui/blob/66bf89f634cad4a275e957d7a2214c7e73ff8c2e/packages/react-components/react-input/src/components/Input/useInputStyles.ts#L215-L235
priority
input contentbefore and contentafter don t work with bundleicons repro the specificity of the content styles in input are more specific than the styles applied in bundled icons the result is that both filled and regular icons are always visible
1
152,119
5,833,346,104
IssuesEvent
2017-05-09 01:10:37
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
closed
ALP: add support for `#replaceUrl` for cache hit improvement
Category: Ads P1: High Priority Type: Bug
ALPs have many client-resolvable query parameters which greatly reduce cache-hit by multiplying cache keys. E.g. `https://host/alp?adpos=X` will multiply on the possible space of `adpos` query parameter, even though it's never used by the page itself. The proposal is to use `https://host/alp` as the cache URL/key and replace it when AMP runtime starts up to `https://host/alp?adpos=X`. The replacing URL will be specified via viewer param as `#replaceUrl=...`. Some important notes: 1. Replacement should be done, ideally, very early on and definitely before `installDocumentStateService` is called. 2. The replacement URL must match the original URLs origin _and_ source origin
1.0
ALP: add support for `#replaceUrl` for cache hit improvement - ALPs have many client-resolvable query parameters which greatly reduce cache-hit by multiplying cache keys. E.g. `https://host/alp?adpos=X` will multiply on the possible space of `adpos` query parameter, even though it's never used by the page itself. The proposal is to use `https://host/alp` as the cache URL/key and replace it when AMP runtime starts up to `https://host/alp?adpos=X`. The replacing URL will be specified via viewer param as `#replaceUrl=...`. Some important notes: 1. Replacement should be done, ideally, very early on and definitely before `installDocumentStateService` is called. 2. The replacement URL must match the original URLs origin _and_ source origin
priority
alp add support for replaceurl for cache hit improvement alps have many client resolvable query parameters which greatly reduce cache hit by multiplying cache keys e g will multiply on the possible space of adpos query parameter even though it s never used by the page itself the proposal is to use as the cache url key and replace it when amp runtime starts up to the replacing url will be specified via viewer param as replaceurl some important notes replacement should be done ideally very early on and definitely before installdocumentstateservice is called the replacement url must match the original urls origin and source origin
1
506,485
14,666,161,824
IssuesEvent
2020-12-29 15:43:57
netlify/next-on-netlify
https://api.github.com/repos/netlify/next-on-netlify
reopened
Preview mode broken in Next 10.0.2 with target: "serverless": TypeError: e[t] is not a function
Next 10 experimental-serverless-trace priority: high
Since we've switched to the latest version of next (10.0.2), [preview mode](https://nextjs.org/docs/advanced-features/preview-mode) has been broken. Everything builds successfully, but when we load our `/api/preview` endpoint we get the following error: ``` [request] /api/preview/ TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) /usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:116 throw ex; ^ abort({}) at Error at jsStackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1070:13) at stackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1087:12) at process.abort (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:8502:44) at process.emit (events.js:310:20) at processPromiseRejections (internal/process/promises.js:209:33) at processTicksAndRejections (internal/process/task_queues.js:98:32) (Use `node --trace-uncaught ...` to show where the exception was thrown) ```
1.0
Preview mode broken in Next 10.0.2 with target: "serverless": TypeError: e[t] is not a function - Since we've switched to the latest version of next (10.0.2), [preview mode](https://nextjs.org/docs/advanced-features/preview-mode) has been broken. Everything builds successfully, but when we load our `/api/preview` endpoint we get the following error: ``` [request] /api/preview/ TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) TypeError: e[t] is not a function at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.293 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:29747) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.118 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:28826) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.692 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:17036) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.28 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:15901) at __webpack_require__ (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:30130) at Object.27 (/Users/florisverloop/Projects/site-sass/site/out_functions/next_api_preview/nextPage.js:43666:3078) /usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:116 throw ex; ^ abort({}) at Error at jsStackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1070:13) at stackTrace (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:1087:12) at process.abort (/usr/local/lib/node_modules/netlify-cli/node_modules/netlify-redirector/lib/redirects.js:8502:44) at process.emit (events.js:310:20) at processPromiseRejections (internal/process/promises.js:209:33) at processTicksAndRejections (internal/process/task_queues.js:98:32) (Use `node --trace-uncaught ...` to show where the exception was thrown) ```
priority
preview mode broken in next with target serverless typeerror e is not a function since we ve switched to the latest version of next has been broken everything builds successfully but when we load our api preview endpoint we get the following error api preview typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js typeerror e is not a function at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js at webpack require users florisverloop projects site sass site out functions next api preview nextpage js at object users florisverloop projects site sass site out functions next api preview nextpage js usr local lib node modules netlify cli node modules netlify redirector lib redirects js throw ex abort at error at jsstacktrace usr local lib node modules netlify cli node modules netlify redirector lib redirects js at stacktrace usr local lib node modules netlify cli node modules netlify redirector lib redirects js at process abort usr local lib node modules netlify cli node modules netlify redirector lib redirects js at process emit events js at processpromiserejections internal process promises js at processticksandrejections internal process task queues js use node trace uncaught to show where the exception was thrown
1
602,482
18,470,078,286
IssuesEvent
2021-10-17 15:34:24
AY2122S1-CS2103-T14-2/tp
https://api.github.com/repos/AY2122S1-CS2103-T14-2/tp
closed
Store CCA enrolments data locally
type.Story priority.High
As a user, after enrolling contacts into CCAs, I want to have it saved locally so I do not have to re-enrol my contacts every time.
1.0
Store CCA enrolments data locally - As a user, after enrolling contacts into CCAs, I want to have it saved locally so I do not have to re-enrol my contacts every time.
priority
store cca enrolments data locally as a user after enrolling contacts into ccas i want to have it saved locally so i do not have to re enrol my contacts every time
1
537,099
15,723,056,021
IssuesEvent
2021-03-29 06:55:01
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.youtube.com - desktop site instead of mobile site
browser-chrome ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Chrome 89.0.4389 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/68920 --> **URL**: https://www.youtube.com/watch?v=u0KQqOai-uA **Browser / Version**: Chrome 89.0.4389 **Operating System**: Windows 10 **Tested Another Browser**: Yes Safari **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.youtube.com - desktop site instead of mobile site - <!-- @browser: Chrome 89.0.4389 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/68920 --> **URL**: https://www.youtube.com/watch?v=u0KQqOai-uA **Browser / Version**: Chrome 89.0.4389 **Operating System**: Windows 10 **Tested Another Browser**: Yes Safari **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
desktop site instead of mobile site url browser version chrome operating system windows tested another browser yes safari problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce sqwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww browser configuration none from with ❤️
1
368,313
10,877,474,279
IssuesEvent
2019-11-16 10:27:27
trustwallet/blockatlas
https://api.github.com/repos/trustwallet/blockatlas
closed
reformat response for GET {chain}/staking/delegations/{addr} endpoint
Priority: High Task Size: S Type: Enhancement
Make response consistent with POST /staking/delegations/ endpoint
1.0
reformat response for GET {chain}/staking/delegations/{addr} endpoint - Make response consistent with POST /staking/delegations/ endpoint
priority
reformat response for get chain staking delegations addr endpoint make response consistent with post staking delegations endpoint
1
556,516
16,485,213,813
IssuesEvent
2021-05-24 16:55:46
sopra-fs21-group-11/sopra-client
https://api.github.com/repos/sopra-fs21-group-11/sopra-client
closed
S16: As a player I want to configure the game before it starts in order to get the most suitable game setup.
high priority user story
Before the player starts a game he must be able to: - [ ] edit the guess countdown before the evaluation. - [ ] edit the number of cards before an evaluation happens. - [ ] Edit the number of evaluations. - [ ] Edit the countdown for doubting a placement. - [ ] Edit the maximal number of players participating - [ ] Edit, how card is visible after it gets involved in a doubting process. - [ ] Edit how long a players turn lasts until he loses a token (S8)
1.0
S16: As a player I want to configure the game before it starts in order to get the most suitable game setup. - Before the player starts a game he must be able to: - [ ] edit the guess countdown before the evaluation. - [ ] edit the number of cards before an evaluation happens. - [ ] Edit the number of evaluations. - [ ] Edit the countdown for doubting a placement. - [ ] Edit the maximal number of players participating - [ ] Edit, how card is visible after it gets involved in a doubting process. - [ ] Edit how long a players turn lasts until he loses a token (S8)
priority
as a player i want to configure the game before it starts in order to get the most suitable game setup before the player starts a game he must be able to edit the guess countdown before the evaluation edit the number of cards before an evaluation happens edit the number of evaluations edit the countdown for doubting a placement edit the maximal number of players participating edit how card is visible after it gets involved in a doubting process edit how long a players turn lasts until he loses a token
1
271,621
8,485,895,492
IssuesEvent
2018-10-26 09:14:26
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
Report the reason why a workflow failed to be acquired
High Priority
When we hit a terminal exception in WorkQueue, we should somehow propagate this information up to CompOps. Maybe in the ReqMgr2 template where it has the state transitions and etc.
1.0
Report the reason why a workflow failed to be acquired - When we hit a terminal exception in WorkQueue, we should somehow propagate this information up to CompOps. Maybe in the ReqMgr2 template where it has the state transitions and etc.
priority
report the reason why a workflow failed to be acquired when we hit a terminal exception in workqueue we should somehow propagate this information up to compops maybe in the template where it has the state transitions and etc
1
639,089
20,746,471,634
IssuesEvent
2022-03-14 23:54:45
SSWConsulting/SSW.SophieBot
https://api.github.com/repos/SSWConsulting/SSW.SophieBot
closed
⭐ Query - When is John Smith's next availability
Priority: high
<!-- These comments automatically delete --> <!-- **Tip:** Delete parts that are not relevant --> <!-- Next to Cc:, @ mention users who should be in the loop --> Cc: @theixdev @AttackOnMorty <!-- add intended user next to **Hi** --> ### Pain <!-- Explain the pain you are experiencing --> People want to know when is someone available ### Tasks (only do the easy ones) <!--Add GitHub tasks--> - [x] When is John Smith's next availability - [x] When is John next available - [x] When will John next be free - [x] When is John next Free/Available - [x] When will John finish his current booking - [x] When is John's next bookable ### Hard Tasks - [x] When is John next free for 3 days? Feedback from @theixdev Thanks!
1.0
⭐ Query - When is John Smith's next availability - <!-- These comments automatically delete --> <!-- **Tip:** Delete parts that are not relevant --> <!-- Next to Cc:, @ mention users who should be in the loop --> Cc: @theixdev @AttackOnMorty <!-- add intended user next to **Hi** --> ### Pain <!-- Explain the pain you are experiencing --> People want to know when is someone available ### Tasks (only do the easy ones) <!--Add GitHub tasks--> - [x] When is John Smith's next availability - [x] When is John next available - [x] When will John next be free - [x] When is John next Free/Available - [x] When will John finish his current booking - [x] When is John's next bookable ### Hard Tasks - [x] When is John next free for 3 days? Feedback from @theixdev Thanks!
priority
⭐ query when is john smith s next availability cc theixdev attackonmorty pain people want to know when is someone available tasks only do the easy ones when is john smith s next availability when is john next available when will john next be free when is john next free available when will john finish his current booking when is john s next bookable hard tasks when is john next free for days feedback from theixdev thanks
1
498,880
14,434,965,239
IssuesEvent
2020-12-07 08:01:19
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
The "amphtml" is not generating on the custom taxonomy and the custom post types
NEXT UPDATE [Priority: HIGH] bug
Ref: https://secure.helpscout.net/conversation/1331694976/164006?folderId=2770545 The amphtml is not generating for the custom taxonomy and the custom post types after the 1.0.63 version update. Custom post type: https://demo6.petclassifieds.com/ad/1-year-old-basset-hound/ Custom taxonomy: https://demo6.petclassifieds.com/ad_category/dogs/ The user is creating the post types and the custom taxonomy using the Adforest theme.
1.0
The "amphtml" is not generating on the custom taxonomy and the custom post types - Ref: https://secure.helpscout.net/conversation/1331694976/164006?folderId=2770545 The amphtml is not generating for the custom taxonomy and the custom post types after the 1.0.63 version update. Custom post type: https://demo6.petclassifieds.com/ad/1-year-old-basset-hound/ Custom taxonomy: https://demo6.petclassifieds.com/ad_category/dogs/ The user is creating the post types and the custom taxonomy using the Adforest theme.
priority
the amphtml is not generating on the custom taxonomy and the custom post types ref the amphtml is not generating for the custom taxonomy and the custom post types after the version update custom post type custom taxonomy the user is creating the post types and the custom taxonomy using the adforest theme
1
509,292
14,727,737,864
IssuesEvent
2021-01-06 08:59:15
MikeVedsted/JoinMe
https://api.github.com/repos/MikeVedsted/JoinMe
opened
[FEAT] Add search event functionality in homepage
Priority: High :bomb: Status: Received :inbox_tray: Status: Selected for development :pencil: Type: Enhancement :rocket:
**🚀 I'm ready for take off** Before submitting, please mark if you: - [ ] Checked that this feature doesn't already exists - [ ] Checked that a feature request doesn't already exists - [ ] Went through the user flow, and understand the impact - [ ] Made sure the request shows why it is important to users but doesn't exaggerate the value
1.0
[FEAT] Add search event functionality in homepage - **🚀 I'm ready for take off** Before submitting, please mark if you: - [ ] Checked that this feature doesn't already exists - [ ] Checked that a feature request doesn't already exists - [ ] Went through the user flow, and understand the impact - [ ] Made sure the request shows why it is important to users but doesn't exaggerate the value
priority
add search event functionality in homepage 🚀 i m ready for take off before submitting please mark if you checked that this feature doesn t already exists checked that a feature request doesn t already exists went through the user flow and understand the impact made sure the request shows why it is important to users but doesn t exaggerate the value
1
689,352
23,617,802,756
IssuesEvent
2022-08-24 17:27:50
Unity-Technologies/com.unity.netcode.gameobjects
https://api.github.com/repos/Unity-Technologies/com.unity.netcode.gameobjects
closed
When a client connects to a server all clients receive all NetworkVariables
type:bug priority:high stat:imported
Thanks @mraue for the update. I get your issue now. Please let me update it and re-open as it seems pretty important. ### Description Every time a client connects to a server all `NetworkVariable`s on all `NetworkObject`s get synced to **all** clients. ### Reproduce Steps 1. Create a NetworkObject with a NetworkVariable. Add a log that tracks `OnValueChanged` 2. Start a server and connect a client 3. Connect a second client 4. The first and the second client show the `OnValueChanged` log ### Actual Outcome All connected clients receive a `NetworkVariable` update. ### Expected Outcome Only the second client connecting should receive the `NetworkVariable` data. ### Additional Context During the `internal void SendSpawnCallForObject(ulong clientId, NetworkObject networkObject)` for the NetworkObject to spawn, we mark: `networkObject.MarkVariablesDirty(true);` This causes the NetworkVariable to be considered dirty for all clients. As a result, the NetworkVariable gets serialized to all clients. The difficulty lies with the fact that current Netcode for GameObjects doesn't have a way to mark dirtiness for some clients only. This will require a bit of refactor.
1.0
When a client connects to a server all clients receive all NetworkVariables - Thanks @mraue for the update. I get your issue now. Please let me update it and re-open as it seems pretty important. ### Description Every time a client connects to a server all `NetworkVariable`s on all `NetworkObject`s get synced to **all** clients. ### Reproduce Steps 1. Create a NetworkObject with a NetworkVariable. Add a log that tracks `OnValueChanged` 2. Start a server and connect a client 3. Connect a second client 4. The first and the second client show the `OnValueChanged` log ### Actual Outcome All connected clients receive a `NetworkVariable` update. ### Expected Outcome Only the second client connecting should receive the `NetworkVariable` data. ### Additional Context During the `internal void SendSpawnCallForObject(ulong clientId, NetworkObject networkObject)` for the NetworkObject to spawn, we mark: `networkObject.MarkVariablesDirty(true);` This causes the NetworkVariable to be considered dirty for all clients. As a result, the NetworkVariable gets serialized to all clients. The difficulty lies with the fact that current Netcode for GameObjects doesn't have a way to mark dirtiness for some clients only. This will require a bit of refactor.
priority
when a client connects to a server all clients receive all networkvariables thanks mraue for the update i get your issue now please let me update it and re open as it seems pretty important description every time a client connects to a server all networkvariable s on all networkobject s get synced to all clients reproduce steps create a networkobject with a networkvariable add a log that tracks onvaluechanged start a server and connect a client connect a second client the first and the second client show the onvaluechanged log actual outcome all connected clients receive a networkvariable update expected outcome only the second client connecting should receive the networkvariable data additional context during the internal void sendspawncallforobject ulong clientid networkobject networkobject for the networkobject to spawn we mark networkobject markvariablesdirty true this causes the networkvariable to be considered dirty for all clients as a result the networkvariable gets serialized to all clients the difficulty lies with the fact that current netcode for gameobjects doesn t have a way to mark dirtiness for some clients only this will require a bit of refactor
1
541,259
15,823,859,508
IssuesEvent
2021-04-06 01:51:00
istopwg/ippsample
https://api.github.com/repos/istopwg/ippsample
reopened
ipptool: Add directive to monitor "printer-state" or "printer-state-reasons" in the background during a test
enhancement priority-high
Recent issues reported against the IPP Everywhere Self Certification Test I-20 and I-20.1 where I-20 deadlocks because PWG Raster is handled as a streaming document format, which prevents I-20.1 from checking the "printer-state-reasons", causes a condition where the tests cannot complete successfully. What is needed is a way for the tests to be run in parallel. But we don't want to create a complex general purpose parallel testing facility - that could create a monster. However, from discussions on the IPP WG reflector and in recent teleconferences, it was discussed that, if "ipptool" were able to monitor "printer-state-reasons" in a background thread while performing the test in the foreground thread, this would be useful for this purpose and for monitoring the completion of the action triggered by the operation. Add to ipptool support for a new "MONITOR-PRINTER-STATE" directive like so: MONITOR-PRINTER-STATE predicate [timeout] where ipptool would perform a polling Get-Printer-Attributes operation looking for the matching Printer state and state reason conditions defined by "predicate", and "timeout" would specify a hard time limit after which the MONITOR-PRINTER-STATE would fail if the matching status wasn't achieved. (We can noodle on the design of this...) Also define a MONITOR-JOB-STATE that could be used to monitor a Job state for completion: MONITOR-JOB-STATE jobid predicate [timeout] The logic would be basically the same but would poll the Job specified by jobid with Get-Job-Attributes until the matching conditions were achieved or the timeout occurred.
1.0
ipptool: Add directive to monitor "printer-state" or "printer-state-reasons" in the background during a test - Recent issues reported against the IPP Everywhere Self Certification Test I-20 and I-20.1 where I-20 deadlocks because PWG Raster is handled as a streaming document format, which prevents I-20.1 from checking the "printer-state-reasons", causes a condition where the tests cannot complete successfully. What is needed is a way for the tests to be run in parallel. But we don't want to create a complex general purpose parallel testing facility - that could create a monster. However, from discussions on the IPP WG reflector and in recent teleconferences, it was discussed that, if "ipptool" were able to monitor "printer-state-reasons" in a background thread while performing the test in the foreground thread, this would be useful for this purpose and for monitoring the completion of the action triggered by the operation. Add to ipptool support for a new "MONITOR-PRINTER-STATE" directive like so: MONITOR-PRINTER-STATE predicate [timeout] where ipptool would perform a polling Get-Printer-Attributes operation looking for the matching Printer state and state reason conditions defined by "predicate", and "timeout" would specify a hard time limit after which the MONITOR-PRINTER-STATE would fail if the matching status wasn't achieved. (We can noodle on the design of this...) Also define a MONITOR-JOB-STATE that could be used to monitor a Job state for completion: MONITOR-JOB-STATE jobid predicate [timeout] The logic would be basically the same but would poll the Job specified by jobid with Get-Job-Attributes until the matching conditions were achieved or the timeout occurred.
priority
ipptool add directive to monitor printer state or printer state reasons in the background during a test recent issues reported against the ipp everywhere self certification test i and i where i deadlocks because pwg raster is handled as a streaming document format which prevents i from checking the printer state reasons causes a condition where the tests cannot complete successfully what is needed is a way for the tests to be run in parallel but we don t want to create a complex general purpose parallel testing facility that could create a monster however from discussions on the ipp wg reflector and in recent teleconferences it was discussed that if ipptool were able to monitor printer state reasons in a background thread while performing the test in the foreground thread this would be useful for this purpose and for monitoring the completion of the action triggered by the operation add to ipptool support for a new monitor printer state directive like so monitor printer state predicate where ipptool would perform a polling get printer attributes operation looking for the matching printer state and state reason conditions defined by predicate and timeout would specify a hard time limit after which the monitor printer state would fail if the matching status wasn t achieved we can noodle on the design of this also define a monitor job state that could be used to monitor a job state for completion monitor job state jobid predicate the logic would be basically the same but would poll the job specified by jobid with get job attributes until the matching conditions were achieved or the timeout occurred
1
620,606
19,565,670,566
IssuesEvent
2022-01-03 23:39:35
bounswe/2021SpringGroup12
https://api.github.com/repos/bounswe/2021SpringGroup12
closed
Search a goal in an exact way
priority: high android
**Description** - Users should be able to see the goals that are published that includes the searched query inside the title, desription or tags
1.0
Search a goal in an exact way - **Description** - Users should be able to see the goals that are published that includes the searched query inside the title, desription or tags
priority
search a goal in an exact way description users should be able to see the goals that are published that includes the searched query inside the title desription or tags
1
376,263
11,140,471,143
IssuesEvent
2019-12-21 14:39:45
wilcommerce/Wilcommerce.Catalog
https://api.github.com/repos/wilcommerce/Wilcommerce.Catalog
closed
Change model description throws an error if description is empty
priority:high status:done type:bug
The method _ChangeDescription_ in the _Brand_, _Category_ and _Product_ entities throws an _ArgumentNullException_ if the description argument is null or empty. This behavior doesn't allow to set an empty description for the specified entities.
1.0
Change model description throws an error if description is empty - The method _ChangeDescription_ in the _Brand_, _Category_ and _Product_ entities throws an _ArgumentNullException_ if the description argument is null or empty. This behavior doesn't allow to set an empty description for the specified entities.
priority
change model description throws an error if description is empty the method changedescription in the brand category and product entities throws an argumentnullexception if the description argument is null or empty this behavior doesn t allow to set an empty description for the specified entities
1
311,622
9,536,674,501
IssuesEvent
2019-04-30 10:21:13
aartiukh/sph
https://api.github.com/repos/aartiukh/sph
opened
Fix unit tests in sph library
area: c++ priority: high type: bug
**Describe the bug** Currently unit tests are disabled in `sph` CMake file https://github.com/aartiukh/sph/blob/develop/sph/CMakeLists.txt#L22 **To Reproduce** Steps to reproduce the behavior: 1. Build `sph-sdk` 2. Run tests 3. Ensure that `sph` tests are not running **Expected behavior** The `sph` tests should work.
1.0
Fix unit tests in sph library - **Describe the bug** Currently unit tests are disabled in `sph` CMake file https://github.com/aartiukh/sph/blob/develop/sph/CMakeLists.txt#L22 **To Reproduce** Steps to reproduce the behavior: 1. Build `sph-sdk` 2. Run tests 3. Ensure that `sph` tests are not running **Expected behavior** The `sph` tests should work.
priority
fix unit tests in sph library describe the bug currently unit tests are disabled in sph cmake file to reproduce steps to reproduce the behavior build sph sdk run tests ensure that sph tests are not running expected behavior the sph tests should work
1
222,403
7,431,931,558
IssuesEvent
2018-03-25 19:32:02
smit-happens/YCP_EVOS
https://api.github.com/repos/smit-happens/YCP_EVOS
closed
Handle Precharge done signal
priority-high size-small stage unitek
<!-- Issue template To Use this Template: * Fill out what you can * Delete what you do not fill out --> ## End goal Transition to energized stage when precharge done line is enabled by the Unitek logic board
1.0
Handle Precharge done signal - <!-- Issue template To Use this Template: * Fill out what you can * Delete what you do not fill out --> ## End goal Transition to energized stage when precharge done line is enabled by the Unitek logic board
priority
handle precharge done signal issue template to use this template fill out what you can delete what you do not fill out end goal transition to energized stage when precharge done line is enabled by the unitek logic board
1
50,922
3,008,142,753
IssuesEvent
2015-07-27 19:41:12
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
PeopleInSpace: decrease or eliminate z-index
Bug Low-Hanging Fruit Priority: High
setting the z-index so high conflicts with other answerbar elements. set it low, like 2, or find a better way to line up the flags without using z-index. ------ http://duck.co/ia/view/people_in_space
1.0
PeopleInSpace: decrease or eliminate z-index - setting the z-index so high conflicts with other answerbar elements. set it low, like 2, or find a better way to line up the flags without using z-index. ------ http://duck.co/ia/view/people_in_space
priority
peopleinspace decrease or eliminate z index setting the z index so high conflicts with other answerbar elements set it low like or find a better way to line up the flags without using z index
1
257,279
8,135,296,338
IssuesEvent
2018-08-20 01:53:03
collinbarrett/FilterLists
https://api.github.com/repos/collinbarrett/FilterLists
opened
fix SnapshotService algorithm trying to insert duplicate raw rules
agent bug high priority
Since adding a unique index on rules.Raw in #353 , I can now see exceptions like below from the [SnapshotService](https://github.com/collinbarrett/FilterLists/tree/master/src/FilterLists.Services/Snapshot) that represent a flaw in the algorithm. It should not try to insert duplicate raw rules, but it is fairly often for some reason. Example: > FilterListId: 297 Exception: An error occurred while updating the entries. See the inner exception for details. at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.ReaderModificationCommandBatch.ExecuteAsync(IRelationalConnection connection, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.Internal.BatchExecutor.ExecuteAsync(DbContext _, ValueTuple`2 parameters, CancellationToken cancellationToken) at Pomelo.EntityFrameworkCore.MySql.Storage.Internal.MySqlExecutionStrategy.ExecuteAsync[TState,TResult](TState state, Func`4 operation, Func`4 verifySucceeded, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(IReadOnlyList`1 entriesToSave, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.DbContext.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at FilterLists.Services.Snapshot.SnapshotBatch.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/SnapshotBatch.cs:line 32 at FilterLists.Services.Snapshot.Snapshot.SaveBatches(IEnumerable`1 batches) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 122 at FilterLists.Services.Snapshot.Snapshot.SaveInBatches(IEnumerable`1 lines) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 114 at FilterLists.Services.Snapshot.Snapshot.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 65 at FilterLists.Services.Snapshot.Snapshot.TrySaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 43 Duplicate entry '###Meebo\:AdElement\.Root' for key 'IX_rules_Raw' at MySql.Data.MySqlClient.MySqlDataReader.ActivateResultSet(ResultSet resultSet) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 93 at MySql.Data.MySqlClient.MySqlDataReader.NextResultAsync(IOBehavior ioBehavior, CancellationToken cancellationToken) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 62 at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken)
1.0
fix SnapshotService algorithm trying to insert duplicate raw rules - Since adding a unique index on rules.Raw in #353 , I can now see exceptions like below from the [SnapshotService](https://github.com/collinbarrett/FilterLists/tree/master/src/FilterLists.Services/Snapshot) that represent a flaw in the algorithm. It should not try to insert duplicate raw rules, but it is fairly often for some reason. Example: > FilterListId: 297 Exception: An error occurred while updating the entries. See the inner exception for details. at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.ReaderModificationCommandBatch.ExecuteAsync(IRelationalConnection connection, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.Update.Internal.BatchExecutor.ExecuteAsync(DbContext _, ValueTuple`2 parameters, CancellationToken cancellationToken) at Pomelo.EntityFrameworkCore.MySql.Storage.Internal.MySqlExecutionStrategy.ExecuteAsync[TState,TResult](TState state, Func`4 operation, Func`4 verifySucceeded, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(IReadOnlyList`1 entriesToSave, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.ChangeTracking.Internal.StateManager.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at Microsoft.EntityFrameworkCore.DbContext.SaveChangesAsync(Boolean acceptAllChangesOnSuccess, CancellationToken cancellationToken) at FilterLists.Services.Snapshot.SnapshotBatch.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/SnapshotBatch.cs:line 32 at FilterLists.Services.Snapshot.Snapshot.SaveBatches(IEnumerable`1 batches) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 122 at FilterLists.Services.Snapshot.Snapshot.SaveInBatches(IEnumerable`1 lines) in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 114 at FilterLists.Services.Snapshot.Snapshot.SaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 65 at FilterLists.Services.Snapshot.Snapshot.TrySaveAsync() in /home/travis/build/collinbarrett/FilterLists/src/FilterLists.Services/Snapshot/Snapshot.cs:line 43 Duplicate entry '###Meebo\:AdElement\.Root' for key 'IX_rules_Raw' at MySql.Data.MySqlClient.MySqlDataReader.ActivateResultSet(ResultSet resultSet) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 93 at MySql.Data.MySqlClient.MySqlDataReader.NextResultAsync(IOBehavior ioBehavior, CancellationToken cancellationToken) in C:\projects\mysqlconnector\src\MySqlConnector\MySql.Data.MySqlClient\MySqlDataReader.cs:line 62 at Microsoft.EntityFrameworkCore.Update.AffectedCountModificationCommandBatch.ConsumeAsync(RelationalDataReader reader, CancellationToken cancellationToken)
priority
fix snapshotservice algorithm trying to insert duplicate raw rules since adding a unique index on rules raw in i can now see exceptions like below from the that represent a flaw in the algorithm it should not try to insert duplicate raw rules but it is fairly often for some reason example filterlistid exception an error occurred while updating the entries see the inner exception for details at microsoft entityframeworkcore update affectedcountmodificationcommandbatch consumeasync relationaldatareader reader cancellationtoken cancellationtoken at microsoft entityframeworkcore update readermodificationcommandbatch executeasync irelationalconnection connection cancellationtoken cancellationtoken at microsoft entityframeworkcore update internal batchexecutor executeasync dbcontext valuetuple parameters cancellationtoken cancellationtoken at pomelo entityframeworkcore mysql storage internal mysqlexecutionstrategy executeasync tstate state func operation func verifysucceeded cancellationtoken cancellationtoken at microsoft entityframeworkcore changetracking internal statemanager savechangesasync ireadonlylist entriestosave cancellationtoken cancellationtoken at microsoft entityframeworkcore changetracking internal statemanager savechangesasync boolean acceptallchangesonsuccess cancellationtoken cancellationtoken at microsoft entityframeworkcore dbcontext savechangesasync boolean acceptallchangesonsuccess cancellationtoken cancellationtoken at filterlists services snapshot snapshotbatch saveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshotbatch cs line at filterlists services snapshot snapshot savebatches ienumerable batches in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot saveinbatches ienumerable lines in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot saveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line at filterlists services snapshot snapshot trysaveasync in home travis build collinbarrett filterlists src filterlists services snapshot snapshot cs line duplicate entry meebo adelement root for key ix rules raw at mysql data mysqlclient mysqldatareader activateresultset resultset resultset in c projects mysqlconnector src mysqlconnector mysql data mysqlclient mysqldatareader cs line at mysql data mysqlclient mysqldatareader nextresultasync iobehavior iobehavior cancellationtoken cancellationtoken in c projects mysqlconnector src mysqlconnector mysql data mysqlclient mysqldatareader cs line at microsoft entityframeworkcore update affectedcountmodificationcommandbatch consumeasync relationaldatareader reader cancellationtoken cancellationtoken
1
283,504
8,719,731,675
IssuesEvent
2018-12-08 03:45:21
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
build_visit2_7_0 fails building python with icc.
bug likelihood medium priority reviewed severity high
Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line: env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \ console nothirdparty novisit \ mesa cmake python vtk qt icet pyside \ hdf5 silo szip \ makeflags -j4 Here is the Python failure: icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC Ibuild/temp.linuxx86_64-2.7/libffi/include Ibuild/temp.linuxx86_64-2.7/libffi I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include I. IInclude I./Include I/usr/local/include I/nfs/tmp2/brugger/build_python/Python2.7.5/Include I/nfs/tmp2/brugger/build_python/Python2.7.5 -c /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c -o build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o Wall fexceptions /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56): error: identifier "__m128" is undefined UINT128 i128; ^ compilation aborted for /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c (code 2) Python build finished, but the necessary bits to build these modules were not found: bsddb185 dl imageop sunaudiodev To find the necessary bits, look in setup.py in detect_modules() for the module's name. Failed to build these modules: _ctypes running build_scripts running install_lib Here is the Qt failure: icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259 DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include I../../../include/QtCore I../../../include/QtXml -o .obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp rm -f libbootstrap.a xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o .obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o .obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o .obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o .obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o .obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o .obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o .obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o .obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o .obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o .obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o .obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o .obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o .obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o .obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o .obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o .obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o .obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o .obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o .obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o .obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o .obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o .obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o .obj/release-static/qfilesystemiterator_unix.o .obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o make[1]: xiar: Command not found make[1]: ***** [libbootstrap.a] Error 127 make[1]: Leaving directory `/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap' make: ***** [subtoolsbootstrapmake_defaultordered] Error 2 -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1651 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: build_visit2_7_0 fails building python with icc. Assigned to: Kathleen Biagas Category: Target version: 2.7.1 Author: Eric Brugger Start: 11/08/2013 Due date: % Done: 0 Estimated time: Created: 11/08/2013 05:32 pm Updated: 12/11/2013 05:33 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: 2.7.0 Impact: Expected Use: OS: All Support Group: Any Description: Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line: env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \ --console --no-thirdparty --no-visit \ --mesa --cmake --python --vtk --qt --icet --pyside \ --hdf5 --silo --szip \ --makeflags -j4 Here is the Python failure: icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC -Ibuild/temp.linux-x86_64-2.7/libffi/include -Ibuild/temp.linux-x86_64-2.7/libffi -I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src -I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include -I. -IInclude -I./Include -I/usr/local/include -I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include -I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c -o build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o -Wall -fexceptions /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56): error: identifier "__m128" is undefined UINT128 i128; ^ compilation aborted for /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c (code 2) Python build finished, but the necessary bits to build these modules were not found: bsddb185 dl imageop sunaudiodev To find the necessary bits, look in setup.py in detect_modules() for the module's name. Failed to build these modules: _ctypes running build_scripts running install_lib Here is the Qt failure: icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing -fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259 -DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII -DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT -DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE -DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES -DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE -D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include -I../../../include/QtCore -I../../../include/QtXml -o .obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp rm -f libbootstrap.a xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o .obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o .obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o .obj/release-static/qutfcodec.o .obj/release-static/qglobal.o .obj/release-static/qmalloc.o .obj/release-static/qnumeric.o .obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o .obj/release-static/qdatastream.o .obj/release-static/qdir.o .obj/release-static/qdiriterator.o .obj/release-static/qfile.o .obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o .obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o .obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o .obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o .obj/release-static/qmetatype.o .obj/release-static/qvariant.o .obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o .obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o .obj/release-static/qdatetime.o .obj/release-static/qhash.o .obj/release-static/qlist.o .obj/release-static/qlocale.o .obj/release-static/qlocale_tools.o .obj/release-static/qmap.o .obj/release-static/qregexp.o .obj/release-static/qstring.o .obj/release-static/qstringlist.o .obj/release-static/qvector.o .obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o .obj/release-static/qxmlstream.o .obj/release-static/qdom.o .obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o .obj/release-static/qfilesystemiterator_unix.o .obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o make[1]: xiar: Command not found make[1]: *** [libbootstrap.a] Error 127 make[1]: Leaving directory `/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap' make: *** [sub-tools-bootstrap-make_default-ordered] Error 2 Comments: Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT). 2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
1.0
build_visit2_7_0 fails building python with icc. - Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line: env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \ console nothirdparty novisit \ mesa cmake python vtk qt icet pyside \ hdf5 silo szip \ makeflags -j4 Here is the Python failure: icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC Ibuild/temp.linuxx86_64-2.7/libffi/include Ibuild/temp.linuxx86_64-2.7/libffi I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include I. IInclude I./Include I/usr/local/include I/nfs/tmp2/brugger/build_python/Python2.7.5/Include I/nfs/tmp2/brugger/build_python/Python2.7.5 -c /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c -o build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o Wall fexceptions /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56): error: identifier "__m128" is undefined UINT128 i128; ^ compilation aborted for /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c (code 2) Python build finished, but the necessary bits to build these modules were not found: bsddb185 dl imageop sunaudiodev To find the necessary bits, look in setup.py in detect_modules() for the module's name. Failed to build these modules: _ctypes running build_scripts running install_lib Here is the Qt failure: icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259 DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include I../../../include/QtCore I../../../include/QtXml -o .obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp rm -f libbootstrap.a xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o .obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o .obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o .obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o .obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o .obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o .obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o .obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o .obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o .obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o .obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o .obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o .obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o .obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o .obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o .obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o .obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o .obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o .obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o .obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o .obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o .obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o .obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o .obj/release-static/qfilesystemiterator_unix.o .obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o make[1]: xiar: Command not found make[1]: ***** [libbootstrap.a] Error 127 make[1]: Leaving directory `/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap' make: ***** [subtoolsbootstrapmake_defaultordered] Error 2 -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1651 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: build_visit2_7_0 fails building python with icc. Assigned to: Kathleen Biagas Category: Target version: 2.7.1 Author: Eric Brugger Start: 11/08/2013 Due date: % Done: 0 Estimated time: Created: 11/08/2013 05:32 pm Updated: 12/11/2013 05:33 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: 2.7.0 Impact: Expected Use: OS: All Support Group: Any Description: Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line: env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \ --console --no-thirdparty --no-visit \ --mesa --cmake --python --vtk --qt --icet --pyside \ --hdf5 --silo --szip \ --makeflags -j4 Here is the Python failure: icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC -Ibuild/temp.linux-x86_64-2.7/libffi/include -Ibuild/temp.linux-x86_64-2.7/libffi -I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src -I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include -I. -IInclude -I./Include -I/usr/local/include -I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include -I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c -o build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o -Wall -fexceptions /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56): error: identifier "__m128" is undefined UINT128 i128; ^ compilation aborted for /nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c (code 2) Python build finished, but the necessary bits to build these modules were not found: bsddb185 dl imageop sunaudiodev To find the necessary bits, look in setup.py in detect_modules() for the module's name. Failed to build these modules: _ctypes running build_scripts running install_lib Here is the Qt failure: icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing -fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259 -DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII -DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT -DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE -DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES -DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE -D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include -I../../../include/QtCore -I../../../include/QtXml -o .obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp rm -f libbootstrap.a xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o .obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o .obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o .obj/release-static/qutfcodec.o .obj/release-static/qglobal.o .obj/release-static/qmalloc.o .obj/release-static/qnumeric.o .obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o .obj/release-static/qdatastream.o .obj/release-static/qdir.o .obj/release-static/qdiriterator.o .obj/release-static/qfile.o .obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o .obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o .obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o .obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o .obj/release-static/qmetatype.o .obj/release-static/qvariant.o .obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o .obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o .obj/release-static/qdatetime.o .obj/release-static/qhash.o .obj/release-static/qlist.o .obj/release-static/qlocale.o .obj/release-static/qlocale_tools.o .obj/release-static/qmap.o .obj/release-static/qregexp.o .obj/release-static/qstring.o .obj/release-static/qstringlist.o .obj/release-static/qvector.o .obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o .obj/release-static/qxmlstream.o .obj/release-static/qdom.o .obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o .obj/release-static/qfilesystemiterator_unix.o .obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o make[1]: xiar: Command not found make[1]: *** [libbootstrap.a] Error 127 make[1]: Leaving directory `/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap' make: *** [sub-tools-bootstrap-make_default-ordered] Error 2 Comments: Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT). 2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
priority
build fails building python with icc rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapichgnu include build console nothirdparty novisit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fnostrictaliasing g dndebug fpic fpic ibuild temp libffi include ibuild temp libffi i nfs brugger build python modules ctypes libffi src i nfs brugger build python visit python icc include i iinclude i include i usr local include i nfs brugger build python include i nfs brugger build python c nfs brugger build python python modules ctypes libffi src c o build temp nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falignfunctions ansialias fstrict aliasing fpic fnojumptables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj releasestatic qlatincodec o obj releasestatic qsimplecodec o obj releasestatic qtextcodec o obj releasestatic qtsciicodec o obj releasestatic qutfcodec o obj releasestatic qglobal o obj releasestatic qmalloc o obj releasestatic qnumeric o obj releasestatic qabstractfileengine o obj releasestatic qbuffer o obj releasestatic qdatastream o obj releasestatic qdir o obj releasestatic qdiriterator o obj releasestatic qfile o obj releasestatic qfileinfo o obj releasestatic qfilesystementry o obj releasestatic qfilesystemengine o obj releasestatic qfsfileengine o obj releasestatic qfsfileengine iterator o obj releasestatic qiodevice o obj releasestatic qtemporaryfile o obj releasestatic qtextstream o obj releasestatic qmetatype o obj releasestatic qvariant o obj releasestatic qsystemerror o obj releasestatic qbitarray o obj releasestatic qbytearray o obj releasestatic qbytearraymatcher o obj releasestatic qdatetime o obj releasestatic qhash o obj releasestatic qlist o obj releasestatic qlocale o obj releasestatic qlocale tools o obj releasestatic qmap o obj releasestatic qregexp o obj releasestatic qstring o obj releasestatic qstringlist o obj releasestatic qvector o obj releasestatic qvsnprintf o obj releasestatic qxmlutils o obj releasestatic qxmlstream o obj releasestatic qdom o obj releasestatic qxml o obj releasestatic qfilesystemengine unix o obj release static qfilesystemiterator unix o obj releasestatic qfsfileengine unix o obj releasestatic qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python src tools bootstrap make error redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject build fails building python with icc assigned to kathleen biagas category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version impact expected use os all support group any description rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapich gnu include build console no thirdparty no visit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fno strict aliasing g dndebug fpic fpic ibuild temp linux libffi include ibuild temp linux libffi i nfs brugger build python python modules ctypes libffi src i nfs brugger build python visit python linux icc include i iinclude i include i usr local include i nfs brugger build python python include i nfs brugger build python python c nfs brugger build python python modules ctypes libffi src c o build temp linux nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falign functions ansi alias fstrict aliasing fpic fno jump tables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs linux icc i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj release static qlatincodec o obj release static qsimplecodec o obj release static qtextcodec o obj release static qtsciicodec o obj release static qutfcodec o obj release static qglobal o obj release static qmalloc o obj release static qnumeric o obj release static qabstractfileengine o obj release static qbuffer o obj release static qdatastream o obj release static qdir o obj release static qdiriterator o obj release static qfile o obj release static qfileinfo o obj release static qfilesystementry o obj release static qfilesystemengine o obj release static qfsfileengine o obj release static qfsfileengine iterator o obj release static qiodevice o obj release static qtemporaryfile o obj release static qtextstream o obj release static qmetatype o obj release static qvariant o obj release static qsystemerror o obj release static qbitarray o obj release static qbytearray o obj release static qbytearraymatcher o obj release static qdatetime o obj release static qhash o obj release static qlist o obj release static qlocale o obj release static qlocale tools o obj release static qmap o obj release static qregexp o obj release static qstring o obj release static qstringlist o obj release static qvector o obj release static qvsnprintf o obj release static qxmlutils o obj release static qxmlstream o obj release static qdom o obj release static qxml o obj release static qfilesystemengine unix o obj release static qfilesystemiterator unix o obj release static qfsfileengine unix o obj release static qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python qt everywhere opensource src src tools bootstrap make error comments hi eric the python error may not be critical at least for now looks like qt is configuring the ar command to be xiar mkspecs linux icc qmake conf qmake ar xiar cqsa quick fix might be to set a simple alias xiar ar or modify the qmake conf file to what ever might be appropriate call to create archives for your system best regards hari the qt build issue eric encountered is due to the lc setup we need to issue a use icc before attempting to build as doing so makes available the full suite of intel compiler tools including xiar used by qt update merge to trunk minor python patch to bv python fixing issue when compiling with icc the patch is from a newer version of library
1
152,423
5,846,832,704
IssuesEvent
2017-05-10 17:03:57
hecoding/Pac-Man
https://api.github.com/repos/hecoding/Pac-Man
closed
Fantasmas siempre "starterghosts" en la partida que usa la GUI para mostrar ejecuciones del mejor individuo encontrado
bug GUI high priority
Fix asap
1.0
Fantasmas siempre "starterghosts" en la partida que usa la GUI para mostrar ejecuciones del mejor individuo encontrado - Fix asap
priority
fantasmas siempre starterghosts en la partida que usa la gui para mostrar ejecuciones del mejor individuo encontrado fix asap
1
381,248
11,275,454,480
IssuesEvent
2020-01-14 20:48:38
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode "wavelets" cause darktable to crash
bug: wip priority: high reproduce: confirmed understood: clear
<!-- IMPORTANT Bug reports that do not make an effort to help the developers will be closed without notice. Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed. A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information. --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode set to "wavelets" cause darktable to crash with "An unhandled exception occured" box showing up. The same behaviour observed when zoom is set to 1600% or 800% and "denoise (profiled)" module with the same mode is activated. **To Reproduce** <!-- Provide detailed steps that can reproduce the behavior, such as:--> 1. Go to 'Correction group' 2. Click on 'Denoise (profiled)' 3. Set 'mode' to 'wavelets' 5. Change zoom to 1600% either by clicking 1600% or scrolling mouse wheel with Ctr 4. See error Alternatively: 1. Go to 'Correction group' 2. Click on 'Denoise (profiled)' 3. Set 'mode' to 'wavelets' 4. Deactivate 'Denoise (profiled)' module 5. Change zoon to 1600% 6. Activate 'Denoise (profiled)' module 7. See error **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> Should not crash **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![Error](https://user-images.githubusercontent.com/59767780/72205556-be39ae00-3484-11ea-883c-f4948360642f.png) **Platform (please complete the following information):** - Darktable Version: <!-- [e.g. 2.6.0] --> darktable 3.0.0 - OS: <!-- [e.g. Windows 8.1, Gentoo Linux] --> Windows 10 - CPU: Intel(R) Xeon(R) CPU E5-1620 v3 @ 3.50GHz **Additional context** <!-- Add any other context about the problem here, for example: - Can you reproduce with another Darktable version? - Can you reproduce with a RAW or Jpeg or both? - Are the steps above reproduce with a fresh edit (removing history)? - Attach an XMP if this is necessary - Did you compile Darktable yourself? If so which compiler was used, with what options? - Is the issue still present using an empty/new config-dir --> This is observable either with previously edited and with new raws. Backtrace file is usually almost empty, I also add another bactrace file generated after the same error but with content. [darktable_bt_68B0D0.txt](https://github.com/darktable-org/darktable/files/4049412/darktable_bt_68B0D0.txt) [darktable_bt_6FB0D0.txt](https://github.com/darktable-org/darktable/files/4049413/darktable_bt_6FB0D0.txt)
1.0
Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode "wavelets" cause darktable to crash - <!-- IMPORTANT Bug reports that do not make an effort to help the developers will be closed without notice. Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed. A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information. --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Zooming in to 800% or 1600% when "denoise (profiled)" module is on with mode set to "wavelets" cause darktable to crash with "An unhandled exception occured" box showing up. The same behaviour observed when zoom is set to 1600% or 800% and "denoise (profiled)" module with the same mode is activated. **To Reproduce** <!-- Provide detailed steps that can reproduce the behavior, such as:--> 1. Go to 'Correction group' 2. Click on 'Denoise (profiled)' 3. Set 'mode' to 'wavelets' 5. Change zoom to 1600% either by clicking 1600% or scrolling mouse wheel with Ctr 4. See error Alternatively: 1. Go to 'Correction group' 2. Click on 'Denoise (profiled)' 3. Set 'mode' to 'wavelets' 4. Deactivate 'Denoise (profiled)' module 5. Change zoon to 1600% 6. Activate 'Denoise (profiled)' module 7. See error **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> Should not crash **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![Error](https://user-images.githubusercontent.com/59767780/72205556-be39ae00-3484-11ea-883c-f4948360642f.png) **Platform (please complete the following information):** - Darktable Version: <!-- [e.g. 2.6.0] --> darktable 3.0.0 - OS: <!-- [e.g. Windows 8.1, Gentoo Linux] --> Windows 10 - CPU: Intel(R) Xeon(R) CPU E5-1620 v3 @ 3.50GHz **Additional context** <!-- Add any other context about the problem here, for example: - Can you reproduce with another Darktable version? - Can you reproduce with a RAW or Jpeg or both? - Are the steps above reproduce with a fresh edit (removing history)? - Attach an XMP if this is necessary - Did you compile Darktable yourself? If so which compiler was used, with what options? - Is the issue still present using an empty/new config-dir --> This is observable either with previously edited and with new raws. Backtrace file is usually almost empty, I also add another bactrace file generated after the same error but with content. [darktable_bt_68B0D0.txt](https://github.com/darktable-org/darktable/files/4049412/darktable_bt_68B0D0.txt) [darktable_bt_6FB0D0.txt](https://github.com/darktable-org/darktable/files/4049413/darktable_bt_6FB0D0.txt)
priority
zooming in to or when denoise profiled module is on with mode wavelets cause darktable to crash important bug reports that do not make an effort to help the developers will be closed without notice make sure that this bug has not already been opened and or closed by searching the issues on github as duplicate bug reports will be closed a bug report simply stating that darktable crashes is unhelpful so please fill in most of the items below and provide detailed information describe the bug zooming in to or when denoise profiled module is on with mode set to wavelets cause darktable to crash with an unhandled exception occured box showing up the same behaviour observed when zoom is set to or and denoise profiled module with the same mode is activated to reproduce go to correction group click on denoise profiled set mode to wavelets change zoom to either by clicking or scrolling mouse wheel with ctr see error alternatively go to correction group click on denoise profiled set mode to wavelets deactivate denoise profiled module change zoon to activate denoise profiled module see error expected behavior should not crash screenshots platform please complete the following information darktable version darktable os windows cpu intel r xeon r cpu additional context add any other context about the problem here for example can you reproduce with another darktable version can you reproduce with a raw or jpeg or both are the steps above reproduce with a fresh edit removing history attach an xmp if this is necessary did you compile darktable yourself if so which compiler was used with what options is the issue still present using an empty new config dir this is observable either with previously edited and with new raws backtrace file is usually almost empty i also add another bactrace file generated after the same error but with content
1
444,415
12,811,911,588
IssuesEvent
2020-07-04 02:15:41
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
opened
JADC got the big blind twice due to new player motion code
bug high priority
odbphatmackakabig got moved from a table with 7 players to our table where we only had 5. There logically was only one place where he could go, which was between JADC, who had just paid the big blind and me, who had just paid the small blind. I should have gotten the button, which I did, odbphatmackakabig should have had to sit out (instead he paid the small blind) and JADC should have gotten the small blind. I think this just shows that my implementation of #247 was flawed. I'll look at it when I'm awake tomorrow, but probably after I've exercised. ``` Dealing Pot-Limit Binglaha deadhead blinds 2000 JADC blinds 4000 1 JADC 2 1850 4000 [ ] [ ] [ ] [ ] 2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ] 3 nnr13 0 24750 0 [ ] [ ] [ ] [ ] 4 B Meeeee2 2 11250 0 [ ] [ ] [ ] [ ] 5 deadhead 2 46000 2000 [ ] [ ] [ ] [ ] My cards are: [Kd] [6c] [8c] [8s] deadhead: gg tombayz folds nnr13: gg grizz nnr13 folds grizz: ty Meeeee2 folds It is 2000 to you deadhead calls JADC checks 1 JADC 2 1850 4000 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 > deadhead 2 44000 4000 [ ] [ ] [ ] [ ] Board: 5s 8d Tc My cards are: [Kd] [6c] [8c] [8s] It is 0 to you deadhead bets 4000 JADC asked for and gets 60 more seconds odbphatmackakabig has come from Table 277 JADC will be turning in in 2 lammers before the next hand JADC calls and is all-in Meeeee2 rolled a 4 1 > JADC 2 0 5850 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ] 6 [odbphatmackakabig] 1 5750[ 0] Board: 5s 8d Tc 7h My cards are: [Kd] [6c] [8c] [8s] 1 > JADC 2 0 5850 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ] 6 [odbphatmackakabig] 1 5750[ 0] Board: 5s 8d Tc 7h Th My cards are: [Kd] [6c] [8c] [8s] Uncalled 2150 returned to deadhead deadhead Kd 6c 8c 8s High Full House (Eights over Tens) 8c 8d 8s Th Tc Low none JADC 5d 7d 2h 9h High Two Pair (Tens and Sevens) Tc Th 7h 7d 9h Low none deadhead wins 11700 for High JADC redeemed a lammer for 5000 JADC redeemed a lammer for 5000 Dealing Pot-Limit Binglaha odbphatmackakabig blinds 2000 JADC blinds 4000 1 JADC 0 6000 4000 [ ] [ ] [ ] [ ] 2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ] 3 nnr13 0 24750 0 [ ] [ ] [ ] [ ] 4 Meeeee2 2 11250 0 [ ] [ ] [ ] [ ] 5 B deadhead 2 53850 0 [ ] [ ] [ ] [ ] 6 odbphatmackakabig 1 3750 2000 [ ] [ ] [ ] [ ] My cards are: [7s] [7h] [4d] [Ac] ```
1.0
JADC got the big blind twice due to new player motion code - odbphatmackakabig got moved from a table with 7 players to our table where we only had 5. There logically was only one place where he could go, which was between JADC, who had just paid the big blind and me, who had just paid the small blind. I should have gotten the button, which I did, odbphatmackakabig should have had to sit out (instead he paid the small blind) and JADC should have gotten the small blind. I think this just shows that my implementation of #247 was flawed. I'll look at it when I'm awake tomorrow, but probably after I've exercised. ``` Dealing Pot-Limit Binglaha deadhead blinds 2000 JADC blinds 4000 1 JADC 2 1850 4000 [ ] [ ] [ ] [ ] 2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ] 3 nnr13 0 24750 0 [ ] [ ] [ ] [ ] 4 B Meeeee2 2 11250 0 [ ] [ ] [ ] [ ] 5 deadhead 2 46000 2000 [ ] [ ] [ ] [ ] My cards are: [Kd] [6c] [8c] [8s] deadhead: gg tombayz folds nnr13: gg grizz nnr13 folds grizz: ty Meeeee2 folds It is 2000 to you deadhead calls JADC checks 1 JADC 2 1850 4000 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 > deadhead 2 44000 4000 [ ] [ ] [ ] [ ] Board: 5s 8d Tc My cards are: [Kd] [6c] [8c] [8s] It is 0 to you deadhead bets 4000 JADC asked for and gets 60 more seconds odbphatmackakabig has come from Table 277 JADC will be turning in in 2 lammers before the next hand JADC calls and is all-in Meeeee2 rolled a 4 1 > JADC 2 0 5850 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ] 6 [odbphatmackakabig] 1 5750[ 0] Board: 5s 8d Tc 7h My cards are: [Kd] [6c] [8c] [8s] 1 > JADC 2 0 5850 [ ] [ ] [ ] [ ] 2 [tombayz] 0 20150[ 0] 3 [nnr13] 0 24750[ 0] 4 B [Meeeee2] 2 11250[ 0] 5 deadhead 2 40000 8000 [ ] [ ] [ ] [ ] 6 [odbphatmackakabig] 1 5750[ 0] Board: 5s 8d Tc 7h Th My cards are: [Kd] [6c] [8c] [8s] Uncalled 2150 returned to deadhead deadhead Kd 6c 8c 8s High Full House (Eights over Tens) 8c 8d 8s Th Tc Low none JADC 5d 7d 2h 9h High Two Pair (Tens and Sevens) Tc Th 7h 7d 9h Low none deadhead wins 11700 for High JADC redeemed a lammer for 5000 JADC redeemed a lammer for 5000 Dealing Pot-Limit Binglaha odbphatmackakabig blinds 2000 JADC blinds 4000 1 JADC 0 6000 4000 [ ] [ ] [ ] [ ] 2 > tombayz 0 20150 0 [ ] [ ] [ ] [ ] 3 nnr13 0 24750 0 [ ] [ ] [ ] [ ] 4 Meeeee2 2 11250 0 [ ] [ ] [ ] [ ] 5 B deadhead 2 53850 0 [ ] [ ] [ ] [ ] 6 odbphatmackakabig 1 3750 2000 [ ] [ ] [ ] [ ] My cards are: [7s] [7h] [4d] [Ac] ```
priority
jadc got the big blind twice due to new player motion code odbphatmackakabig got moved from a table with players to our table where we only had there logically was only one place where he could go which was between jadc who had just paid the big blind and me who had just paid the small blind i should have gotten the button which i did odbphatmackakabig should have had to sit out instead he paid the small blind and jadc should have gotten the small blind i think this just shows that my implementation of was flawed i ll look at it when i m awake tomorrow but probably after i ve exercised dealing pot limit binglaha deadhead blinds jadc blinds jadc tombayz b deadhead my cards are deadhead gg tombayz folds gg grizz folds grizz ty folds it is to you deadhead calls jadc checks jadc b deadhead board tc my cards are it is to you deadhead bets jadc asked for and gets more seconds odbphatmackakabig has come from table jadc will be turning in in lammers before the next hand jadc calls and is all in rolled a jadc b deadhead board tc my cards are jadc b deadhead board tc th my cards are uncalled returned to deadhead deadhead kd high full house eights over tens th tc low none jadc high two pair tens and sevens tc th low none deadhead wins for high jadc redeemed a lammer for jadc redeemed a lammer for dealing pot limit binglaha odbphatmackakabig blinds jadc blinds jadc tombayz b deadhead odbphatmackakabig my cards are
1
73,286
3,410,538,691
IssuesEvent
2015-12-04 20:38:46
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
closed
Edit Dataset: Restricting a file on edit files tab results in TOA popup behind the white background.
Component: File Upload & Handling Component: UX & Upgrade Priority: High Status: QA Type: Bug
The behavior varies, depending on whether you restrict the file on creating the dataset or just editing the file metadata. It works if you restrict from the view dataset mode. 1. On create dataset: create a dataset, upload a file, restrict that file, save. The TOA popup is completely behind the whiteout background and you cannot interact with the popup. 2. On edit file metadata: On save, the TOA popup is half behind the whiteout background with the buttons visible and working and a > next to the cancel button.
1.0
Edit Dataset: Restricting a file on edit files tab results in TOA popup behind the white background. - The behavior varies, depending on whether you restrict the file on creating the dataset or just editing the file metadata. It works if you restrict from the view dataset mode. 1. On create dataset: create a dataset, upload a file, restrict that file, save. The TOA popup is completely behind the whiteout background and you cannot interact with the popup. 2. On edit file metadata: On save, the TOA popup is half behind the whiteout background with the buttons visible and working and a > next to the cancel button.
priority
edit dataset restricting a file on edit files tab results in toa popup behind the white background the behavior varies depending on whether you restrict the file on creating the dataset or just editing the file metadata it works if you restrict from the view dataset mode on create dataset create a dataset upload a file restrict that file save the toa popup is completely behind the whiteout background and you cannot interact with the popup on edit file metadata on save the toa popup is half behind the whiteout background with the buttons visible and working and a next to the cancel button
1
33,659
2,770,742,634
IssuesEvent
2015-05-01 16:44:39
biocore/qiita
https://api.github.com/repos/biocore/qiita
closed
Clean up instantiation of connection objects
priority: high
Once #1131 is merged, we should go through and clean up instances in the codebase of: ```python conn_handler = SQLConnectionHandler() if conn_handler is None else conn_handler ``` And the variants. Either the `conn_handler` should be passed in, or a new object created. Since the database connections are not reopened on instantiation (unless necessary), the overhead of just doing `SQLConnectionHandler()` will be minor cc @antgonza, who originally recommended opening this
1.0
Clean up instantiation of connection objects - Once #1131 is merged, we should go through and clean up instances in the codebase of: ```python conn_handler = SQLConnectionHandler() if conn_handler is None else conn_handler ``` And the variants. Either the `conn_handler` should be passed in, or a new object created. Since the database connections are not reopened on instantiation (unless necessary), the overhead of just doing `SQLConnectionHandler()` will be minor cc @antgonza, who originally recommended opening this
priority
clean up instantiation of connection objects once is merged we should go through and clean up instances in the codebase of python conn handler sqlconnectionhandler if conn handler is none else conn handler and the variants either the conn handler should be passed in or a new object created since the database connections are not reopened on instantiation unless necessary the overhead of just doing sqlconnectionhandler will be minor cc antgonza who originally recommended opening this
1
261,496
8,234,254,772
IssuesEvent
2018-09-08 12:15:12
Theophilix/event-table-edit
https://api.github.com/repos/Theophilix/event-table-edit
closed
Appointment tables: Bug with time / leading zero
bug high priority
When the user chooses a time with "10:00", like 08:00 - 10:00, it is displayed like this: "08:00 - 010:00". Outlook interprets the time like 01:00 (one o'clock!). This problem only appears, when 10:00 is the ending time. ![screenshot-2018-8-28 appointmenttable 1](https://user-images.githubusercontent.com/17642779/44725284-37757280-aad5-11e8-8872-5757d3e713d8.png) ![screenshot-2018-8-28 appointmenttable](https://user-images.githubusercontent.com/17642779/44725291-3ba19000-aad5-11e8-97ca-3a568422272b.png) ![capture_001_28082018_150634](https://user-images.githubusercontent.com/17642779/44725302-40feda80-aad5-11e8-9277-37f4c8dd578d.jpg)
1.0
Appointment tables: Bug with time / leading zero - When the user chooses a time with "10:00", like 08:00 - 10:00, it is displayed like this: "08:00 - 010:00". Outlook interprets the time like 01:00 (one o'clock!). This problem only appears, when 10:00 is the ending time. ![screenshot-2018-8-28 appointmenttable 1](https://user-images.githubusercontent.com/17642779/44725284-37757280-aad5-11e8-8872-5757d3e713d8.png) ![screenshot-2018-8-28 appointmenttable](https://user-images.githubusercontent.com/17642779/44725291-3ba19000-aad5-11e8-97ca-3a568422272b.png) ![capture_001_28082018_150634](https://user-images.githubusercontent.com/17642779/44725302-40feda80-aad5-11e8-9277-37f4c8dd578d.jpg)
priority
appointment tables bug with time leading zero when the user chooses a time with like it is displayed like this outlook interprets the time like one o clock this problem only appears when is the ending time
1
695,279
23,850,793,852
IssuesEvent
2022-09-06 17:43:37
guynir42/virtualobserver
https://api.github.com/repos/guynir42/virtualobserver
opened
Automatic naming for data files and keys
question high priority
When downloading data we need to save it inside files (often with internal keys, like HDF5 files). The naming should probably be automatic, and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object's data is saved. It should also conform to reasonable file size and number, e.g., stick 1000 lightcurves in each file.
1.0
Automatic naming for data files and keys - When downloading data we need to save it inside files (often with internal keys, like HDF5 files). The naming should probably be automatic, and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object's data is saved. It should also conform to reasonable file size and number, e.g., stick 1000 lightcurves in each file.
priority
automatic naming for data files and keys when downloading data we need to save it inside files often with internal keys like files the naming should probably be automatic and should be consistent enough so that it is easy to locate objects even if the database is corrupt and no longer keeps a record of where each object s data is saved it should also conform to reasonable file size and number e g stick lightcurves in each file
1
185,966
6,732,350,377
IssuesEvent
2017-10-18 11:08:25
DOAJ/doaj
https://api.github.com/repos/DOAJ/doaj
opened
DOAJ data download issue
bug high priority
Dear CL I received the following email and I will need your help: Hi Team, We are not able to download DOAJ records. We are getting error while downloading records. PFB error. Is there any server issue from your side? Please let us know. We are able to download some records but for most of records we are getting this error. PFA for sample downloaded file. Error:- <html> <head><title>503 Service Temporarily Unavailable</title></head> <body bgcolor="white"> <center><h1>503 Service Temporarily Unavailable</h1></center> <hr><center>nginx/1.4.6 (Ubuntu)</center> </body> </html> ;503 Steps to download data are as follows: Step 1 https://doaj.org/oai.article?verb=ListSets Step 2 FIND medicine https://doaj.org/oai.article?verb=ListIdentifiers&metadataPrefix=oai_dc&set=RE9BSjpJbnRlcm5hbCBtZWRpY2luZQ~~ Step 3 - resumption token https://doaj.org/oai.article?verb=ListIdentifiers&resumptionToken=eyJzIjogIlJFOUJTanBKYm5SbGNtNWhiQ0J0WldScFkybHVaUX5-IiwgIm0iOiAib2FpX2RjIiwgIm4iOiA2MDB9 Step 4 https://doaj.org/oai.article?verb=GetRecord&metadataPrefix=oai_dc&identifier=oai:doaj.org/article:fcb85b13cb3a44e6ab1f9dfa2fe605a6
1.0
DOAJ data download issue - Dear CL I received the following email and I will need your help: Hi Team, We are not able to download DOAJ records. We are getting error while downloading records. PFB error. Is there any server issue from your side? Please let us know. We are able to download some records but for most of records we are getting this error. PFA for sample downloaded file. Error:- <html> <head><title>503 Service Temporarily Unavailable</title></head> <body bgcolor="white"> <center><h1>503 Service Temporarily Unavailable</h1></center> <hr><center>nginx/1.4.6 (Ubuntu)</center> </body> </html> ;503 Steps to download data are as follows: Step 1 https://doaj.org/oai.article?verb=ListSets Step 2 FIND medicine https://doaj.org/oai.article?verb=ListIdentifiers&metadataPrefix=oai_dc&set=RE9BSjpJbnRlcm5hbCBtZWRpY2luZQ~~ Step 3 - resumption token https://doaj.org/oai.article?verb=ListIdentifiers&resumptionToken=eyJzIjogIlJFOUJTanBKYm5SbGNtNWhiQ0J0WldScFkybHVaUX5-IiwgIm0iOiAib2FpX2RjIiwgIm4iOiA2MDB9 Step 4 https://doaj.org/oai.article?verb=GetRecord&metadataPrefix=oai_dc&identifier=oai:doaj.org/article:fcb85b13cb3a44e6ab1f9dfa2fe605a6
priority
doaj data download issue dear cl i received the following email and i will need your help hi team we are not able to download doaj records we are getting error while downloading records pfb error is there any server issue from your side please let us know we are able to download some records but for most of records we are getting this error pfa for sample downloaded file error service temporarily unavailable service temporarily unavailable nginx ubuntu steps to download data are as follows step step find medicine step resumption token step
1
201,451
7,031,613,941
IssuesEvent
2017-12-26 19:11:37
qhacks/hacker-dashboard
https://api.github.com/repos/qhacks/hacker-dashboard
opened
Fix mongo issues (you know what they are ahah)
priority: blocker (high)
**Problem** Fix the database issues **Requirements** - [ ] Resolve data inconsistencies
1.0
Fix mongo issues (you know what they are ahah) - **Problem** Fix the database issues **Requirements** - [ ] Resolve data inconsistencies
priority
fix mongo issues you know what they are ahah problem fix the database issues requirements resolve data inconsistencies
1
799,331
28,304,440,442
IssuesEvent
2023-04-10 09:33:10
bounswe/bounswe2023group6
https://api.github.com/repos/bounswe/bounswe2023group6
closed
Filling the Project Plan Chart
priority: high type: task status: inprogress area: wiki area: milestone
### Problem Some of group members prepared well structured Project Plan Chart. It must be filled with related information, individually. ### Solution I'll fill the chart by reviewing all the issues I've contributed to so far. ### Documentation You can reach the solution [here](). ### Additional notes _No response_ ### Reviewers Ömer Talip Akalın ### Deadline 10.04.2023 - 20.00
1.0
Filling the Project Plan Chart - ### Problem Some of group members prepared well structured Project Plan Chart. It must be filled with related information, individually. ### Solution I'll fill the chart by reviewing all the issues I've contributed to so far. ### Documentation You can reach the solution [here](). ### Additional notes _No response_ ### Reviewers Ömer Talip Akalın ### Deadline 10.04.2023 - 20.00
priority
filling the project plan chart problem some of group members prepared well structured project plan chart it must be filled with related information individually solution i ll fill the chart by reviewing all the issues i ve contributed to so far documentation you can reach the solution additional notes no response reviewers ömer talip akalın deadline
1
365,903
10,799,514,235
IssuesEvent
2019-11-06 12:21:57
ADFC-Hamburg/adfc-t30-paten-frontend
https://api.github.com/repos/ADFC-Hamburg/adfc-t30-paten-frontend
closed
Liveversion auf aktuellste Version
high-priority
Bitte die aktuelle Version, (am besten) nach Änderung des Datums zum Stichwort, in die Liveversion einspielen, damit die Kampagne los gehen kann.
1.0
Liveversion auf aktuellste Version - Bitte die aktuelle Version, (am besten) nach Änderung des Datums zum Stichwort, in die Liveversion einspielen, damit die Kampagne los gehen kann.
priority
liveversion auf aktuellste version bitte die aktuelle version am besten nach änderung des datums zum stichwort in die liveversion einspielen damit die kampagne los gehen kann
1
402,434
11,809,716,916
IssuesEvent
2020-03-19 15:22:47
cloudfoundry-incubator/kubecf
https://api.github.com/repos/cloudfoundry-incubator/kubecf
opened
PRs to fix patches in *php-buildpack* upstream project
Priority: High Type: Patch
./api/php-buildpack/jobs/create_bosh_pre_start.sh ./api/php-buildpack/ig_resolver/patch_job_mf.sh
1.0
PRs to fix patches in *php-buildpack* upstream project - ./api/php-buildpack/jobs/create_bosh_pre_start.sh ./api/php-buildpack/ig_resolver/patch_job_mf.sh
priority
prs to fix patches in php buildpack upstream project api php buildpack jobs create bosh pre start sh api php buildpack ig resolver patch job mf sh
1
9,934
2,608,968,736
IssuesEvent
2015-02-26 11:21:09
GrannyCookies/scratchext2
https://api.github.com/repos/GrannyCookies/scratchext2
opened
Extension not loaded when using "Create" to make a new project
bug high priority JavaScript todo
Found by @Felizolinha: The new loading of the extension does not work when using "Create" to make a new project until the user clicks "See the project page" or reloads it. A workaround for it would be redefining JSredirectTo() to load ScratchExt when the url ends in "projects/editor". Please can somebody fix this, probably in the gh-pages branch (scratchext.user.js)
1.0
Extension not loaded when using "Create" to make a new project - Found by @Felizolinha: The new loading of the extension does not work when using "Create" to make a new project until the user clicks "See the project page" or reloads it. A workaround for it would be redefining JSredirectTo() to load ScratchExt when the url ends in "projects/editor". Please can somebody fix this, probably in the gh-pages branch (scratchext.user.js)
priority
extension not loaded when using create to make a new project found by felizolinha the new loading of the extension does not work when using create to make a new project until the user clicks see the project page or reloads it a workaround for it would be redefining jsredirectto to load scratchext when the url ends in projects editor please can somebody fix this probably in the gh pages branch scratchext user js
1
499,177
14,442,342,523
IssuesEvent
2020-12-07 18:00:57
Edgeryders-Participio/realities
https://api.github.com/repos/Edgeryders-Participio/realities
closed
Create orgs and users that are missing in the mongodb
Priority: High
Right now e.g. I don't exist in the prod plato-core db. Prod doesn't seem to crash though so that's fine for now. But we want orgs and users to exist there so we can properly use orgIds. My thinking atm (middle of the night): * Orgs: The single placeholder org we have now can be created in a db migration (built this for neo4j migrations but using it for mongo works fine as well I guess). Additional orgs we'll worry about when we add ui for creating those * Users: Create them in the same place we're creating them for neo4j (i.e. in auth-callback. make sure to not run it if we don't have an orgId(silent refresh), we want the user to have an org)
1.0
Create orgs and users that are missing in the mongodb - Right now e.g. I don't exist in the prod plato-core db. Prod doesn't seem to crash though so that's fine for now. But we want orgs and users to exist there so we can properly use orgIds. My thinking atm (middle of the night): * Orgs: The single placeholder org we have now can be created in a db migration (built this for neo4j migrations but using it for mongo works fine as well I guess). Additional orgs we'll worry about when we add ui for creating those * Users: Create them in the same place we're creating them for neo4j (i.e. in auth-callback. make sure to not run it if we don't have an orgId(silent refresh), we want the user to have an org)
priority
create orgs and users that are missing in the mongodb right now e g i don t exist in the prod plato core db prod doesn t seem to crash though so that s fine for now but we want orgs and users to exist there so we can properly use orgids my thinking atm middle of the night orgs the single placeholder org we have now can be created in a db migration built this for migrations but using it for mongo works fine as well i guess additional orgs we ll worry about when we add ui for creating those users create them in the same place we re creating them for i e in auth callback make sure to not run it if we don t have an orgid silent refresh we want the user to have an org
1
706,647
24,280,360,143
IssuesEvent
2022-09-28 16:50:56
AY2223S1-CS2103T-W15-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-W15-3/tp
closed
Update Developer Guide (Use Cases)
priority.High type.Task
# TODO Give use cases (textual form) for a few representative user stories that need multiple steps to complete. _e.g. Adding a tag to a person (assume the user needs to find the person first)_ See sample [here](https://se-education.org/addressbook-level3/DeveloperGuide.html#use-cases). ## Things to note (and/or avoid) - Important use cases missing (a use case is important if it involves a user interaction that is worthy of documenting e.g., it has multiple extensions -- this is not the same as the feature being important) - Formatting/notational errors - Incorrect step numbering - Unnecessary UI details mentioned - Missing/unnecessary steps - Missing extensions
1.0
Update Developer Guide (Use Cases) - # TODO Give use cases (textual form) for a few representative user stories that need multiple steps to complete. _e.g. Adding a tag to a person (assume the user needs to find the person first)_ See sample [here](https://se-education.org/addressbook-level3/DeveloperGuide.html#use-cases). ## Things to note (and/or avoid) - Important use cases missing (a use case is important if it involves a user interaction that is worthy of documenting e.g., it has multiple extensions -- this is not the same as the feature being important) - Formatting/notational errors - Incorrect step numbering - Unnecessary UI details mentioned - Missing/unnecessary steps - Missing extensions
priority
update developer guide use cases todo give use cases textual form for a few representative user stories that need multiple steps to complete e g adding a tag to a person assume the user needs to find the person first see sample things to note and or avoid important use cases missing a use case is important if it involves a user interaction that is worthy of documenting e g it has multiple extensions this is not the same as the feature being important formatting notational errors incorrect step numbering unnecessary ui details mentioned missing unnecessary steps missing extensions
1
798,545
28,289,461,848
IssuesEvent
2023-04-09 02:23:43
KATO-Hiro/AtCoderClans
https://api.github.com/repos/KATO-Hiro/AtCoderClans
closed
一部のページのタイトルが飛んでいるかも?
invalid priority high
## WHAT - サイトジェネレータをv9に変えたあたりから? - 一部のページのみ該当している - 本家の#4824が関連しているか? ## TODO - [ ] 該当するページを列挙 - [ ] ヘッダを追加?
1.0
一部のページのタイトルが飛んでいるかも? - ## WHAT - サイトジェネレータをv9に変えたあたりから? - 一部のページのみ該当している - 本家の#4824が関連しているか? ## TODO - [ ] 該当するページを列挙 - [ ] ヘッダを追加?
priority
一部のページのタイトルが飛んでいるかも what 一部のページのみ該当している 本家の todo 該当するページを列挙 ヘッダを追加
1
659,689
21,937,739,193
IssuesEvent
2022-05-23 15:06:18
ooni/ooni.org
https://api.github.com/repos/ooni/ooni.org
opened
Prepare for Internet without Borders events
priority/high workshop community
OONI is participating in the Internet without Borders events (https://internetborders.net/) for the Russian diaspora. As part of these events, we'll be giving OONI presentations and facilitating OONI hackathons. In preparation for these events, we're working on the following: - [ ] Create (and share) slides for OONI presentations - [ ] Finalize challenges for OONI hackathons - [ ] Create video for OONI hackathons - [ ] Create OONI channel on the event Mattermost (for hackathon coordination) - [ ] Update relevant spreadsheet (coordination with organizers)
1.0
Prepare for Internet without Borders events - OONI is participating in the Internet without Borders events (https://internetborders.net/) for the Russian diaspora. As part of these events, we'll be giving OONI presentations and facilitating OONI hackathons. In preparation for these events, we're working on the following: - [ ] Create (and share) slides for OONI presentations - [ ] Finalize challenges for OONI hackathons - [ ] Create video for OONI hackathons - [ ] Create OONI channel on the event Mattermost (for hackathon coordination) - [ ] Update relevant spreadsheet (coordination with organizers)
priority
prepare for internet without borders events ooni is participating in the internet without borders events for the russian diaspora as part of these events we ll be giving ooni presentations and facilitating ooni hackathons in preparation for these events we re working on the following create and share slides for ooni presentations finalize challenges for ooni hackathons create video for ooni hackathons create ooni channel on the event mattermost for hackathon coordination update relevant spreadsheet coordination with organizers
1
74,049
3,427,685,633
IssuesEvent
2015-12-10 03:34:59
phetsims/website
https://api.github.com/repos/phetsims/website
closed
Change directions for comments on activity pages
High Priority Teacher Resources pages
Currently, on any activity page, users can add a comment. The text next to the entry box says: "What do you think about this activity? How did you use it or change it for your class? Professionally constructive comments welcome." We would like to change it to say: "How did you use this activity or change it for your class? Answer keys are often not available because the activities are inquiry based and lead to multiple answers. Professionally constructive comments welcome."
1.0
Change directions for comments on activity pages - Currently, on any activity page, users can add a comment. The text next to the entry box says: "What do you think about this activity? How did you use it or change it for your class? Professionally constructive comments welcome." We would like to change it to say: "How did you use this activity or change it for your class? Answer keys are often not available because the activities are inquiry based and lead to multiple answers. Professionally constructive comments welcome."
priority
change directions for comments on activity pages currently on any activity page users can add a comment the text next to the entry box says what do you think about this activity how did you use it or change it for your class professionally constructive comments welcome we would like to change it to say how did you use this activity or change it for your class answer keys are often not available because the activities are inquiry based and lead to multiple answers professionally constructive comments welcome
1
541,653
15,830,997,789
IssuesEvent
2021-04-06 13:10:09
bryntum/support
https://api.github.com/repos/bryntum/support
opened
Popups misaligned
bug high-priority
Reproducible in blog inside a web component <img width="906" alt="Screenshot 2021-04-06 at 15 09 17" src="https://user-images.githubusercontent.com/218570/113715687-15f7b100-96ea-11eb-8f9f-d45c63d95626.png">
1.0
Popups misaligned - Reproducible in blog inside a web component <img width="906" alt="Screenshot 2021-04-06 at 15 09 17" src="https://user-images.githubusercontent.com/218570/113715687-15f7b100-96ea-11eb-8f9f-d45c63d95626.png">
priority
popups misaligned reproducible in blog inside a web component img width alt screenshot at src
1
428,183
12,404,316,628
IssuesEvent
2020-05-21 15:20:17
nthopinion/covid19
https://api.github.com/repos/nthopinion/covid19
closed
Physician Registration: read-only version of physician page for physicians with "level 0" access
High Priority Trello [Sync]
* they can rate answers * they can't answer questions (disabled fields) Context: https://docs.google.com/document/d/1y-26s8otakzhvMqKYpj_3xNFNdGsa1REpwlt4sNLL2I/edit Related tasks in Trello attachments. ┆Issue is synchronized with this [Trello card](https://trello.com/c/PYZEjPin) by [Unito](https://www.unito.io/learn-more) ┆Attachments: <a href="https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access">https:&#x2F;&#x2F;trello.com&#x2F;c&#x2F;VIoaBpVK&#x2F;157-physician-registration-introduce-layered-access</a> | <a href="https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information">https:&#x2F;&#x2F;trello.com&#x2F;c&#x2F;0fSXMht4&#x2F;159-physician-registration-provide-a-way-for-the-user-to-update-their-information</a>
1.0
Physician Registration: read-only version of physician page for physicians with "level 0" access - * they can rate answers * they can't answer questions (disabled fields) Context: https://docs.google.com/document/d/1y-26s8otakzhvMqKYpj_3xNFNdGsa1REpwlt4sNLL2I/edit Related tasks in Trello attachments. ┆Issue is synchronized with this [Trello card](https://trello.com/c/PYZEjPin) by [Unito](https://www.unito.io/learn-more) ┆Attachments: <a href="https://trello.com/c/VIoaBpVK/157-physician-registration-introduce-layered-access">https:&#x2F;&#x2F;trello.com&#x2F;c&#x2F;VIoaBpVK&#x2F;157-physician-registration-introduce-layered-access</a> | <a href="https://trello.com/c/0fSXMht4/159-physician-registration-provide-a-way-for-the-user-to-update-their-information">https:&#x2F;&#x2F;trello.com&#x2F;c&#x2F;0fSXMht4&#x2F;159-physician-registration-provide-a-way-for-the-user-to-update-their-information</a>
priority
physician registration read only version of physician page for physicians with level access they can rate answers they can t answer questions disabled fields context related tasks in trello attachments ┆issue is synchronized with this by ┆attachments a href a href
1
364,270
10,761,538,483
IssuesEvent
2019-10-31 21:00:54
semperfiwebdesign/all-in-one-seo-pack
https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack
closed
Fatal error: Uncaught Error in PHP 7.1
Priority | High
Reported here: https://wordpress.org/support/topic/php-7-1-problem/ I'm not exactly sure what's prompting this, but I don't think this has to do with PHP 7.1. My guess is that the edit array key simply doesn't exist already for some reason. It may be as simple as checking for it first. unset( $actions['edit'] );
1.0
Fatal error: Uncaught Error in PHP 7.1 - Reported here: https://wordpress.org/support/topic/php-7-1-problem/ I'm not exactly sure what's prompting this, but I don't think this has to do with PHP 7.1. My guess is that the edit array key simply doesn't exist already for some reason. It may be as simple as checking for it first. unset( $actions['edit'] );
priority
fatal error uncaught error in php reported here i m not exactly sure what s prompting this but i don t think this has to do with php my guess is that the edit array key simply doesn t exist already for some reason it may be as simple as checking for it first unset actions
1
725,881
24,979,279,007
IssuesEvent
2022-11-02 10:23:48
AY2223S1-CS2113-W13-1/tp
https://api.github.com/repos/AY2223S1-CS2113-W13-1/tp
closed
Be able to cancel login
type.Bug priority.High
Once you are in the process of logging in but do not want to log in you should be able to exit
1.0
Be able to cancel login - Once you are in the process of logging in but do not want to log in you should be able to exit
priority
be able to cancel login once you are in the process of logging in but do not want to log in you should be able to exit
1
684,072
23,405,996,569
IssuesEvent
2022-08-12 12:54:15
cassproject/cass-editor
https://api.github.com/repos/cassproject/cass-editor
closed
Missing a few competencies for DOL Frameworks and not able to export them with this error "TypeError: Cannot read property 'ceasn:isChildOf' of undefined"
blocker High Priority Credential Engine CaSS
@gloverkari There are the two competencies i am not able to export and seeing the same error DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139 CASS-- https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139 DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d CASS-https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
1.0
Missing a few competencies for DOL Frameworks and not able to export them with this error "TypeError: Cannot read property 'ceasn:isChildOf' of undefined" - @gloverkari There are the two competencies i am not able to export and seeing the same error DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139 CASS-- https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-7b219ae4-3c4a-4e64-9100-db31e5e18139 DOL-https://careeronestop.org/CompetencyModel/ctdlasn/graph/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d CASS-https://cass.credentialengine.org/cass-editor/?user=wait&editorRoot=/cass-editor/&server=https://cass.credentialengine.org/api/&ceasnDataFields=true&&newObjectEndpoint=https://credentialengineregistry.org/resources/ce-&show=all&origin=https://apps.credentialengine.org#?frameworkId=http://www.careeronestop.org/CompetencyModel/ctdlasn/resources/ce-e61116f7-5770-4431-bfc0-fb0fa5bd8d5d
priority
missing a few competencies for dol frameworks and not able to export them with this error typeerror cannot read property ceasn ischildof of undefined gloverkari there are the two competencies i am not able to export and seeing the same error dol cass dol cass
1
716,026
24,618,672,844
IssuesEvent
2022-10-15 16:40:51
onesoft-sudo/sudobot
https://api.github.com/repos/onesoft-sudo/sudobot
closed
Profile filter interferes with muting users
bug help wanted priority:high important status:ready-for-fixing semver:minor
After a mute by profile filter, if a user gets muted by a moderator, they will be unmuted back after profile filter takeback. Add a check if the moderators has run mute command.
1.0
Profile filter interferes with muting users - After a mute by profile filter, if a user gets muted by a moderator, they will be unmuted back after profile filter takeback. Add a check if the moderators has run mute command.
priority
profile filter interferes with muting users after a mute by profile filter if a user gets muted by a moderator they will be unmuted back after profile filter takeback add a check if the moderators has run mute command
1
555,705
16,463,730,436
IssuesEvent
2021-05-22 01:50:52
ncopenpass/CampaignFinance
https://api.github.com/repos/ncopenpass/CampaignFinance
opened
design: create design for amount filter on tables
Class Prototype High Priority ui
### Task Update the figma prototype to include the amount filter.
1.0
design: create design for amount filter on tables - ### Task Update the figma prototype to include the amount filter.
priority
design create design for amount filter on tables task update the figma prototype to include the amount filter
1
263,626
8,299,271,563
IssuesEvent
2018-09-21 01:53:55
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] documentation button under help in context nav goes to invalid documentation link
bug priority: high
### Expected behavior Documentation site should load ### Actual behavior Button loads URL: https://docs.craftercms.org/en/3.0/ URL is missing index.html User gets S3's 404 ### Steps to reproduce the problem * log in to Studio * click help * click documentation ### Log/stack trace (use https://gist.github.com) N/A ### Specs #### Version Studio Version Number: 3.0.17-SNAPSHOT-5e8acf Build Number: 5e8acf27a697e7a6a52f4156a5c390a3538c6687 Build Date/Time: 09-19-2018 16:41:26 -0400
1.0
[studio-ui] documentation button under help in context nav goes to invalid documentation link - ### Expected behavior Documentation site should load ### Actual behavior Button loads URL: https://docs.craftercms.org/en/3.0/ URL is missing index.html User gets S3's 404 ### Steps to reproduce the problem * log in to Studio * click help * click documentation ### Log/stack trace (use https://gist.github.com) N/A ### Specs #### Version Studio Version Number: 3.0.17-SNAPSHOT-5e8acf Build Number: 5e8acf27a697e7a6a52f4156a5c390a3538c6687 Build Date/Time: 09-19-2018 16:41:26 -0400
priority
documentation button under help in context nav goes to invalid documentation link expected behavior documentation site should load actual behavior button loads url url is missing index html user gets s steps to reproduce the problem log in to studio click help click documentation log stack trace use n a specs version studio version number snapshot build number build date time
1
196,845
6,949,968,918
IssuesEvent
2017-12-06 09:05:37
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
remoteshell is very slow
component:coral priority:high sprint2 status:pending type:usability
`remoteshell` has a LOT of long-ish sleeps in it. ``` [root@mgmt1 ~]# time updatenode h41n01 -P remoteshell h41n01: xcatdsklspost: downloaded postscripts successfully h41n01: Fri Dec 1 01:06:17 EST 2017 Running postscript: remoteshell h41n01: h41n01: postscript: remoteshell exited with code 0 h41n01: Running of postscripts has completed. real 2m17.663s user 0m0.566s sys 0m0.029s ``` I tried to enable xCAT flow support, but that actually took longer. ``` [root@mgmt1 ~]# time updatenode h41n01 -P remoteshell h41n01: xcatdsklspost: downloaded postscripts successfully h41n01: Fri Dec 1 01:10:07 EST 2017 Running postscript: remoteshell h41n01: h41n01: postscript: remoteshell exited with code 0 h41n01: Running of postscripts has completed. real 3m0.601s user 0m0.588s sys 0m0.010s ``` I hopped on the node and saw 14 concurrent `xcatflowrequest` commands that were reparented to init. I didn't tune `xcatmaxconnections` or `xcatmaxbatchconnections`, but I only ran this on a single node, so the defaults should be ok. It would be great if `remoteshell` could get all the keys in a single command instead of one sleep+RTT for each key.
1.0
remoteshell is very slow - `remoteshell` has a LOT of long-ish sleeps in it. ``` [root@mgmt1 ~]# time updatenode h41n01 -P remoteshell h41n01: xcatdsklspost: downloaded postscripts successfully h41n01: Fri Dec 1 01:06:17 EST 2017 Running postscript: remoteshell h41n01: h41n01: postscript: remoteshell exited with code 0 h41n01: Running of postscripts has completed. real 2m17.663s user 0m0.566s sys 0m0.029s ``` I tried to enable xCAT flow support, but that actually took longer. ``` [root@mgmt1 ~]# time updatenode h41n01 -P remoteshell h41n01: xcatdsklspost: downloaded postscripts successfully h41n01: Fri Dec 1 01:10:07 EST 2017 Running postscript: remoteshell h41n01: h41n01: postscript: remoteshell exited with code 0 h41n01: Running of postscripts has completed. real 3m0.601s user 0m0.588s sys 0m0.010s ``` I hopped on the node and saw 14 concurrent `xcatflowrequest` commands that were reparented to init. I didn't tune `xcatmaxconnections` or `xcatmaxbatchconnections`, but I only ran this on a single node, so the defaults should be ok. It would be great if `remoteshell` could get all the keys in a single command instead of one sleep+RTT for each key.
priority
remoteshell is very slow remoteshell has a lot of long ish sleeps in it time updatenode p remoteshell xcatdsklspost downloaded postscripts successfully fri dec est running postscript remoteshell postscript remoteshell exited with code running of postscripts has completed real user sys i tried to enable xcat flow support but that actually took longer time updatenode p remoteshell xcatdsklspost downloaded postscripts successfully fri dec est running postscript remoteshell postscript remoteshell exited with code running of postscripts has completed real user sys i hopped on the node and saw concurrent xcatflowrequest commands that were reparented to init i didn t tune xcatmaxconnections or xcatmaxbatchconnections but i only ran this on a single node so the defaults should be ok it would be great if remoteshell could get all the keys in a single command instead of one sleep rtt for each key
1
313,126
9,557,487,695
IssuesEvent
2019-05-03 11:43:19
fritzing/fritzing-app
https://api.github.com/repos/fritzing/fritzing-app
closed
disappearing parts--when changing z order of a wire in bb view
Priority-High bug imported
_From [irasc...@gmail.com](https://code.google.com/u/104729248032245122687/) on June 11, 2013 12:52:04_ Attached sketch no longer exhibits the behavior. Something is broken in the program state, but saving and loading cleans up the problem. When the striped wire was raised in z-order, the resistor disappeared in bb view. The part was still visible in the other views. **Attachment:** [DC_motor_control---disapearing parts.fzz](http://code.google.com/p/fritzing/issues/detail?id=2623) _Original issue: http://code.google.com/p/fritzing/issues/detail?id=2623_
1.0
disappearing parts--when changing z order of a wire in bb view - _From [irasc...@gmail.com](https://code.google.com/u/104729248032245122687/) on June 11, 2013 12:52:04_ Attached sketch no longer exhibits the behavior. Something is broken in the program state, but saving and loading cleans up the problem. When the striped wire was raised in z-order, the resistor disappeared in bb view. The part was still visible in the other views. **Attachment:** [DC_motor_control---disapearing parts.fzz](http://code.google.com/p/fritzing/issues/detail?id=2623) _Original issue: http://code.google.com/p/fritzing/issues/detail?id=2623_
priority
disappearing parts when changing z order of a wire in bb view from on june attached sketch no longer exhibits the behavior something is broken in the program state but saving and loading cleans up the problem when the striped wire was raised in z order the resistor disappeared in bb view the part was still visible in the other views attachment original issue
1
660,878
22,034,515,033
IssuesEvent
2022-05-28 10:58:06
naturalcrit/homebrewery
https://api.github.com/repos/naturalcrit/homebrewery
closed
Columns not displaying correctly on Chrome 102.0.5005.63
bug solution found P1 - high priority
## Environment **Browser(s):** Chrome Version 102.0.5005.63 (Official Build) (64-bit) **Operating System:** Windows 10 (x64) **Legacy or v3 Renderer:** v3 Renderer ## Description Manual column breaks are rendered with too much spacing and are pushed off the page on Chrome 102.0.5005.63 but not on earlier versions. ## Images [Chrome 101.0.4951.67](https://i.imgur.com/gKY8d44.png) [Chrome 102.0.5005.63](https://i.imgur.com/q21LTWZ.png) **Share Link** : https://homebrewery.naturalcrit.com/share/19O72H7NrXlGJ17rXmBXZ13QkDBSNGyw0fd5PofH4aTmO
1.0
Columns not displaying correctly on Chrome 102.0.5005.63 - ## Environment **Browser(s):** Chrome Version 102.0.5005.63 (Official Build) (64-bit) **Operating System:** Windows 10 (x64) **Legacy or v3 Renderer:** v3 Renderer ## Description Manual column breaks are rendered with too much spacing and are pushed off the page on Chrome 102.0.5005.63 but not on earlier versions. ## Images [Chrome 101.0.4951.67](https://i.imgur.com/gKY8d44.png) [Chrome 102.0.5005.63](https://i.imgur.com/q21LTWZ.png) **Share Link** : https://homebrewery.naturalcrit.com/share/19O72H7NrXlGJ17rXmBXZ13QkDBSNGyw0fd5PofH4aTmO
priority
columns not displaying correctly on chrome environment browser s chrome version official build bit operating system windows legacy or renderer renderer description manual column breaks are rendered with too much spacing and are pushed off the page on chrome but not on earlier versions images share link
1
636,913
20,612,987,606
IssuesEvent
2022-03-07 10:27:23
Soulcialize/souldragonknight
https://api.github.com/repos/Soulcialize/souldragonknight
opened
Implement AI for Alpha stage enemies
type.Enhancement priority.High
There are four types of enemies planned for Alpha stage. Two of them will target the Knight, while the other two will target the Dragon. **Dragon:** - [ ] Flying enemy, floats towards the Dragon before charging up a melee attack - [ ] Ground enemy, protected/shielded such that the Dragon is unable to damage it. The Dragon must empower the knight's attack so that the knight can destroy this enemy. This enemy will fire projectiles at the Dragon. **Knight:** - [ ] Ground enemy, runs towards the Knight before charging up a melee attack - [ ] Flying enemy, fire projectiles at the Knight. The enemy will be flying out of the Knight's reach. The Knight must ride the Dragon so that the Knight can attack and destroy this enemy.
1.0
Implement AI for Alpha stage enemies - There are four types of enemies planned for Alpha stage. Two of them will target the Knight, while the other two will target the Dragon. **Dragon:** - [ ] Flying enemy, floats towards the Dragon before charging up a melee attack - [ ] Ground enemy, protected/shielded such that the Dragon is unable to damage it. The Dragon must empower the knight's attack so that the knight can destroy this enemy. This enemy will fire projectiles at the Dragon. **Knight:** - [ ] Ground enemy, runs towards the Knight before charging up a melee attack - [ ] Flying enemy, fire projectiles at the Knight. The enemy will be flying out of the Knight's reach. The Knight must ride the Dragon so that the Knight can attack and destroy this enemy.
priority
implement ai for alpha stage enemies there are four types of enemies planned for alpha stage two of them will target the knight while the other two will target the dragon dragon flying enemy floats towards the dragon before charging up a melee attack ground enemy protected shielded such that the dragon is unable to damage it the dragon must empower the knight s attack so that the knight can destroy this enemy this enemy will fire projectiles at the dragon knight ground enemy runs towards the knight before charging up a melee attack flying enemy fire projectiles at the knight the enemy will be flying out of the knight s reach the knight must ride the dragon so that the knight can attack and destroy this enemy
1
390,271
11,541,155,897
IssuesEvent
2020-02-18 02:56:37
ritsec/cluster-duck
https://api.github.com/repos/ritsec/cluster-duck
closed
Change Wan network IP range
competition configuration high-priority management
Change Wan network IP range ============== Change Wan network IP range to the upper 2/3rds of our current /24 network
1.0
Change Wan network IP range - Change Wan network IP range ============== Change Wan network IP range to the upper 2/3rds of our current /24 network
priority
change wan network ip range change wan network ip range change wan network ip range to the upper of our current network
1
117,903
4,728,890,290
IssuesEvent
2016-10-18 17:10:00
MRN-Code/penny-collector
https://api.github.com/repos/MRN-Code/penny-collector
closed
Rework file transfer
enhancement high priority
File in question: https://github.com/MRN-Code/penny-collector/blob/develop/src/transferFile.js. It’s pretty difficult to reason about. Turns out, the [Node.js tus client supports streams](https://github.com/tus/tus-js-client/blob/master/demo/node.js). We should be able to do something like: ```js const promisify = require('bluebird').promisify; const tar = require('tar-fs'); const tus = require('tus'); module.exports = transferFile(filePath) { const file = tar.pack(filePath); return promisify(fs.stat)(file).then(stat => { return new Promise((resolve, reject) => { const upload = new tus.Client(file, { endpoint: '', // Get form configuration metadata: { file: filePath, }, onError: reject, onSuccess: resolve, resume: true, uploadSize: stat.size, }); upload.start(); }); }); }; ```
1.0
Rework file transfer - File in question: https://github.com/MRN-Code/penny-collector/blob/develop/src/transferFile.js. It’s pretty difficult to reason about. Turns out, the [Node.js tus client supports streams](https://github.com/tus/tus-js-client/blob/master/demo/node.js). We should be able to do something like: ```js const promisify = require('bluebird').promisify; const tar = require('tar-fs'); const tus = require('tus'); module.exports = transferFile(filePath) { const file = tar.pack(filePath); return promisify(fs.stat)(file).then(stat => { return new Promise((resolve, reject) => { const upload = new tus.Client(file, { endpoint: '', // Get form configuration metadata: { file: filePath, }, onError: reject, onSuccess: resolve, resume: true, uploadSize: stat.size, }); upload.start(); }); }); }; ```
priority
rework file transfer file in question it’s pretty difficult to reason about turns out the we should be able to do something like js const promisify require bluebird promisify const tar require tar fs const tus require tus module exports transferfile filepath const file tar pack filepath return promisify fs stat file then stat return new promise resolve reject const upload new tus client file endpoint get form configuration metadata file filepath onerror reject onsuccess resolve resume true uploadsize stat size upload start
1
180,922
6,654,438,823
IssuesEvent
2017-09-29 12:50:28
CoderDojo/community-platform
https://api.github.com/repos/CoderDojo/community-platform
closed
Minify all static files
high priority technical
We are including separately all of the javascript and css files we need. We could make the site a lot faster and it would help our SEO to minify and combine the bower files. e.g. http://stackoverflow.com/questions/25322326/combine-and-minify-all-bower-libraries-with-gruntjs
1.0
Minify all static files - We are including separately all of the javascript and css files we need. We could make the site a lot faster and it would help our SEO to minify and combine the bower files. e.g. http://stackoverflow.com/questions/25322326/combine-and-minify-all-bower-libraries-with-gruntjs
priority
minify all static files we are including separately all of the javascript and css files we need we could make the site a lot faster and it would help our seo to minify and combine the bower files e g
1
796,999
28,134,890,142
IssuesEvent
2023-04-01 08:58:36
telstra/open-kilda
https://api.github.com/repos/telstra/open-kilda
opened
Add ability to validate HA flow
priority/2-high feature
Ha flow validation must use rile manager to build expected rules Part of #5061
1.0
Add ability to validate HA flow - Ha flow validation must use rile manager to build expected rules Part of #5061
priority
add ability to validate ha flow ha flow validation must use rile manager to build expected rules part of
1
160,790
6,102,586,658
IssuesEvent
2017-06-20 16:48:53
crowdAI/crowdai
https://api.github.com/repos/crowdAI/crowdai
closed
CrowdAI logo on mobile different
high priority v2
For some reason, the crowdAI samurAI loses his eyes on mobile... <img width="402" alt="screen shot 2017-06-17 at 1 56 20 pm" src="https://user-images.githubusercontent.com/215057/27252691-e14b90e2-5364-11e7-94b8-61f0a3bd6339.png">
1.0
CrowdAI logo on mobile different - For some reason, the crowdAI samurAI loses his eyes on mobile... <img width="402" alt="screen shot 2017-06-17 at 1 56 20 pm" src="https://user-images.githubusercontent.com/215057/27252691-e14b90e2-5364-11e7-94b8-61f0a3bd6339.png">
priority
crowdai logo on mobile different for some reason the crowdai samurai loses his eyes on mobile img width alt screen shot at pm src
1