Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
675,771
23,105,212,852
IssuesEvent
2022-07-27 08:13:12
wfau/aglais
https://api.github.com/repos/wfau/aglais
closed
DR3 data sets for inclusion in automated deploy
high priority live-20220725 20220723-zrq-science-data
At this stage the new data sets that should be included in automated deploy from Gaia DR3 (2048 partitioned versions) are GDR3_ASTROPHYSICAL_PARAMETERS GDR3_EPOCH_PHOTOMETRY GDR3_RVS_MEAN_SPECTRUM GDR3_XP_SAMPLED_MEAN_SPECTRUM GDR3_ASTROPHYSICAL_PARAMETERS_SUPP GDR3_GAIA_SOURCE GDR3_XP_CONTINUOUS_MEAN_SPECTRUM GDR3_XP_SUMMARY as available at /user/nch/PARQUET/GDR3. Copies/links should be made as appropriate. At some point in the future there will be more, but for now this is what we have, and what is assumed to be available in gaiadmpsetup.
1.0
DR3 data sets for inclusion in automated deploy - At this stage the new data sets that should be included in automated deploy from Gaia DR3 (2048 partitioned versions) are GDR3_ASTROPHYSICAL_PARAMETERS GDR3_EPOCH_PHOTOMETRY GDR3_RVS_MEAN_SPECTRUM GDR3_XP_SAMPLED_MEAN_SPECTRUM GDR3_ASTROPHYSICAL_PARAMETERS_SUPP GDR3_GAIA_SOURCE GDR3_XP_CONTINUOUS_MEAN_SPECTRUM GDR3_XP_SUMMARY as available at /user/nch/PARQUET/GDR3. Copies/links should be made as appropriate. At some point in the future there will be more, but for now this is what we have, and what is assumed to be available in gaiadmpsetup.
priority
data sets for inclusion in automated deploy at this stage the new data sets that should be included in automated deploy from gaia partitioned versions are astrophysical parameters epoch photometry rvs mean spectrum xp sampled mean spectrum astrophysical parameters supp gaia source xp continuous mean spectrum xp summary as available at user nch parquet copies links should be made as appropriate at some point in the future there will be more but for now this is what we have and what is assumed to be available in gaiadmpsetup
1
370,083
10,925,051,002
IssuesEvent
2019-11-22 11:33:18
bounswe/bounswe2019group7
https://api.github.com/repos/bounswe/bounswe2019group7
closed
Currency Page on Android
Android Effort: More Than a Day Priority: High
In addition to currency converter already implemented, a currencies page will be added which shows currency comparisons with some plots and past data. Deadline: 25.11.2019 @23.59
1.0
Currency Page on Android - In addition to currency converter already implemented, a currencies page will be added which shows currency comparisons with some plots and past data. Deadline: 25.11.2019 @23.59
priority
currency page on android in addition to currency converter already implemented a currencies page will be added which shows currency comparisons with some plots and past data deadline
1
577,051
17,102,329,414
IssuesEvent
2021-07-09 13:04:54
georchestra/mapstore2-georchestra
https://api.github.com/repos/georchestra/mapstore2-georchestra
closed
Plugins & DateTimePicker works only with nativ mapstore2
Priority: High User feedback estimate needed investigation
Hi, **Describe** I get some troubles when i try to use the DateTimePicker component (react-widgets) available with mapstore2 in a plugin. When i try to use the SampleExtension plugin with the DateTimePicker in a mapstore2-georchestra context, the app is broken (see screenshot below). In dev mode from MapStoreExtension and MapStore2 without georchestra (npm start), all works perfectly (see below). **Steps** 1. Clone MapStoreExtension 2. Add simple DateTimePicker component 3. Start webpack (npm start) 4. Test plugin and change date ==> OK 5. Build plugin (get ZIP) 6. Open context and test calendar 5. App crash as below **Screenshot** * First part : webpack dev server ==> OK * Second part : Rennes Métropole mapstore2-georchestra ==> KO ![issue113_2](https://user-images.githubusercontent.com/16317988/120352176-cf48c080-c300-11eb-93a6-e02b04f12fbc.gif) **Versions** Mapstore2-geochestra 1.2.x (Rennes Métropole test env). **Logs** ``` TypeError: Cannot read property 'time' of undefined at Object.getFormat (99.7a608ef7.js:2) at p (99.7a608ef7.js:2) at t._dates (99.7a608ef7.js:2) at t.getInitialState (99.7a608ef7.js:2) at new t (99.7a608ef7.js:2) at bi (geOrchestra.js?a15da53b940d7f0e4793:2) at Ia (geOrchestra.js?a15da53b940d7f0e4793:2) TypeError: Cannot read property 'header' of undefined at Object.getFormat (99.7a608ef7.js:2) at z (99.7a608ef7.js:2) at t._label (99.7a608ef7.js:2) at t.render (99.7a608ef7.js:2) at La (geOrchestra.js?a15da53b940d7f0e4793:2) ``` **Code to reproduce** In SampleExtension plugin, Replace component.jsx content by this code : ``` import React from "react"; import Message from "@mapstore/components/I18N/Message"; import src from "../assets/markers_default.png"; import { DateTimePicker } from 'react-widgets'; const Extension = ({ value = 0, onIncrease, changeZoomLevel }) => { return (<div id="SAMPLE_EXTENSION" > <h2>Extension Sample</h2> <div>This is a sample extension plugin. The following tools demonstrate the correct binding inside MapStore</div> <h3>State and epics</h3> <div>A button like this should be visualized also in the toolbar: <br /><button onClick={onIncrease} className="btn-primary square-button btn">INC</button><br /> </div> <div>Clicking on that button (or on the button above) should increase this counter value: <b>{value}</b>. <br />The counter value updates should be logged in console by an epic</div> <i>note: the button in the toolbar can be hidden, in this case click on the "..." button</i> <h3>Localization</h3> The following text should be localized: <b><Message msgId="extension.message" /></b><br /> (you should see something like "Message!" if your language is set to en-US) <h3>Core action</h3> This button should change the zoom level to "1" <button onClick={() => { changeZoomLevel(1); }}>zoom to level 1</button> Here a sample image with several markers: <DateTimePicker type="date" dropUp calendar={true} time={false} culture="fr" format="DD/MM/YYYY" /> <img src={src}/> </div>); }; ```
1.0
Plugins & DateTimePicker works only with nativ mapstore2 - Hi, **Describe** I get some troubles when i try to use the DateTimePicker component (react-widgets) available with mapstore2 in a plugin. When i try to use the SampleExtension plugin with the DateTimePicker in a mapstore2-georchestra context, the app is broken (see screenshot below). In dev mode from MapStoreExtension and MapStore2 without georchestra (npm start), all works perfectly (see below). **Steps** 1. Clone MapStoreExtension 2. Add simple DateTimePicker component 3. Start webpack (npm start) 4. Test plugin and change date ==> OK 5. Build plugin (get ZIP) 6. Open context and test calendar 5. App crash as below **Screenshot** * First part : webpack dev server ==> OK * Second part : Rennes Métropole mapstore2-georchestra ==> KO ![issue113_2](https://user-images.githubusercontent.com/16317988/120352176-cf48c080-c300-11eb-93a6-e02b04f12fbc.gif) **Versions** Mapstore2-geochestra 1.2.x (Rennes Métropole test env). **Logs** ``` TypeError: Cannot read property 'time' of undefined at Object.getFormat (99.7a608ef7.js:2) at p (99.7a608ef7.js:2) at t._dates (99.7a608ef7.js:2) at t.getInitialState (99.7a608ef7.js:2) at new t (99.7a608ef7.js:2) at bi (geOrchestra.js?a15da53b940d7f0e4793:2) at Ia (geOrchestra.js?a15da53b940d7f0e4793:2) TypeError: Cannot read property 'header' of undefined at Object.getFormat (99.7a608ef7.js:2) at z (99.7a608ef7.js:2) at t._label (99.7a608ef7.js:2) at t.render (99.7a608ef7.js:2) at La (geOrchestra.js?a15da53b940d7f0e4793:2) ``` **Code to reproduce** In SampleExtension plugin, Replace component.jsx content by this code : ``` import React from "react"; import Message from "@mapstore/components/I18N/Message"; import src from "../assets/markers_default.png"; import { DateTimePicker } from 'react-widgets'; const Extension = ({ value = 0, onIncrease, changeZoomLevel }) => { return (<div id="SAMPLE_EXTENSION" > <h2>Extension Sample</h2> <div>This is a sample extension plugin. The following tools demonstrate the correct binding inside MapStore</div> <h3>State and epics</h3> <div>A button like this should be visualized also in the toolbar: <br /><button onClick={onIncrease} className="btn-primary square-button btn">INC</button><br /> </div> <div>Clicking on that button (or on the button above) should increase this counter value: <b>{value}</b>. <br />The counter value updates should be logged in console by an epic</div> <i>note: the button in the toolbar can be hidden, in this case click on the "..." button</i> <h3>Localization</h3> The following text should be localized: <b><Message msgId="extension.message" /></b><br /> (you should see something like "Message!" if your language is set to en-US) <h3>Core action</h3> This button should change the zoom level to "1" <button onClick={() => { changeZoomLevel(1); }}>zoom to level 1</button> Here a sample image with several markers: <DateTimePicker type="date" dropUp calendar={true} time={false} culture="fr" format="DD/MM/YYYY" /> <img src={src}/> </div>); }; ```
priority
plugins datetimepicker works only with nativ hi describe i get some troubles when i try to use the datetimepicker component react widgets available with in a plugin when i try to use the sampleextension plugin with the datetimepicker in a georchestra context the app is broken see screenshot below in dev mode from mapstoreextension and without georchestra npm start all works perfectly see below steps clone mapstoreextension add simple datetimepicker component start webpack npm start test plugin and change date ok build plugin get zip open context and test calendar app crash as below screenshot first part webpack dev server ok second part rennes métropole georchestra ko versions geochestra x rennes métropole test env logs typeerror cannot read property time of undefined at object getformat js at p js at t dates js at t getinitialstate js at new t js at bi georchestra js at ia georchestra js typeerror cannot read property header of undefined at object getformat js at z js at t label js at t render js at la georchestra js code to reproduce in sampleextension plugin replace component jsx content by this code import react from react import message from mapstore components message import src from assets markers default png import datetimepicker from react widgets const extension value onincrease changezoomlevel return extension sample this is a sample extension plugin the following tools demonstrate the correct binding inside mapstore state and epics a button like this should be visualized also in the toolbar inc clicking on that button or on the button above should increase this counter value value the counter value updates should be logged in console by an epic note the button in the toolbar can be hidden in this case click on the button localization the following text should be localized you should see something like message if your language is set to en us core action this button should change the zoom level to changezoomlevel zoom to level here a sample image with several markers datetimepicker type date dropup calendar true time false culture fr format dd mm yyyy
1
770,268
27,035,281,158
IssuesEvent
2023-02-12 17:54:42
SlimeVR/SlimeVR-Server
https://api.github.com/repos/SlimeVR/SlimeVR-Server
closed
Trackers aren't detected
Type: Bug Priority: High
On Poly's PC literally nothing happen when you plug in the tracker, even though the serial exists. ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067112371014873168/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067112456658370571/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067113024495816785/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067113388930519130/image.png) On my laptop I see a popup but it can't provision the tracker. ![image](https://user-images.githubusercontent.com/1257939/214099168-8588c2af-bf2e-4f84-a848-bf52b6707d22.png) Even though I can actually open serial console and send GET INFO. No idea how to even start debugging it.
1.0
Trackers aren't detected - On Poly's PC literally nothing happen when you plug in the tracker, even though the serial exists. ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067112371014873168/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067112456658370571/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067113024495816785/image.png) ![img](https://cdn.discordapp.com/attachments/946392672791318528/1067113388930519130/image.png) On my laptop I see a popup but it can't provision the tracker. ![image](https://user-images.githubusercontent.com/1257939/214099168-8588c2af-bf2e-4f84-a848-bf52b6707d22.png) Even though I can actually open serial console and send GET INFO. No idea how to even start debugging it.
priority
trackers aren t detected on poly s pc literally nothing happen when you plug in the tracker even though the serial exists on my laptop i see a popup but it can t provision the tracker even though i can actually open serial console and send get info no idea how to even start debugging it
1
224,905
7,473,778,883
IssuesEvent
2018-04-03 16:17:23
EvictionLab/eviction-maps
https://api.github.com/repos/EvictionLab/eviction-maps
closed
Don't show rankings numbers for areas with unavailable data
enhancement feedback needed high priority
We could potentially replace this with N/A or just leave it blank?
1.0
Don't show rankings numbers for areas with unavailable data - We could potentially replace this with N/A or just leave it blank?
priority
don t show rankings numbers for areas with unavailable data we could potentially replace this with n a or just leave it blank
1
698,195
23,969,501,198
IssuesEvent
2022-09-13 06:21:25
JGuetschow/UNFCCC_non-AnnexI_data
https://api.github.com/repos/JGuetschow/UNFCCC_non-AnnexI_data
closed
Republic of Korea BUR4 / NIR 2020
Priority high BUR Needs review
Integrate data already read from NIR xlsx file (non-UNFCCC folder) and adapt code to publish it here.
1.0
Republic of Korea BUR4 / NIR 2020 - Integrate data already read from NIR xlsx file (non-UNFCCC folder) and adapt code to publish it here.
priority
republic of korea nir integrate data already read from nir xlsx file non unfccc folder and adapt code to publish it here
1
785,685
27,622,634,165
IssuesEvent
2023-03-10 02:19:33
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
closed
[BUG] Janitor can use holosign to force doors open.
Type: Bug Priority: High
### There is no other issue reporting the same problem. - [X] I have searched the existing issues ### Game Version 23022319 ### I was Client ### OS Windows 8, 10, 11 ### What happened? By trying to open a door you don't have access to you can force it open using a janitor holosign. ### What should have happened instead? Holosign should not force doors open. ### Steps to reproduce the issue 1. Get the janitor holo sign maker thing 2. Spam tries to open door while space clicking the holosign maker 3. Opens door ### Additional context only seen someone else do this dunno how it really works
1.0
[BUG] Janitor can use holosign to force doors open. - ### There is no other issue reporting the same problem. - [X] I have searched the existing issues ### Game Version 23022319 ### I was Client ### OS Windows 8, 10, 11 ### What happened? By trying to open a door you don't have access to you can force it open using a janitor holosign. ### What should have happened instead? Holosign should not force doors open. ### Steps to reproduce the issue 1. Get the janitor holo sign maker thing 2. Spam tries to open door while space clicking the holosign maker 3. Opens door ### Additional context only seen someone else do this dunno how it really works
priority
janitor can use holosign to force doors open there is no other issue reporting the same problem i have searched the existing issues game version i was client os windows what happened by trying to open a door you don t have access to you can force it open using a janitor holosign what should have happened instead holosign should not force doors open steps to reproduce the issue get the janitor holo sign maker thing spam tries to open door while space clicking the holosign maker opens door additional context only seen someone else do this dunno how it really works
1
399,227
11,744,954,765
IssuesEvent
2020-03-12 08:53:53
level73/membernet
https://api.github.com/repos/level73/membernet
opened
ordering of contacts on member profiles
Platform: Membernet Priority: High Type: Development
please ensure that the focal point is always the first visible contact (now i believe it sometimes puts alternate fp first). the need is also arising for the website, so they only pull the first contact on member profiles on their end. in the case that there are two fps, it would be good if it could the newest one (as there is more chance that thats updated) or alphabetical ordering. on member profiles, it would be good if all the special roles get listed first: fp, alt. fp, admin/comms focal point, etc...
1.0
ordering of contacts on member profiles - please ensure that the focal point is always the first visible contact (now i believe it sometimes puts alternate fp first). the need is also arising for the website, so they only pull the first contact on member profiles on their end. in the case that there are two fps, it would be good if it could the newest one (as there is more chance that thats updated) or alphabetical ordering. on member profiles, it would be good if all the special roles get listed first: fp, alt. fp, admin/comms focal point, etc...
priority
ordering of contacts on member profiles please ensure that the focal point is always the first visible contact now i believe it sometimes puts alternate fp first the need is also arising for the website so they only pull the first contact on member profiles on their end in the case that there are two fps it would be good if it could the newest one as there is more chance that thats updated or alphabetical ordering on member profiles it would be good if all the special roles get listed first fp alt fp admin comms focal point etc
1
825,994
31,494,735,379
IssuesEvent
2023-08-31 00:44:04
geopm/geopm
https://api.github.com/repos/geopm/geopm
closed
Build requirements mention a package with the wrong capitalization
bug bug-priority-high bug-exposure-high bug-quality-high
**Describe the bug** I tried to paste the `yum install` instructions from https://geopm.github.io/requires.html. The name of `python3-Sphinx` should be `python3-sphinx` instead. **GEOPM version** Online documentation in August 2023. **Expected behavior** Dependency installation instructions can be pasted into a shell session. **Actual behavior** The sphinx dependency has the wrong capitalization.
1.0
Build requirements mention a package with the wrong capitalization - **Describe the bug** I tried to paste the `yum install` instructions from https://geopm.github.io/requires.html. The name of `python3-Sphinx` should be `python3-sphinx` instead. **GEOPM version** Online documentation in August 2023. **Expected behavior** Dependency installation instructions can be pasted into a shell session. **Actual behavior** The sphinx dependency has the wrong capitalization.
priority
build requirements mention a package with the wrong capitalization describe the bug i tried to paste the yum install instructions from the name of sphinx should be sphinx instead geopm version online documentation in august expected behavior dependency installation instructions can be pasted into a shell session actual behavior the sphinx dependency has the wrong capitalization
1
694,731
23,827,188,860
IssuesEvent
2022-09-05 15:51:59
python/mypy
https://api.github.com/repos/python/mypy
closed
internal error when inferring type of callable resulting from untyped ternary expression
crash priority-0-high topic-ternary-expression
Consider the following example: ``` import random from typing import Union, Callable class Wrapper: def __init__(self, x): self.x = x def main() -> Union[Wrapper, int]: wrap = Wrapper if random.random() < 0.5 else lambda x: x return wrap(2) ``` Running mypy (from the master branch in the repo) on it crashes with the following stack trace: ``` mypytest.py:18: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.rtfd.io/en/latest/common_issues.html#using-a-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.790+dev.cc46692020ef6b672b761512095fbba281fdefc7 Traceback (most recent call last): File "/home/edo/miniconda3/envs/abm2/bin/mypy", line 8, in <module> sys.exit(console_entry()) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/__main__.py", line 8, in console_entry main(None, sys.stdout, sys.stderr) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/main.py", line 89, in main res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 180, in build result = _build( File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 252, in _build graph = dispatch(sources, manager, stdout) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2626, in dispatch process_graph(graph, manager) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2949, in process_graph process_stale_scc(graph, scc, manager) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 3047, in process_stale_scc graph[id].type_check_first_pass() File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2107, in type_check_first_pass self.type_checker().check_first_pass() File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 294, in check_first_pass self.accept(d) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 677, in accept return visitor.visit_func_def(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 726, in visit_func_def self._visit_func_def(defn) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 730, in _visit_func_def self.check_func_item(defn, name=defn.name) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 792, in check_func_item self.check_func_def(defn, typ, name) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 975, in check_func_def self.accept(item.body) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1005, in accept return visitor.visit_block(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 1973, in visit_block self.accept(s) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1141, in accept return visitor.visit_return_stmt(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 3156, in visit_return_stmt self.check_return_stmt(s) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 3188, in check_return_stmt typ = get_proper_type(self.expr_checker.accept( File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 3768, in accept typ = node.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1545, in accept return visitor.visit_call_expr(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 263, in visit_call_expr return self.visit_call_expr_inner(e, allow_none_return=allow_none_return) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 340, in visit_call_expr_inner ret_type = self.check_call_expr_with_callee_type(callee_type, e, fullname, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 817, in check_call_expr_with_callee_type return self.check_call(callee_type, e.args, e.arg_kinds, e, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 876, in check_call return self.check_callable_call(callee, args, arg_kinds, context, arg_names, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 941, in check_callable_call if (callee.is_type_obj() and callee.type_object().is_abstract File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/types.py", line 1094, in type_object assert isinstance(ret, Instance) AssertionError: mypytest.py:18: : note: use --pdb to drop into pdb ``` Where `ret` is `Any` and `callee` in the frame above is `def (x: Any) -> Any`. The error can be prevented either by assigning the lambda to a variable ``` identity = lambda x: x wrap = Wrapper if random.random() < 0.5 else identity ``` where in this case no issues are reported, or the wrapper: ``` class Wrapper: def __init__(self, x: int) -> None: self.x = x def main() -> Union[Wrapper, int]: wrap = Wrapper if random.random() < 0.5 else lambda x: x ``` but in this case two errors are reported (line 12 is `wrap = ` and line 14 is `return` ``` mypytest.py:12: error: Incompatible return value type (got "int", expected "Wrapper") mypytest.py:14: error: Incompatible return value type (got "object", expected "Union[Wrapper, int]") ``` these erros do not occur when both fixes are applied. expected behavior: no errors reported in all three cases, or at least no crash.
1.0
internal error when inferring type of callable resulting from untyped ternary expression - Consider the following example: ``` import random from typing import Union, Callable class Wrapper: def __init__(self, x): self.x = x def main() -> Union[Wrapper, int]: wrap = Wrapper if random.random() < 0.5 else lambda x: x return wrap(2) ``` Running mypy (from the master branch in the repo) on it crashes with the following stack trace: ``` mypytest.py:18: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.rtfd.io/en/latest/common_issues.html#using-a-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.790+dev.cc46692020ef6b672b761512095fbba281fdefc7 Traceback (most recent call last): File "/home/edo/miniconda3/envs/abm2/bin/mypy", line 8, in <module> sys.exit(console_entry()) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/__main__.py", line 8, in console_entry main(None, sys.stdout, sys.stderr) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/main.py", line 89, in main res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 180, in build result = _build( File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 252, in _build graph = dispatch(sources, manager, stdout) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2626, in dispatch process_graph(graph, manager) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2949, in process_graph process_stale_scc(graph, scc, manager) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 3047, in process_stale_scc graph[id].type_check_first_pass() File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/build.py", line 2107, in type_check_first_pass self.type_checker().check_first_pass() File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 294, in check_first_pass self.accept(d) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 677, in accept return visitor.visit_func_def(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 726, in visit_func_def self._visit_func_def(defn) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 730, in _visit_func_def self.check_func_item(defn, name=defn.name) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 792, in check_func_item self.check_func_def(defn, typ, name) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 975, in check_func_def self.accept(item.body) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1005, in accept return visitor.visit_block(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 1973, in visit_block self.accept(s) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 401, in accept stmt.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1141, in accept return visitor.visit_return_stmt(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 3156, in visit_return_stmt self.check_return_stmt(s) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checker.py", line 3188, in check_return_stmt typ = get_proper_type(self.expr_checker.accept( File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 3768, in accept typ = node.accept(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/nodes.py", line 1545, in accept return visitor.visit_call_expr(self) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 263, in visit_call_expr return self.visit_call_expr_inner(e, allow_none_return=allow_none_return) File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 340, in visit_call_expr_inner ret_type = self.check_call_expr_with_callee_type(callee_type, e, fullname, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 817, in check_call_expr_with_callee_type return self.check_call(callee_type, e.args, e.arg_kinds, e, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 876, in check_call return self.check_callable_call(callee, args, arg_kinds, context, arg_names, File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/checkexpr.py", line 941, in check_callable_call if (callee.is_type_obj() and callee.type_object().is_abstract File "/home/edo/miniconda3/envs/abm2/lib/python3.8/site-packages/mypy/types.py", line 1094, in type_object assert isinstance(ret, Instance) AssertionError: mypytest.py:18: : note: use --pdb to drop into pdb ``` Where `ret` is `Any` and `callee` in the frame above is `def (x: Any) -> Any`. The error can be prevented either by assigning the lambda to a variable ``` identity = lambda x: x wrap = Wrapper if random.random() < 0.5 else identity ``` where in this case no issues are reported, or the wrapper: ``` class Wrapper: def __init__(self, x: int) -> None: self.x = x def main() -> Union[Wrapper, int]: wrap = Wrapper if random.random() < 0.5 else lambda x: x ``` but in this case two errors are reported (line 12 is `wrap = ` and line 14 is `return` ``` mypytest.py:12: error: Incompatible return value type (got "int", expected "Wrapper") mypytest.py:14: error: Incompatible return value type (got "object", expected "Union[Wrapper, int]") ``` these erros do not occur when both fixes are applied. expected behavior: no errors reported in all three cases, or at least no crash.
priority
internal error when inferring type of callable resulting from untyped ternary expression consider the following example import random from typing import union callable class wrapper def init self x self x x def main union wrap wrapper if random random else lambda x x return wrap running mypy from the master branch in the repo on it crashes with the following stack trace mypytest py error internal error please try using mypy master on github please report a bug at version dev traceback most recent call last file home edo envs bin mypy line in sys exit console entry file home edo envs lib site packages mypy main py line in console entry main none sys stdout sys stderr file home edo envs lib site packages mypy main py line in main res build build sources options none flush errors fscache stdout stderr file home edo envs lib site packages mypy build py line in build result build file home edo envs lib site packages mypy build py line in build graph dispatch sources manager stdout file home edo envs lib site packages mypy build py line in dispatch process graph graph manager file home edo envs lib site packages mypy build py line in process graph process stale scc graph scc manager file home edo envs lib site packages mypy build py line in process stale scc graph type check first pass file home edo envs lib site packages mypy build py line in type check first pass self type checker check first pass file home edo envs lib site packages mypy checker py line in check first pass self accept d file home edo envs lib site packages mypy checker py line in accept stmt accept self file home edo envs lib site packages mypy nodes py line in accept return visitor visit func def self file home edo envs lib site packages mypy checker py line in visit func def self visit func def defn file home edo envs lib site packages mypy checker py line in visit func def self check func item defn name defn name file home edo envs lib site packages mypy checker py line in check func item self check func def defn typ name file home edo envs lib site packages mypy checker py line in check func def self accept item body file home edo envs lib site packages mypy checker py line in accept stmt accept self file home edo envs lib site packages mypy nodes py line in accept return visitor visit block self file home edo envs lib site packages mypy checker py line in visit block self accept s file home edo envs lib site packages mypy checker py line in accept stmt accept self file home edo envs lib site packages mypy nodes py line in accept return visitor visit return stmt self file home edo envs lib site packages mypy checker py line in visit return stmt self check return stmt s file home edo envs lib site packages mypy checker py line in check return stmt typ get proper type self expr checker accept file home edo envs lib site packages mypy checkexpr py line in accept typ node accept self file home edo envs lib site packages mypy nodes py line in accept return visitor visit call expr self file home edo envs lib site packages mypy checkexpr py line in visit call expr return self visit call expr inner e allow none return allow none return file home edo envs lib site packages mypy checkexpr py line in visit call expr inner ret type self check call expr with callee type callee type e fullname file home edo envs lib site packages mypy checkexpr py line in check call expr with callee type return self check call callee type e args e arg kinds e file home edo envs lib site packages mypy checkexpr py line in check call return self check callable call callee args arg kinds context arg names file home edo envs lib site packages mypy checkexpr py line in check callable call if callee is type obj and callee type object is abstract file home edo envs lib site packages mypy types py line in type object assert isinstance ret instance assertionerror mypytest py note use pdb to drop into pdb where ret is any and callee in the frame above is def x any any the error can be prevented either by assigning the lambda to a variable identity lambda x x wrap wrapper if random random else identity where in this case no issues are reported or the wrapper class wrapper def init self x int none self x x def main union wrap wrapper if random random else lambda x x but in this case two errors are reported line is wrap and line is return mypytest py error incompatible return value type got int expected wrapper mypytest py error incompatible return value type got object expected union these erros do not occur when both fixes are applied expected behavior no errors reported in all three cases or at least no crash
1
404,565
11,859,282,187
IssuesEvent
2020-03-25 13:06:37
wso2/docker-apim
https://api.github.com/repos/wso2/docker-apim
closed
Use WSO2 product pack downloadable links to binaries available at GitHub release pages
Priority/High Type/Task
**Description:** Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds. It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default. **Affected Product Version:** Docker resources for WSO2 API Management v3.1.0 and beyond **Sub Tasks:** - [x] Integrate to Alpine based Docker resources - [x] Integrate to CentOS based Docker resources - [x] Integrate to Ubuntu based Docker resources - [x] Code/Peer review
1.0
Use WSO2 product pack downloadable links to binaries available at GitHub release pages - **Description:** Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds. It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default. **Affected Product Version:** Docker resources for WSO2 API Management v3.1.0 and beyond **Sub Tasks:** - [x] Integrate to Alpine based Docker resources - [x] Integrate to CentOS based Docker resources - [x] Integrate to Ubuntu based Docker resources - [x] Code/Peer review
priority
use product pack downloadable links to binaries available at github release pages description currently product binaries available at jfrog bintray are used for the docker image builds it has been suggested to use product pack downloadable links to binaries available at github product release pages as the default affected product version docker resources for api management and beyond sub tasks integrate to alpine based docker resources integrate to centos based docker resources integrate to ubuntu based docker resources code peer review
1
564,582
16,736,462,834
IssuesEvent
2021-06-11 02:36:48
1Hive/honeyswap
https://api.github.com/repos/1Hive/honeyswap
opened
[POLYGON] Switch blockchain explorer links to PolygonScan
Frontend Polygon good first issue priority: HIGH
## User story As a Honeyswap user, I would like blockchain explorer links to send to PolygonScan, so that I can see transactions and accounts on the support explorer ## Acceptance Criteria GIVEN I am on the Honeyswap charts page AND I have selected the Polygon network from the dropdown WHEN I click an account THEN I am connected to the account page on PolygonScan explorer GIVEN I am on the Honeyswap charts page AND I have selected the Polygon network from the dropdown WHEN I click a transaction THEN I am connected to the transaction page on PolygonScan explorer ## Notes * MaticVigil is being deprecated and replaced by PolygonScan.com
1.0
[POLYGON] Switch blockchain explorer links to PolygonScan - ## User story As a Honeyswap user, I would like blockchain explorer links to send to PolygonScan, so that I can see transactions and accounts on the support explorer ## Acceptance Criteria GIVEN I am on the Honeyswap charts page AND I have selected the Polygon network from the dropdown WHEN I click an account THEN I am connected to the account page on PolygonScan explorer GIVEN I am on the Honeyswap charts page AND I have selected the Polygon network from the dropdown WHEN I click a transaction THEN I am connected to the transaction page on PolygonScan explorer ## Notes * MaticVigil is being deprecated and replaced by PolygonScan.com
priority
switch blockchain explorer links to polygonscan user story as a honeyswap user i would like blockchain explorer links to send to polygonscan so that i can see transactions and accounts on the support explorer acceptance criteria given i am on the honeyswap charts page and i have selected the polygon network from the dropdown when i click an account then i am connected to the account page on polygonscan explorer given i am on the honeyswap charts page and i have selected the polygon network from the dropdown when i click a transaction then i am connected to the transaction page on polygonscan explorer notes maticvigil is being deprecated and replaced by polygonscan com
1
694,189
23,805,864,195
IssuesEvent
2022-09-04 02:36:57
uchikoshi-fes/uchikoshi-fes.jp-2022
https://api.github.com/repos/uchikoshi-fes/uchikoshi-fes.jp-2022
closed
[要望]: トップページを充実させる
enhancement priority: HIGH
### リクエスト - [x] ファーストビュー - [x] アクセス (事前予約ページへのリンクを含む) - [x] スローガンについて - [x] PV タブ - [x] アサノラジオ局 随時追加予定
1.0
[要望]: トップページを充実させる - ### リクエスト - [x] ファーストビュー - [x] アクセス (事前予約ページへのリンクを含む) - [x] スローガンについて - [x] PV タブ - [x] アサノラジオ局 随時追加予定
priority
トップページを充実させる リクエスト ファーストビュー アクセス 事前予約ページへのリンクを含む スローガンについて pv タブ アサノラジオ局 随時追加予定
1
51,867
3,014,962,672
IssuesEvent
2015-07-29 17:14:20
meetinghouse/cms
https://api.github.com/repos/meetinghouse/cms
closed
Dark Theme: Portfolio page has blank spaces
High Priority
On the main portfolio page, there are apparently random blank spaces. We haven't noticed these before and don't know why they are showing up. We're not asking about the place-holder project where there is no image (because we know those need to have images uploaded). The concern is the completely empty spaces. Here is a screenshot: ![screenshot 2015-07-28 at 2 18 12 pm](https://cloud.githubusercontent.com/assets/5215893/8939626/8097bcfe-3533-11e5-9ad9-b431dd1d277a.png)
1.0
Dark Theme: Portfolio page has blank spaces - On the main portfolio page, there are apparently random blank spaces. We haven't noticed these before and don't know why they are showing up. We're not asking about the place-holder project where there is no image (because we know those need to have images uploaded). The concern is the completely empty spaces. Here is a screenshot: ![screenshot 2015-07-28 at 2 18 12 pm](https://cloud.githubusercontent.com/assets/5215893/8939626/8097bcfe-3533-11e5-9ad9-b431dd1d277a.png)
priority
dark theme portfolio page has blank spaces on the main portfolio page there are apparently random blank spaces we haven t noticed these before and don t know why they are showing up we re not asking about the place holder project where there is no image because we know those need to have images uploaded the concern is the completely empty spaces here is a screenshot
1
342,091
10,311,761,289
IssuesEvent
2019-08-29 18:09:41
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Inconsistent model behavior when loading model using load_state_dict in different ways
high priority module: random module: serialization triaged
## 🐛 Bug <!-- A clear and concise description of what the bug is. --> I have a model which has two embedding layers. And I initialize the model, train it on GPU, and save its state dict to disk using `torch.save(model.state_dict, 'tmp.pt')`. And now I want to use previous model to do inference jobs, I find two ways below can cause different results: 1. initialize model on cpu -> load state dict -> model to gpu 2. initialize model on cpu -> model to gpu -> load state dict **However, using just one embedding layer in my model will solve this.** ## To Reproduce Run the code below, and it does these things: 1. fix random seeds 2. generate fake embedding weights and samples 3. initialize one model, train it, save it on disk 4. three different ways of doing inference, and the first two asserts will fail ```python import torch import torch.nn as nn import torch.optim as optim def seed_torch(seed): import random random.seed(seed) import os os.environ['PYTHONHASHSEED'] = str(seed) import numpy as np np.random.seed(seed) torch.manual_seed(seed) torch.cuda.manual_seed(seed) torch.cuda.manual_seed_all(seed) # if you are using multi-GPU. torch.backends.cudnn.benchmark = False torch.backends.cudnn.deterministic = True seed_torch(2019) class Model(nn.Module): def __init__(self, vectors): super(Model, self).__init__() self.emb = nn.Embedding.from_pretrained( vectors, freeze=False, padding_idx=0 ) # dummy embedding layers, just initialized and not used self.emb2 = nn.Embedding.from_pretrained( vectors, freeze=False, padding_idx=0 ) self.linear = nn.Linear(128, 2) def forward(self, s): s = self.emb(s) s = self.linear(s) s = s.sum(dim=1) return s vectors = torch.randn(1000, 128) # fake samples train_s = torch.randint(1, 1000, size=(100, 30)).to('cuda') train_y = torch.ones(100).to(dtype=torch.long).to('cuda') # train model = Model(vectors).to('cuda') criterion = nn.CrossEntropyLoss(reduction='sum') optimizer = torch.optim.SGD(model.parameters(), lr=1e-3) model.train() model.zero_grad() out = model(train_s) loss = criterion(out, train_y) loss.backward() optimizer.step() torch.save(model.state_dict(), 'tmp.pt') # inference, model init -> to cuda -> load state dict m1 = Model(vectors).to('cuda') m1.load_state_dict(torch.load('tmp.pt')) m1.eval() with torch.no_grad(): o1 = m1(train_s.to('cuda')) # inference, model init -> load state dit -> to cuda m2 = Model(vectors) m2.load_state_dict(torch.load('tmp.pt')) m2 = m2.to('cuda') m2.eval() with torch.no_grad(): o2 = m2(train_s.to('cuda')) # inference on cpu m3 = Model(vectors) m3.load_state_dict(torch.load('tmp.pt')) m3.eval() with torch.no_grad(): o3 = m3(train_s.to('cpu')) assert torch.allclose(o1, o2) assert torch.allclose(o1.cpu(), o3) assert torch.allclose(o2.cpu(), o3) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior The assert in above code should all pass. <!-- A clear and concise description of what you expected to happen. --> ## Environment PyTorch version: 1.1.0 Is debug build: No CUDA used to build PyTorch: 10.0.130 OS: Ubuntu 16.04.6 LTS GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.11) 5.4.0 20160609 CMake version: version 3.5.1 Python version: 3.6 Is CUDA available: Yes CUDA runtime version: 10.0.130 GPU models and configuration: GPU 0: Tesla K80 Nvidia driver version: 410.104 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.4.2 Versions of relevant libraries: [pip] numpy==1.16.2 [pip] pytorch-pretrained-bert==0.6.2 [pip] torch==1.1.0 [pip] torchvision==0.3.0 [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.0.2 py36h7b6447c_0 [conda] mkl_fft 1.0.12 py36ha843d7b_0 [conda] mkl_random 1.0.2 py36hd81dba3_0 [conda] pytorch 1.1.0 py3.6_cuda10.0.130_cudnn7.5.1_0 pytorch [conda] torchvision 0.3.0 py36_cu10.0.130_1 pytorch ## Additional context <!-- Add any other context about the problem here. -->
1.0
Inconsistent model behavior when loading model using load_state_dict in different ways - ## 🐛 Bug <!-- A clear and concise description of what the bug is. --> I have a model which has two embedding layers. And I initialize the model, train it on GPU, and save its state dict to disk using `torch.save(model.state_dict, 'tmp.pt')`. And now I want to use previous model to do inference jobs, I find two ways below can cause different results: 1. initialize model on cpu -> load state dict -> model to gpu 2. initialize model on cpu -> model to gpu -> load state dict **However, using just one embedding layer in my model will solve this.** ## To Reproduce Run the code below, and it does these things: 1. fix random seeds 2. generate fake embedding weights and samples 3. initialize one model, train it, save it on disk 4. three different ways of doing inference, and the first two asserts will fail ```python import torch import torch.nn as nn import torch.optim as optim def seed_torch(seed): import random random.seed(seed) import os os.environ['PYTHONHASHSEED'] = str(seed) import numpy as np np.random.seed(seed) torch.manual_seed(seed) torch.cuda.manual_seed(seed) torch.cuda.manual_seed_all(seed) # if you are using multi-GPU. torch.backends.cudnn.benchmark = False torch.backends.cudnn.deterministic = True seed_torch(2019) class Model(nn.Module): def __init__(self, vectors): super(Model, self).__init__() self.emb = nn.Embedding.from_pretrained( vectors, freeze=False, padding_idx=0 ) # dummy embedding layers, just initialized and not used self.emb2 = nn.Embedding.from_pretrained( vectors, freeze=False, padding_idx=0 ) self.linear = nn.Linear(128, 2) def forward(self, s): s = self.emb(s) s = self.linear(s) s = s.sum(dim=1) return s vectors = torch.randn(1000, 128) # fake samples train_s = torch.randint(1, 1000, size=(100, 30)).to('cuda') train_y = torch.ones(100).to(dtype=torch.long).to('cuda') # train model = Model(vectors).to('cuda') criterion = nn.CrossEntropyLoss(reduction='sum') optimizer = torch.optim.SGD(model.parameters(), lr=1e-3) model.train() model.zero_grad() out = model(train_s) loss = criterion(out, train_y) loss.backward() optimizer.step() torch.save(model.state_dict(), 'tmp.pt') # inference, model init -> to cuda -> load state dict m1 = Model(vectors).to('cuda') m1.load_state_dict(torch.load('tmp.pt')) m1.eval() with torch.no_grad(): o1 = m1(train_s.to('cuda')) # inference, model init -> load state dit -> to cuda m2 = Model(vectors) m2.load_state_dict(torch.load('tmp.pt')) m2 = m2.to('cuda') m2.eval() with torch.no_grad(): o2 = m2(train_s.to('cuda')) # inference on cpu m3 = Model(vectors) m3.load_state_dict(torch.load('tmp.pt')) m3.eval() with torch.no_grad(): o3 = m3(train_s.to('cpu')) assert torch.allclose(o1, o2) assert torch.allclose(o1.cpu(), o3) assert torch.allclose(o2.cpu(), o3) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior The assert in above code should all pass. <!-- A clear and concise description of what you expected to happen. --> ## Environment PyTorch version: 1.1.0 Is debug build: No CUDA used to build PyTorch: 10.0.130 OS: Ubuntu 16.04.6 LTS GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.11) 5.4.0 20160609 CMake version: version 3.5.1 Python version: 3.6 Is CUDA available: Yes CUDA runtime version: 10.0.130 GPU models and configuration: GPU 0: Tesla K80 Nvidia driver version: 410.104 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.4.2 Versions of relevant libraries: [pip] numpy==1.16.2 [pip] pytorch-pretrained-bert==0.6.2 [pip] torch==1.1.0 [pip] torchvision==0.3.0 [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.0.2 py36h7b6447c_0 [conda] mkl_fft 1.0.12 py36ha843d7b_0 [conda] mkl_random 1.0.2 py36hd81dba3_0 [conda] pytorch 1.1.0 py3.6_cuda10.0.130_cudnn7.5.1_0 pytorch [conda] torchvision 0.3.0 py36_cu10.0.130_1 pytorch ## Additional context <!-- Add any other context about the problem here. -->
priority
inconsistent model behavior when loading model using load state dict in different ways 🐛 bug i have a model which has two embedding layers and i initialize the model train it on gpu and save its state dict to disk using torch save model state dict tmp pt and now i want to use previous model to do inference jobs i find two ways below can cause different results initialize model on cpu load state dict model to gpu initialize model on cpu model to gpu load state dict however using just one embedding layer in my model will solve this to reproduce run the code below and it does these things fix random seeds generate fake embedding weights and samples initialize one model train it save it on disk three different ways of doing inference and the first two asserts will fail python import torch import torch nn as nn import torch optim as optim def seed torch seed import random random seed seed import os os environ str seed import numpy as np np random seed seed torch manual seed seed torch cuda manual seed seed torch cuda manual seed all seed if you are using multi gpu torch backends cudnn benchmark false torch backends cudnn deterministic true seed torch class model nn module def init self vectors super model self init self emb nn embedding from pretrained vectors freeze false padding idx dummy embedding layers just initialized and not used self nn embedding from pretrained vectors freeze false padding idx self linear nn linear def forward self s s self emb s s self linear s s s sum dim return s vectors torch randn fake samples train s torch randint size to cuda train y torch ones to dtype torch long to cuda train model model vectors to cuda criterion nn crossentropyloss reduction sum optimizer torch optim sgd model parameters lr model train model zero grad out model train s loss criterion out train y loss backward optimizer step torch save model state dict tmp pt inference model init to cuda load state dict model vectors to cuda load state dict torch load tmp pt eval with torch no grad train s to cuda inference model init load state dit to cuda model vectors load state dict torch load tmp pt to cuda eval with torch no grad train s to cuda inference on cpu model vectors load state dict torch load tmp pt eval with torch no grad train s to cpu assert torch allclose assert torch allclose cpu assert torch allclose cpu expected behavior the assert in above code should all pass environment pytorch version is debug build no cuda used to build pytorch os ubuntu lts gcc version ubuntu cmake version version python version is cuda available yes cuda runtime version gpu models and configuration gpu tesla nvidia driver version cudnn version usr lib linux gnu libcudnn so versions of relevant libraries numpy pytorch pretrained bert torch torchvision blas mkl mkl mkl service mkl fft mkl random pytorch pytorch torchvision pytorch additional context
1
283,910
8,725,553,208
IssuesEvent
2018-12-10 09:44:38
playasoft/volunteers
https://api.github.com/repos/playasoft/volunteers
closed
Adding shift leads to departments
feature request help wanted major change priority: high
In the current Apogaea Volunteer Database, it's possible to add shift leads to a department. These leads then have access to change the options of the department, create new shifts, and kick users from slots. This would require some pretty major changes in laravel-voldb's system for handling roles.
1.0
Adding shift leads to departments - In the current Apogaea Volunteer Database, it's possible to add shift leads to a department. These leads then have access to change the options of the department, create new shifts, and kick users from slots. This would require some pretty major changes in laravel-voldb's system for handling roles.
priority
adding shift leads to departments in the current apogaea volunteer database it s possible to add shift leads to a department these leads then have access to change the options of the department create new shifts and kick users from slots this would require some pretty major changes in laravel voldb s system for handling roles
1
216,270
7,302,713,637
IssuesEvent
2018-02-27 10:35:17
crowdAI/crowdai
https://api.github.com/repos/crowdAI/crowdai
closed
Submission initiated status
high priority
As some NIPS submissions are taking ~45 minutes to upload the model, the grading process should be updated so that an API call is made at the beginning of the submission process, with a status of 'initiated' to correctly record the timestamp of the submission.
1.0
Submission initiated status - As some NIPS submissions are taking ~45 minutes to upload the model, the grading process should be updated so that an API call is made at the beginning of the submission process, with a status of 'initiated' to correctly record the timestamp of the submission.
priority
submission initiated status as some nips submissions are taking minutes to upload the model the grading process should be updated so that an api call is made at the beginning of the submission process with a status of initiated to correctly record the timestamp of the submission
1
23,350
2,658,443,438
IssuesEvent
2015-03-18 15:39:13
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
closed
Geoconnect: Error when clicking map it button when logged in but no edit perms.
Priority: High Status: QA Type: Bug
When I am logged in as a registered user but do not have edit perms on the dataset, I click map it on a shape file and see: {"status":"ERROR","message":"The user does not have permission to edit metadata for this file."}
1.0
Geoconnect: Error when clicking map it button when logged in but no edit perms. - When I am logged in as a registered user but do not have edit perms on the dataset, I click map it on a shape file and see: {"status":"ERROR","message":"The user does not have permission to edit metadata for this file."}
priority
geoconnect error when clicking map it button when logged in but no edit perms when i am logged in as a registered user but do not have edit perms on the dataset i click map it on a shape file and see status error message the user does not have permission to edit metadata for this file
1
69,369
3,297,853,878
IssuesEvent
2015-11-02 11:17:56
Maslosoft/Mangan
https://api.github.com/repos/Maslosoft/Mangan
opened
Create criteria interface
high-priority improvement refactor
To allow custom criteria classes there should be interface (s) for criteria related operations. Best use case is a Conditions class which is in fact different approach to definining criteria. All classes accepting Criteria must now accept CriteriaInterface. TODO: * Consider what should be required for such interface.
1.0
Create criteria interface - To allow custom criteria classes there should be interface (s) for criteria related operations. Best use case is a Conditions class which is in fact different approach to definining criteria. All classes accepting Criteria must now accept CriteriaInterface. TODO: * Consider what should be required for such interface.
priority
create criteria interface to allow custom criteria classes there should be interface s for criteria related operations best use case is a conditions class which is in fact different approach to definining criteria all classes accepting criteria must now accept criteriainterface todo consider what should be required for such interface
1
2,459
2,525,862,607
IssuesEvent
2015-01-21 06:55:22
graybeal/ont
https://api.github.com/repos/graybeal/ont
closed
Original metadata values shouldn't be changed for unversioned response
1 star bug imported ont Priority-High
_From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on May 22, 2009 19:34:30_ For the unversioned response form of any registered ontology a few metadata attributes are changed including: * omv:name (full title in UI): If the original title is for example "Argo QA/QC Flags", this get converted to "Unversioned form of: Argo QA/QC Flags" in the unversioned response. * omv:description (brief description in UI): "QA/QC flags used in Argo" gets converted to "An unversioned ontology containing the latest terms as of the request time, for the ontology containing: QA/QC flags used in Argo". These attributes should not be changed. If we want to make clear that it is an unversioned form, perhaps we can include a (new) special metadata attribute, say, omvmmi:versionAnnotation. The point is to avoid any alteration to original metadata values while avoiding unnecessary confusion. _Original issue: http://code.google.com/p/mmisw/issues/detail?id=138_
1.0
Original metadata values shouldn't be changed for unversioned response - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on May 22, 2009 19:34:30_ For the unversioned response form of any registered ontology a few metadata attributes are changed including: * omv:name (full title in UI): If the original title is for example "Argo QA/QC Flags", this get converted to "Unversioned form of: Argo QA/QC Flags" in the unversioned response. * omv:description (brief description in UI): "QA/QC flags used in Argo" gets converted to "An unversioned ontology containing the latest terms as of the request time, for the ontology containing: QA/QC flags used in Argo". These attributes should not be changed. If we want to make clear that it is an unversioned form, perhaps we can include a (new) special metadata attribute, say, omvmmi:versionAnnotation. The point is to avoid any alteration to original metadata values while avoiding unnecessary confusion. _Original issue: http://code.google.com/p/mmisw/issues/detail?id=138_
priority
original metadata values shouldn t be changed for unversioned response from on may for the unversioned response form of any registered ontology a few metadata attributes are changed including omv name full title in ui if the original title is for example argo qa qc flags this get converted to unversioned form of argo qa qc flags in the unversioned response omv description brief description in ui qa qc flags used in argo gets converted to an unversioned ontology containing the latest terms as of the request time for the ontology containing qa qc flags used in argo these attributes should not be changed if we want to make clear that it is an unversioned form perhaps we can include a new special metadata attribute say omvmmi versionannotation the point is to avoid any alteration to original metadata values while avoiding unnecessary confusion original issue
1
383,720
11,361,757,320
IssuesEvent
2020-01-26 16:55:25
deep-learning-indaba/Baobab
https://api.github.com/repos/deep-learning-indaba/Baobab
closed
Link user profiles to organisations
High Priority back-end enhancement
Depends on #448 As per the [multi-tenant spec](https://docs.google.com/document/d/19Wr5rPT-zzJ18h3AC0dKQYAJpO3zWLThs6Kku5NoTl8/edit#heading=h.v07jx0wja9vt), User profiles will now be specific to organisations. We need to make the following changes to implement this: - Add a foreign key reference from AppUser to Organisation. This should be non-nullable and existing data can be populated with the "Deep Learning Indaba" organisation - Update the POST method in the user controller to set the organisation to the current organisation - Update the AuthenticationAPI to filter user profiles by the current organisation - Check whether any other controllers in api/app/users/api.py need to be updated - Search for any places in the code that query the AppUser model using "email address" as a surrogate key, and change it to use email address + organisation_id
1.0
Link user profiles to organisations - Depends on #448 As per the [multi-tenant spec](https://docs.google.com/document/d/19Wr5rPT-zzJ18h3AC0dKQYAJpO3zWLThs6Kku5NoTl8/edit#heading=h.v07jx0wja9vt), User profiles will now be specific to organisations. We need to make the following changes to implement this: - Add a foreign key reference from AppUser to Organisation. This should be non-nullable and existing data can be populated with the "Deep Learning Indaba" organisation - Update the POST method in the user controller to set the organisation to the current organisation - Update the AuthenticationAPI to filter user profiles by the current organisation - Check whether any other controllers in api/app/users/api.py need to be updated - Search for any places in the code that query the AppUser model using "email address" as a surrogate key, and change it to use email address + organisation_id
priority
link user profiles to organisations depends on as per the user profiles will now be specific to organisations we need to make the following changes to implement this add a foreign key reference from appuser to organisation this should be non nullable and existing data can be populated with the deep learning indaba organisation update the post method in the user controller to set the organisation to the current organisation update the authenticationapi to filter user profiles by the current organisation check whether any other controllers in api app users api py need to be updated search for any places in the code that query the appuser model using email address as a surrogate key and change it to use email address organisation id
1
676,017
23,113,525,864
IssuesEvent
2022-07-27 14:48:52
NIAEFEUP/tts-revamp-fe
https://api.github.com/repos/NIAEFEUP/tts-revamp-fe
opened
Lesson conflict update when hiding individual courses
medium priority high effort
When a hide individual class box is ticked, conflicts should be updated: run the conflicts algorithm once again, not considering the hidden classes. (refreshHiddenLessons @ LessonBox.tsx) Suggestion: change CheckedCourse type to this ```ts export type CheckedCourse = { checked: boolean hidden: boolean // new info: Course } ```
1.0
Lesson conflict update when hiding individual courses - When a hide individual class box is ticked, conflicts should be updated: run the conflicts algorithm once again, not considering the hidden classes. (refreshHiddenLessons @ LessonBox.tsx) Suggestion: change CheckedCourse type to this ```ts export type CheckedCourse = { checked: boolean hidden: boolean // new info: Course } ```
priority
lesson conflict update when hiding individual courses when a hide individual class box is ticked conflicts should be updated run the conflicts algorithm once again not considering the hidden classes refreshhiddenlessons lessonbox tsx suggestion change checkedcourse type to this ts export type checkedcourse checked boolean hidden boolean new info course
1
167,593
6,342,072,105
IssuesEvent
2017-07-27 14:48:28
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Trigger notification if a map is an old one asking for updating it
enhancement pending review Priority: High
**requirements:** - The user must have write access for the resource map. - Must be a mapstore map (old mapstore) - remove the refresh tools from the TOC **behaviour:** if the requirements are met then show a notification in top center position with warning style (yellow) containing: - an update button that will update the map in background (with all the options selected). ![image](https://user-images.githubusercontent.com/11991428/28072857-778c3082-6654-11e7-8d1a-8d0080a62096.png) - a body text "This is an old map, thus not all the functionalities can be enabled. Click on update button in order to update the map or simply dismiss this notification if you don't want to update it."
1.0
Trigger notification if a map is an old one asking for updating it - **requirements:** - The user must have write access for the resource map. - Must be a mapstore map (old mapstore) - remove the refresh tools from the TOC **behaviour:** if the requirements are met then show a notification in top center position with warning style (yellow) containing: - an update button that will update the map in background (with all the options selected). ![image](https://user-images.githubusercontent.com/11991428/28072857-778c3082-6654-11e7-8d1a-8d0080a62096.png) - a body text "This is an old map, thus not all the functionalities can be enabled. Click on update button in order to update the map or simply dismiss this notification if you don't want to update it."
priority
trigger notification if a map is an old one asking for updating it requirements the user must have write access for the resource map must be a mapstore map old mapstore remove the refresh tools from the toc behaviour if the requirements are met then show a notification in top center position with warning style yellow containing an update button that will update the map in background with all the options selected a body text this is an old map thus not all the functionalities can be enabled click on update button in order to update the map or simply dismiss this notification if you don t want to update it
1
411,933
12,033,704,693
IssuesEvent
2020-04-13 14:44:42
ooni/backend
https://api.github.com/repos/ooni/backend
closed
Fastpath: keep rows in fastpath table
effort/L enhancement epic ooni/pipeline priority/high
Delete json files of disk but keep the row in the db table Check how the API is going to handle this. Implement integ tests in the API
1.0
Fastpath: keep rows in fastpath table - Delete json files of disk but keep the row in the db table Check how the API is going to handle this. Implement integ tests in the API
priority
fastpath keep rows in fastpath table delete json files of disk but keep the row in the db table check how the api is going to handle this implement integ tests in the api
1
168,137
6,362,591,444
IssuesEvent
2017-07-31 15:16:14
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Layer opacity setting in Leaflet
bug pending review Priority: High
Using Leaflet each layer opacity changes from layer settings produce a new WMS request with the OPACITY parameter. This must be avoided. Steps to reproduce: - open a map using Leaflet - add a WMS layer or use an existing one in TOC - open the layer properties tool and change the opacity, then check the performed requests using firebug or the chrome dev tool
1.0
Layer opacity setting in Leaflet - Using Leaflet each layer opacity changes from layer settings produce a new WMS request with the OPACITY parameter. This must be avoided. Steps to reproduce: - open a map using Leaflet - add a WMS layer or use an existing one in TOC - open the layer properties tool and change the opacity, then check the performed requests using firebug or the chrome dev tool
priority
layer opacity setting in leaflet using leaflet each layer opacity changes from layer settings produce a new wms request with the opacity parameter this must be avoided steps to reproduce open a map using leaflet add a wms layer or use an existing one in toc open the layer properties tool and change the opacity then check the performed requests using firebug or the chrome dev tool
1
743,456
25,899,720,590
IssuesEvent
2022-12-15 03:48:45
julianmichael/debate
https://api.github.com/repos/julianmichael/debate
closed
Fix intermittent websocket disconnects
bug high priority
* Happened to one person using Firefox and Chrome but not others. Not sure if we can reliably reproduce but putting the issue here for further reference in case we reproduce it. * It's possible we need to add heartbeats. I don't think it's necessary though? * The likely fix will be getting the client to seamlessly auto-reconnect.
1.0
Fix intermittent websocket disconnects - * Happened to one person using Firefox and Chrome but not others. Not sure if we can reliably reproduce but putting the issue here for further reference in case we reproduce it. * It's possible we need to add heartbeats. I don't think it's necessary though? * The likely fix will be getting the client to seamlessly auto-reconnect.
priority
fix intermittent websocket disconnects happened to one person using firefox and chrome but not others not sure if we can reliably reproduce but putting the issue here for further reference in case we reproduce it it s possible we need to add heartbeats i don t think it s necessary though the likely fix will be getting the client to seamlessly auto reconnect
1
366,279
10,819,229,472
IssuesEvent
2019-11-08 13:57:58
AY1920S1-CS2103T-W13-3/main
https://api.github.com/repos/AY1920S1-CS2103T-W13-3/main
closed
Implement test cases for NoteFragment methods
priority.High type.Task
Includes: - ParserUtil NoteFragment methods - NoteFragment class - NoteFragment storage (if any) - NoteFragment integration with CheatSheet (if necessary)
1.0
Implement test cases for NoteFragment methods - Includes: - ParserUtil NoteFragment methods - NoteFragment class - NoteFragment storage (if any) - NoteFragment integration with CheatSheet (if necessary)
priority
implement test cases for notefragment methods includes parserutil notefragment methods notefragment class notefragment storage if any notefragment integration with cheatsheet if necessary
1
247,254
7,915,789,238
IssuesEvent
2018-07-04 01:43:34
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: Treasury
High Priority
**Version:** 0.7.5.1 beta **Steps to Reproduce:** set tax, gain money in treasury, click allocate **Expected behavior:** it should have given me a pop up to allocate money **Actual behavior:** The following error message popped up: `Server encountered an exceptio: Message:Original type is Object. Cannot convert from Object to Guid Source:Eco.Shared` Server encountered an exception: <size=60.00%>Exception: Exception Message:Original type is Object. Cannot convert from Object to Guid Source:Eco.Shared System.Exception: Original type is Object. Cannot convert from Object to Guid at Eco.Shared.Serialization.BSONValue.get_GuidValue() at Eco.Gameplay.Economy.Currency.<>c.<.cctor>b__45_0(BSONValue v) at Eco.Shared.Serialization.BsonManipulator.FromBson(BSONValue value, Type type) at Eco.Shared.Networking.RPCManager.UnpackArgs(BSONObject bson, RPCMethod[] methods, Object extensionMethodObject, RPCMethod& targetMethod)</size> ```
1.0
USER ISSUE: Treasury - **Version:** 0.7.5.1 beta **Steps to Reproduce:** set tax, gain money in treasury, click allocate **Expected behavior:** it should have given me a pop up to allocate money **Actual behavior:** The following error message popped up: `Server encountered an exceptio: Message:Original type is Object. Cannot convert from Object to Guid Source:Eco.Shared` Server encountered an exception: <size=60.00%>Exception: Exception Message:Original type is Object. Cannot convert from Object to Guid Source:Eco.Shared System.Exception: Original type is Object. Cannot convert from Object to Guid at Eco.Shared.Serialization.BSONValue.get_GuidValue() at Eco.Gameplay.Economy.Currency.<>c.<.cctor>b__45_0(BSONValue v) at Eco.Shared.Serialization.BsonManipulator.FromBson(BSONValue value, Type type) at Eco.Shared.Networking.RPCManager.UnpackArgs(BSONObject bson, RPCMethod[] methods, Object extensionMethodObject, RPCMethod& targetMethod)</size> ```
priority
user issue treasury version beta steps to reproduce set tax gain money in treasury click allocate expected behavior it should have given me a pop up to allocate money actual behavior the following error message popped up server encountered an exceptio message original type is object cannot convert from object to guid source eco shared server encountered an exception exception exception message original type is object cannot convert from object to guid source eco shared system exception original type is object cannot convert from object to guid at eco shared serialization bsonvalue get guidvalue at eco gameplay economy currency c b bsonvalue v at eco shared serialization bsonmanipulator frombson bsonvalue value type type at eco shared networking rpcmanager unpackargs bsonobject bson rpcmethod methods object extensionmethodobject rpcmethod targetmethod
1
111,020
4,448,644,102
IssuesEvent
2016-08-22 01:14:25
AtlasOfLivingAustralia/spatial-portal
https://api.github.com/repos/AtlasOfLivingAustralia/spatial-portal
closed
Area report - endemics not 100% correct
bug priority-high
Area: POLYGON((115.21 -20.927,115.556 -20.927,115.556 -20.662,115.21 -20.662,115.21 -20.927)). Genus Austroteneriffia not endemic to area. Other 14 reported taxa are endemic but two taxa Barrowammo waldockae reports 3 occurrences but 6 appear to be there. Ditto, genus Dictus reports 1 occurrence and 6 appear to be there. [Endemics - Barrow Island area.xlsx](https://github.com/AtlasOfLivingAustralia/spatial-portal/files/368230/Endemics.-.Barrow.Island.area.xlsx)
1.0
Area report - endemics not 100% correct - Area: POLYGON((115.21 -20.927,115.556 -20.927,115.556 -20.662,115.21 -20.662,115.21 -20.927)). Genus Austroteneriffia not endemic to area. Other 14 reported taxa are endemic but two taxa Barrowammo waldockae reports 3 occurrences but 6 appear to be there. Ditto, genus Dictus reports 1 occurrence and 6 appear to be there. [Endemics - Barrow Island area.xlsx](https://github.com/AtlasOfLivingAustralia/spatial-portal/files/368230/Endemics.-.Barrow.Island.area.xlsx)
priority
area report endemics not correct area polygon genus austroteneriffia not endemic to area other reported taxa are endemic but two taxa barrowammo waldockae reports occurrences but appear to be there ditto genus dictus reports occurrence and appear to be there
1
371,881
10,986,923,137
IssuesEvent
2019-12-02 08:01:38
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
addon-knobs not working with vue from v4.0.8
addon: knobs app: vue bug high priority
Upgraded from 4.0.4 to 4.0.12 and knobs were no longer visible or working. Tested each version and 4.0.7 works but 4.0.8 does not. I assume it's something to do with #4773. Typical use of knobs looks something like this (if that helps): ``` template: '<simple-knob-example :age="age" />', data: () => ({ age: 40, }), ``` I don't see any error messages or anything, but if there is anything else I can do to help please let us know.
1.0
addon-knobs not working with vue from v4.0.8 - Upgraded from 4.0.4 to 4.0.12 and knobs were no longer visible or working. Tested each version and 4.0.7 works but 4.0.8 does not. I assume it's something to do with #4773. Typical use of knobs looks something like this (if that helps): ``` template: '<simple-knob-example :age="age" />', data: () => ({ age: 40, }), ``` I don't see any error messages or anything, but if there is anything else I can do to help please let us know.
priority
addon knobs not working with vue from upgraded from to and knobs were no longer visible or working tested each version and works but does not i assume it s something to do with typical use of knobs looks something like this if that helps template data age i don t see any error messages or anything but if there is anything else i can do to help please let us know
1
322,432
9,817,641,662
IssuesEvent
2019-06-13 17:13:55
chrisly-bear/PEBRApp
https://api.github.com/repos/chrisly-bear/PEBRApp
closed
Display Title above Calender after ART Refill Done
high priority
When the user clicks "Refill Done" on the ART Refill Screen, the calendar pops up. It might not be clear to the user that they have to select the date of the _next_ ART Refill (instead they might select the date when the refill was done). Display a title above the calendar or inform the user otherwise to select the next ART Refill date.
1.0
Display Title above Calender after ART Refill Done - When the user clicks "Refill Done" on the ART Refill Screen, the calendar pops up. It might not be clear to the user that they have to select the date of the _next_ ART Refill (instead they might select the date when the refill was done). Display a title above the calendar or inform the user otherwise to select the next ART Refill date.
priority
display title above calender after art refill done when the user clicks refill done on the art refill screen the calendar pops up it might not be clear to the user that they have to select the date of the next art refill instead they might select the date when the refill was done display a title above the calendar or inform the user otherwise to select the next art refill date
1
141,032
5,428,305,335
IssuesEvent
2017-03-03 15:36:25
GoogleCloudPlatform/google-cloud-eclipse
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse
closed
Publish plugin to Eclipse marketplace
GA high priority task
so users don't have to use our download site: http://marketplace.eclipse.org/quickstart (via JavaOne)
1.0
Publish plugin to Eclipse marketplace - so users don't have to use our download site: http://marketplace.eclipse.org/quickstart (via JavaOne)
priority
publish plugin to eclipse marketplace so users don t have to use our download site via javaone
1
362,714
10,731,051,643
IssuesEvent
2019-10-28 18:41:17
ComPWA/ComPWA
https://api.github.com/repos/ComPWA/ComPWA
closed
Fix conflicts with tbb when building as static library
Priority: High Type: Bug
Building with static libraries sometimes results in weird runtime behaviour (the executable keeps running) is you use more than one ComPWA module that implement some parallelisation from `tbb`. See [this travis job result](https://travis-ci.com/ComPWA/ComPWA/jobs/249466271#L1514) of [pull request 237](https://github.com/ComPWA/ComPWA/pull/237) in particular. Actually, `tbb` strongly advises not to build static libraries, but we need to build ComPWA as static for pycompwa. @spflueger knows more about the details.
1.0
Fix conflicts with tbb when building as static library - Building with static libraries sometimes results in weird runtime behaviour (the executable keeps running) is you use more than one ComPWA module that implement some parallelisation from `tbb`. See [this travis job result](https://travis-ci.com/ComPWA/ComPWA/jobs/249466271#L1514) of [pull request 237](https://github.com/ComPWA/ComPWA/pull/237) in particular. Actually, `tbb` strongly advises not to build static libraries, but we need to build ComPWA as static for pycompwa. @spflueger knows more about the details.
priority
fix conflicts with tbb when building as static library building with static libraries sometimes results in weird runtime behaviour the executable keeps running is you use more than one compwa module that implement some parallelisation from tbb see of in particular actually tbb strongly advises not to build static libraries but we need to build compwa as static for pycompwa spflueger knows more about the details
1
432,850
12,499,189,571
IssuesEvent
2020-06-01 19:41:35
Code-and-Response/ClusterDuck-Protocol
https://api.github.com/repos/Code-and-Response/ClusterDuck-Protocol
opened
Sensor Integrations
feature high-priority
The ClusterDuck Protocol provides a foundation that can enable a variety of sensor applications for different purposes. An example of this are the SolarDucks in Puerto Rico - many of these have BMP180 (temperature, pressure, altimeter) sensors or an Air Quality PM2.5 sensor. Most sensors that can be integrated into Arduino can be integrated into a DuckLink. ![solarduck_github](https://user-images.githubusercontent.com/1772065/83447331-46adf080-a41e-11ea-9a39-3f9ab4e496b1.jpg) **This issue focuses on two main goals** 1. Constructing the process for future integrations of sensors 2. Building the integration of numerous common sensors already used. There is already an example for the BMP180 sensor. Others we are looking to build out - MQXX Gas sensors (MQ-1, MQ-2, MQ-3, etc). - Accelerometers - Air Quality PM2.5 sensors - Infrared heath thermometers - Motion/visual sensors
1.0
Sensor Integrations - The ClusterDuck Protocol provides a foundation that can enable a variety of sensor applications for different purposes. An example of this are the SolarDucks in Puerto Rico - many of these have BMP180 (temperature, pressure, altimeter) sensors or an Air Quality PM2.5 sensor. Most sensors that can be integrated into Arduino can be integrated into a DuckLink. ![solarduck_github](https://user-images.githubusercontent.com/1772065/83447331-46adf080-a41e-11ea-9a39-3f9ab4e496b1.jpg) **This issue focuses on two main goals** 1. Constructing the process for future integrations of sensors 2. Building the integration of numerous common sensors already used. There is already an example for the BMP180 sensor. Others we are looking to build out - MQXX Gas sensors (MQ-1, MQ-2, MQ-3, etc). - Accelerometers - Air Quality PM2.5 sensors - Infrared heath thermometers - Motion/visual sensors
priority
sensor integrations the clusterduck protocol provides a foundation that can enable a variety of sensor applications for different purposes an example of this are the solarducks in puerto rico many of these have temperature pressure altimeter sensors or an air quality sensor most sensors that can be integrated into arduino can be integrated into a ducklink this issue focuses on two main goals constructing the process for future integrations of sensors building the integration of numerous common sensors already used there is already an example for the sensor others we are looking to build out mqxx gas sensors mq mq mq etc accelerometers air quality sensors infrared heath thermometers motion visual sensors
1
291,797
8,949,703,688
IssuesEvent
2019-01-25 08:37:00
cekit/cekit
https://api.github.com/repos/cekit/cekit
opened
Investigate changes in image size after switching to applying modules individually
complexity/high priority/high
This is related to #329. We need to ensure that the resulting image is not bigger than the original. If it is, we need to investigate what is the root cause and what can we do about it.
1.0
Investigate changes in image size after switching to applying modules individually - This is related to #329. We need to ensure that the resulting image is not bigger than the original. If it is, we need to investigate what is the root cause and what can we do about it.
priority
investigate changes in image size after switching to applying modules individually this is related to we need to ensure that the resulting image is not bigger than the original if it is we need to investigate what is the root cause and what can we do about it
1
694,318
23,809,630,185
IssuesEvent
2022-09-04 15:09:09
connect-ring/js-swa-sample
https://api.github.com/repos/connect-ring/js-swa-sample
opened
【機能追加】Functions API の返り値を変更
Type: enhancement Priority: high
## やること - Functions API の返り値を変更する ## クローズ条件 - 上記やることが完了し、PR 承認完了したらクローズとする ## その他 - プルリクエストのURL:https://xxx
1.0
【機能追加】Functions API の返り値を変更 - ## やること - Functions API の返り値を変更する ## クローズ条件 - 上記やることが完了し、PR 承認完了したらクローズとする ## その他 - プルリクエストのURL:https://xxx
priority
【機能追加】functions api の返り値を変更 やること functions api の返り値を変更する クローズ条件 上記やることが完了し、pr 承認完了したらクローズとする その他 プルリクエストのurl:
1
25,358
2,679,607,585
IssuesEvent
2015-03-26 17:32:21
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
sct_apply_transfo uses absolute path to warping fields
enhancement priority: high sct_apply_transfo
lines 106-109 ```python for i in range(len(fname_warp_list_invert)): sct.check_file_exist(fname_warp_list_invert[i]) # get absolute path fname_warp_list_invert[i] = os.path.abspath(fname_warp_list_invert[i]) ```
1.0
sct_apply_transfo uses absolute path to warping fields - lines 106-109 ```python for i in range(len(fname_warp_list_invert)): sct.check_file_exist(fname_warp_list_invert[i]) # get absolute path fname_warp_list_invert[i] = os.path.abspath(fname_warp_list_invert[i]) ```
priority
sct apply transfo uses absolute path to warping fields lines python for i in range len fname warp list invert sct check file exist fname warp list invert get absolute path fname warp list invert os path abspath fname warp list invert
1
316,000
9,634,719,921
IssuesEvent
2019-05-15 22:04:52
IQSS/DataTaggingServer
https://api.github.com/repos/IQSS/DataTaggingServer
opened
Report and Summary tag value views: Maintain policy space ordering
Priority: High
Currently the order is not as defined in the policy model space definition.
1.0
Report and Summary tag value views: Maintain policy space ordering - Currently the order is not as defined in the policy model space definition.
priority
report and summary tag value views maintain policy space ordering currently the order is not as defined in the policy model space definition
1
69,761
3,314,460,622
IssuesEvent
2015-11-06 05:25:19
lonh/lon-workspace
https://api.github.com/repos/lonh/lon-workspace
closed
Google image api
auto-migrated Priority-High Type-Task
``` Utilize google image service ``` Original issue reported on code.google.com by `lon...@gmail.com` on 2 Feb 2012 at 2:38
1.0
Google image api - ``` Utilize google image service ``` Original issue reported on code.google.com by `lon...@gmail.com` on 2 Feb 2012 at 2:38
priority
google image api utilize google image service original issue reported on code google com by lon gmail com on feb at
1
343,884
10,337,946,554
IssuesEvent
2019-09-03 15:49:47
status-im/status-react
https://api.github.com/repos/status-im/status-react
opened
"Null is not an object" crash when tapping on "Learn more" screen during onboarding
bug high-priority onboarding v1 release
### Description *Type*: Bug *Summary*: tapping on "Learn more" link during onboarding leads to crash (on "Choose key and name", "Select key storage") For e2e it is required to add accessibility-id to this link to prevent this to happen in the future. #### Expected behavior no crash #### Actual behavior <img width="467" alt="cvxvc" src="https://user-images.githubusercontent.com/4557972/64188262-5b9b2a80-ce72-11e9-9c6f-636682a03a4f.png"> ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) - Open Status - Tap on "Generate new key" - Tap on "learn more" ### Additional Information [comment]: # (Please do your best to fill this out.) - Status version: nightly 03/09/2019 - Operating System: Android, iOS
1.0
"Null is not an object" crash when tapping on "Learn more" screen during onboarding - ### Description *Type*: Bug *Summary*: tapping on "Learn more" link during onboarding leads to crash (on "Choose key and name", "Select key storage") For e2e it is required to add accessibility-id to this link to prevent this to happen in the future. #### Expected behavior no crash #### Actual behavior <img width="467" alt="cvxvc" src="https://user-images.githubusercontent.com/4557972/64188262-5b9b2a80-ce72-11e9-9c6f-636682a03a4f.png"> ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) - Open Status - Tap on "Generate new key" - Tap on "learn more" ### Additional Information [comment]: # (Please do your best to fill this out.) - Status version: nightly 03/09/2019 - Operating System: Android, iOS
priority
null is not an object crash when tapping on learn more screen during onboarding description type bug summary tapping on learn more link during onboarding leads to crash on choose key and name select key storage for it is required to add accessibility id to this link to prevent this to happen in the future expected behavior no crash actual behavior img width alt cvxvc src reproduction describe how we can replicate the bug step by step open status tap on generate new key tap on learn more additional information please do your best to fill this out status version nightly operating system android ios
1
782,798
27,507,252,925
IssuesEvent
2023-03-06 05:16:16
AY2223S2-CS2103T-T12-2/tp
https://api.github.com/repos/AY2223S2-CS2103T-T12-2/tp
opened
Remove the discharge date of a patient
type.Story priority.High
As a staff member, I want to remove the discharge date of a patient so that I can put the patient's discharge date on hold while awaiting further assessment.
1.0
Remove the discharge date of a patient - As a staff member, I want to remove the discharge date of a patient so that I can put the patient's discharge date on hold while awaiting further assessment.
priority
remove the discharge date of a patient as a staff member i want to remove the discharge date of a patient so that i can put the patient s discharge date on hold while awaiting further assessment
1
638,021
20,693,624,693
IssuesEvent
2022-03-11 04:54:04
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[DocDB] Potential deadlock in rocksdb when applying intents
area/docdb priority/high
### Description ``` @ 0x7faf175a9b39 __lll_lock_wait @ 0x7faf175a46e2 __GI___pthread_mutex_lock @ 0x33b8370 rocksdb::InstrumentedMutex::Lock() @ 0x32aae5f rocksdb::ColumnFamilyData::GetThreadLocalSuperVersion() @ 0x32d3de7 rocksdb::DBImpl::NewIterator() @ 0x30c68ca yb::docdb::BoundedRocksDbIterator::BoundedRocksDbIterator() @ 0x3120018 yb::docdb::CreateRocksDBIterator() @ 0x315ff68 yb::docdb::IntentsWriter::Apply() @ 0x3362adc rocksdb::WriteBatch::Iterate() @ 0x32e3e1f rocksdb::DBImpl::WriteImpl() @ 0x32d2700 rocksdb::DBImpl::Write() @ 0x34c0624 yb::tablet::Tablet::WriteToRocksDB() @ 0x34bd9fe yb::tablet::Tablet::ApplyIntents() @ 0x34be1b1 yb::tablet::Tablet::ApplyIntents() @ 0x354d83d yb::tablet::TransactionParticipant::Impl::ProcessReplicated() @ 0x34a7e37 yb::tablet::UpdateTxnOperation::DoReplicated() Total number of threads: 1 ``` Slack link for more context + dump of all threads: https://yugabyte.slack.com/archives/C01CB38CZHU/p1646669491969579?thread_ts=1646378629.443139&cid=C01CB38CZHU > version 2.13.0.0 build 21 revision 3156825304c9ad4b7ea20464bab99a272dba758d build_type RELEASE built at 01 Mar 2022 02:16:00 UTC
1.0
[DocDB] Potential deadlock in rocksdb when applying intents - ### Description ``` @ 0x7faf175a9b39 __lll_lock_wait @ 0x7faf175a46e2 __GI___pthread_mutex_lock @ 0x33b8370 rocksdb::InstrumentedMutex::Lock() @ 0x32aae5f rocksdb::ColumnFamilyData::GetThreadLocalSuperVersion() @ 0x32d3de7 rocksdb::DBImpl::NewIterator() @ 0x30c68ca yb::docdb::BoundedRocksDbIterator::BoundedRocksDbIterator() @ 0x3120018 yb::docdb::CreateRocksDBIterator() @ 0x315ff68 yb::docdb::IntentsWriter::Apply() @ 0x3362adc rocksdb::WriteBatch::Iterate() @ 0x32e3e1f rocksdb::DBImpl::WriteImpl() @ 0x32d2700 rocksdb::DBImpl::Write() @ 0x34c0624 yb::tablet::Tablet::WriteToRocksDB() @ 0x34bd9fe yb::tablet::Tablet::ApplyIntents() @ 0x34be1b1 yb::tablet::Tablet::ApplyIntents() @ 0x354d83d yb::tablet::TransactionParticipant::Impl::ProcessReplicated() @ 0x34a7e37 yb::tablet::UpdateTxnOperation::DoReplicated() Total number of threads: 1 ``` Slack link for more context + dump of all threads: https://yugabyte.slack.com/archives/C01CB38CZHU/p1646669491969579?thread_ts=1646378629.443139&cid=C01CB38CZHU > version 2.13.0.0 build 21 revision 3156825304c9ad4b7ea20464bab99a272dba758d build_type RELEASE built at 01 Mar 2022 02:16:00 UTC
priority
potential deadlock in rocksdb when applying intents description lll lock wait gi pthread mutex lock rocksdb instrumentedmutex lock rocksdb columnfamilydata getthreadlocalsuperversion rocksdb dbimpl newiterator yb docdb boundedrocksdbiterator boundedrocksdbiterator yb docdb createrocksdbiterator yb docdb intentswriter apply rocksdb writebatch iterate rocksdb dbimpl writeimpl rocksdb dbimpl write yb tablet tablet writetorocksdb yb tablet tablet applyintents yb tablet tablet applyintents yb tablet transactionparticipant impl processreplicated yb tablet updatetxnoperation doreplicated total number of threads slack link for more context dump of all threads version build revision build type release built at mar utc
1
636,637
20,604,645,553
IssuesEvent
2022-03-06 19:54:26
BlueCodeSystems/opensrp-client-ecap-chw
https://api.github.com/repos/BlueCodeSystems/opensrp-client-ecap-chw
closed
The questions on TB when the option "Yes" is selected on "Does the child/adolescent have their TB screening results on file?" The questions that follows are duplicated.
bug High Priority
![Screenshot_20220121-100612_Ecap II](https://user-images.githubusercontent.com/86519642/150490297-93b0f23f-d9cb-447e-974a-0865d5a7b488.jpg)
1.0
The questions on TB when the option "Yes" is selected on "Does the child/adolescent have their TB screening results on file?" The questions that follows are duplicated. - ![Screenshot_20220121-100612_Ecap II](https://user-images.githubusercontent.com/86519642/150490297-93b0f23f-d9cb-447e-974a-0865d5a7b488.jpg)
priority
the questions on tb when the option yes is selected on does the child adolescent have their tb screening results on file the questions that follows are duplicated
1
196,483
6,928,407,657
IssuesEvent
2017-12-01 04:40:09
DMS-Aus/Roam
https://api.github.com/repos/DMS-Aus/Roam
opened
Reinstate mulit image widget.
enhancement :) priority/high
This widget was disabled for 2.6 pending more testing but will need to be came back into the next release as some people are using it.
1.0
Reinstate mulit image widget. - This widget was disabled for 2.6 pending more testing but will need to be came back into the next release as some people are using it.
priority
reinstate mulit image widget this widget was disabled for pending more testing but will need to be came back into the next release as some people are using it
1
518,004
15,022,344,147
IssuesEvent
2021-02-01 16:50:40
owncloud/web
https://api.github.com/repos/owncloud/web
closed
web 1.0.2
Priority:p2-high
### Reason - Release needed for the correct version bump. ### Product approval - [x] Verify transifex commits are in master: If a folder https://github.com/owncloud/<APPNAME>/tree/master/l10n exists, then https://drone.owncloud.com/owncloud/<APPNAME> should receive nightly '[tx] updated from transifex` ... - [x] Ping pmaier to check `info.xml` description and screenshot - [x] Get approval on translations from pmaier - send the date of the last tx commit from drone to Patrick ### QA - [ ] Generate Changelog - [x] Create **QA** Milestone and set the description to $version - [ ] Move items from `development` to `qa` - [ ] Create release branch with the format "release-$version" - [x] Create "$version Testplan" #4638 ### Documentation - [ ] Changes in the docs needed? If so create issue in owncloud/docs. ### Marketing - [x] Ping @owncloud/marketing ### Build All actions to be done on the release branch from here: - [x] Bump version in `info.xml` (no version suffix) - [x] Check php version dependency in `info.xml` - [x] Check minimum ownCloud version dependency in `info.xml` - [ ] Change CHANGELOG.md 'Unreleased' to '$version' (today's date) #### Beta/RC - [ ] Create github tag for `v${version}RC1` on release branch - [ ] Build and sign RC tarball from github tag `v${version}RC1` (see handbook for how to build) - [ ] Upload as Github release using changelog entries - [ ] Test and document in testplan => Final "Go" from QA (set 'QA-ready' label) => else do another RC after fixing issues in the "release-$version" branch #### Final - [x] Check that no documentation issues are pending - [x] Rename last RC tarball to final - [x] Create github tag for final `v${version}` on release branch - [x] Upload as Github release using changelog entries ### Publishing - [x] Upload tarball to marketplace (see handbook in confluence for details) - [ ] Publish tarball for customers (see handbook in confluence for details) - https://customer.owncloud.com and https://portal.owncloud.com - [x] Send mail to release-coordination with link to this ticket and to the changelog - [x] Copy email to rocket chat `#updates` channel ### Post-release - [ ] Rename QA milestone to `$version` - [ ] Add a new 'Unreleased' section in CHANGELOG.md (and also update links at the bottom!) - [ ] Create PR to merge changes back to the `master` branch with description "Closes #XYZ" with the release ticket number - [ ] Close the milestone - [ ] Ensure [release template](https://github.com/owncloud/handbook/edit/master/modules/ROOT/examples/releases/server_app_release_template.md) is up to date
1.0
web 1.0.2 - ### Reason - Release needed for the correct version bump. ### Product approval - [x] Verify transifex commits are in master: If a folder https://github.com/owncloud/<APPNAME>/tree/master/l10n exists, then https://drone.owncloud.com/owncloud/<APPNAME> should receive nightly '[tx] updated from transifex` ... - [x] Ping pmaier to check `info.xml` description and screenshot - [x] Get approval on translations from pmaier - send the date of the last tx commit from drone to Patrick ### QA - [ ] Generate Changelog - [x] Create **QA** Milestone and set the description to $version - [ ] Move items from `development` to `qa` - [ ] Create release branch with the format "release-$version" - [x] Create "$version Testplan" #4638 ### Documentation - [ ] Changes in the docs needed? If so create issue in owncloud/docs. ### Marketing - [x] Ping @owncloud/marketing ### Build All actions to be done on the release branch from here: - [x] Bump version in `info.xml` (no version suffix) - [x] Check php version dependency in `info.xml` - [x] Check minimum ownCloud version dependency in `info.xml` - [ ] Change CHANGELOG.md 'Unreleased' to '$version' (today's date) #### Beta/RC - [ ] Create github tag for `v${version}RC1` on release branch - [ ] Build and sign RC tarball from github tag `v${version}RC1` (see handbook for how to build) - [ ] Upload as Github release using changelog entries - [ ] Test and document in testplan => Final "Go" from QA (set 'QA-ready' label) => else do another RC after fixing issues in the "release-$version" branch #### Final - [x] Check that no documentation issues are pending - [x] Rename last RC tarball to final - [x] Create github tag for final `v${version}` on release branch - [x] Upload as Github release using changelog entries ### Publishing - [x] Upload tarball to marketplace (see handbook in confluence for details) - [ ] Publish tarball for customers (see handbook in confluence for details) - https://customer.owncloud.com and https://portal.owncloud.com - [x] Send mail to release-coordination with link to this ticket and to the changelog - [x] Copy email to rocket chat `#updates` channel ### Post-release - [ ] Rename QA milestone to `$version` - [ ] Add a new 'Unreleased' section in CHANGELOG.md (and also update links at the bottom!) - [ ] Create PR to merge changes back to the `master` branch with description "Closes #XYZ" with the release ticket number - [ ] Close the milestone - [ ] Ensure [release template](https://github.com/owncloud/handbook/edit/master/modules/ROOT/examples/releases/server_app_release_template.md) is up to date
priority
web reason release needed for the correct version bump product approval verify transifex commits are in master if a folder exists then should receive nightly updated from transifex ping pmaier to check info xml description and screenshot get approval on translations from pmaier send the date of the last tx commit from drone to patrick qa generate changelog create qa milestone and set the description to version move items from development to qa create release branch with the format release version create version testplan documentation changes in the docs needed if so create issue in owncloud docs marketing ping owncloud marketing build all actions to be done on the release branch from here bump version in info xml no version suffix check php version dependency in info xml check minimum owncloud version dependency in info xml change changelog md unreleased to version today s date beta rc create github tag for v version on release branch build and sign rc tarball from github tag v version see handbook for how to build upload as github release using changelog entries test and document in testplan final go from qa set qa ready label else do another rc after fixing issues in the release version branch final check that no documentation issues are pending rename last rc tarball to final create github tag for final v version on release branch upload as github release using changelog entries publishing upload tarball to marketplace see handbook in confluence for details publish tarball for customers see handbook in confluence for details and send mail to release coordination with link to this ticket and to the changelog copy email to rocket chat updates channel post release rename qa milestone to version add a new unreleased section in changelog md and also update links at the bottom create pr to merge changes back to the master branch with description closes xyz with the release ticket number close the milestone ensure is up to date
1
728,786
25,092,899,378
IssuesEvent
2022-11-08 08:03:25
quickwit-oss/quickwit
https://api.github.com/repos/quickwit-oss/quickwit
closed
Shut down delete pipeline supervisors when shutting down the delete pipeline
bug high-priority
When we detect that an index has been deleted, we shut down the corresponding pipeline. But supervisors are not shut down. We need to shut down each supervisor on the delete pipeline finalization.
1.0
Shut down delete pipeline supervisors when shutting down the delete pipeline - When we detect that an index has been deleted, we shut down the corresponding pipeline. But supervisors are not shut down. We need to shut down each supervisor on the delete pipeline finalization.
priority
shut down delete pipeline supervisors when shutting down the delete pipeline when we detect that an index has been deleted we shut down the corresponding pipeline but supervisors are not shut down we need to shut down each supervisor on the delete pipeline finalization
1
397,878
11,734,526,130
IssuesEvent
2020-03-11 09:30:36
WoWManiaUK/Blackwing-Lair
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
closed
[BOSS/Spell] Lightning blast - The Bastian of Twilight
Class Dungeon/Raid Fixed Confirmed Fixed in Dev Priority-High
**Links:** https://wow.gamepedia.com/Ascendant_Council ![lightning blast](https://user-images.githubusercontent.com/39439201/74115497-c5e99100-4baf-11ea-9aaf-c459947b5f3a.jpg) **What is happening:** - Lightning blast is hitting the whole raid **What should happen:** while it should only target the one who's tanking Arion
1.0
[BOSS/Spell] Lightning blast - The Bastian of Twilight - **Links:** https://wow.gamepedia.com/Ascendant_Council ![lightning blast](https://user-images.githubusercontent.com/39439201/74115497-c5e99100-4baf-11ea-9aaf-c459947b5f3a.jpg) **What is happening:** - Lightning blast is hitting the whole raid **What should happen:** while it should only target the one who's tanking Arion
priority
lightning blast the bastian of twilight links what is happening lightning blast is hitting the whole raid what should happen while it should only target the one who s tanking arion
1
793,282
27,988,949,234
IssuesEvent
2023-03-27 00:40:52
AY2223S2-CS2103T-T14-4/tp
https://api.github.com/repos/AY2223S2-CS2103T-T14-4/tp
closed
As a new TA, I need help tracking time
priority.High
I need something that reminds me how much time I have left for the class.
1.0
As a new TA, I need help tracking time - I need something that reminds me how much time I have left for the class.
priority
as a new ta i need help tracking time i need something that reminds me how much time i have left for the class
1
288,368
8,838,861,042
IssuesEvent
2019-01-05 22:04:05
rcmaples/sputnikapp
https://api.github.com/repos/rcmaples/sputnikapp
closed
Registration form error messages aren't showing.
Priority: High Status: Pending Type: Bug
When visiting https://sputnikapp.com/signup error messages should display if registration fails. Currently the icons change color as expected, but the messages are not displayed.
1.0
Registration form error messages aren't showing. - When visiting https://sputnikapp.com/signup error messages should display if registration fails. Currently the icons change color as expected, but the messages are not displayed.
priority
registration form error messages aren t showing when visiting error messages should display if registration fails currently the icons change color as expected but the messages are not displayed
1
234,324
7,719,764,048
IssuesEvent
2018-05-23 20:31:12
WordImpress/wp-business-reviews
https://api.github.com/repos/WordImpress/wp-business-reviews
opened
fix(admin-review): ensure single reviews can be saved without error
high-priority
## User Story As a WPBR user, I want to create and edit single reviews without errors appearing so my changes are saved. ## Acceptance Criteria - [ ] I can add new reviews without error. - [ ] I can edit existing reviews without error. - [ ] There are no console errors on page load of the review edit screen. ## Current Behavior I currently see a PHP notice is displayed on save: > Notice: Trying to get property 'platform' of non-object in /Users/kevin/Sites/wpbrclient5/wp-content/plugins/wp-business-reviews/includes/admin/class-admin-reviews-single.php on line 318 ## Expected Behavior I expect to save a review and have my changes preserved. ## Visuals ### Video https://drive.google.com/file/d/1EXUBRWQ09R_mJmksq9Q53pTNCtemUly6/view?usp=sharing ### Screenshots #### PHP error on save ![image](https://user-images.githubusercontent.com/1938671/40449291-148759e4-5ea6-11e8-931a-0af2e71b2fed.png) #### Console error on page load ![image](https://user-images.githubusercontent.com/1938671/40449348-3ba15aac-5ea6-11e8-9d3e-40d1b2ea2e6d.png) ## Steps to Recreate for New Review 1. Go to Reviews > Add Review 2. Add a title and set a rating. 3. Click `Publish` and note the error on screen. ## Steps to Recreate for Existing Review 1. Go to Reviews and open an existing review to edit. 2. Click `Update` and note the error on screen.
1.0
fix(admin-review): ensure single reviews can be saved without error - ## User Story As a WPBR user, I want to create and edit single reviews without errors appearing so my changes are saved. ## Acceptance Criteria - [ ] I can add new reviews without error. - [ ] I can edit existing reviews without error. - [ ] There are no console errors on page load of the review edit screen. ## Current Behavior I currently see a PHP notice is displayed on save: > Notice: Trying to get property 'platform' of non-object in /Users/kevin/Sites/wpbrclient5/wp-content/plugins/wp-business-reviews/includes/admin/class-admin-reviews-single.php on line 318 ## Expected Behavior I expect to save a review and have my changes preserved. ## Visuals ### Video https://drive.google.com/file/d/1EXUBRWQ09R_mJmksq9Q53pTNCtemUly6/view?usp=sharing ### Screenshots #### PHP error on save ![image](https://user-images.githubusercontent.com/1938671/40449291-148759e4-5ea6-11e8-931a-0af2e71b2fed.png) #### Console error on page load ![image](https://user-images.githubusercontent.com/1938671/40449348-3ba15aac-5ea6-11e8-9d3e-40d1b2ea2e6d.png) ## Steps to Recreate for New Review 1. Go to Reviews > Add Review 2. Add a title and set a rating. 3. Click `Publish` and note the error on screen. ## Steps to Recreate for Existing Review 1. Go to Reviews and open an existing review to edit. 2. Click `Update` and note the error on screen.
priority
fix admin review ensure single reviews can be saved without error user story as a wpbr user i want to create and edit single reviews without errors appearing so my changes are saved acceptance criteria i can add new reviews without error i can edit existing reviews without error there are no console errors on page load of the review edit screen current behavior i currently see a php notice is displayed on save notice trying to get property platform of non object in users kevin sites wp content plugins wp business reviews includes admin class admin reviews single php on line expected behavior i expect to save a review and have my changes preserved visuals video screenshots php error on save console error on page load steps to recreate for new review go to reviews add review add a title and set a rating click publish and note the error on screen steps to recreate for existing review go to reviews and open an existing review to edit click update and note the error on screen
1
221,319
7,381,902,116
IssuesEvent
2018-03-15 01:29:35
comses/core.comses.net
https://api.github.com/repos/comses/core.comses.net
closed
generate download statistics
feature high priority
create a management command to look at CodebaseReleaseDownloads for the past N months (or a specific start/end date range) and generate a csv with the start/end date range and `codebase URL, name, number of downloads` Might be good to list all the IP addresses as well either in a second table farther or a second sheet.
1.0
generate download statistics - create a management command to look at CodebaseReleaseDownloads for the past N months (or a specific start/end date range) and generate a csv with the start/end date range and `codebase URL, name, number of downloads` Might be good to list all the IP addresses as well either in a second table farther or a second sheet.
priority
generate download statistics create a management command to look at codebasereleasedownloads for the past n months or a specific start end date range and generate a csv with the start end date range and codebase url name number of downloads might be good to list all the ip addresses as well either in a second table farther or a second sheet
1
424,392
12,310,286,120
IssuesEvent
2020-05-12 10:21:05
harmony-one/harmony
https://api.github.com/repos/harmony-one/harmony
closed
[mainnet] mainnet rolling upgrade
devops high priority
Mainnet rolling upgrade, https://github.com/harmony-one/harmony/releases/tag/v1.2.7 - [x] send out an announcement about the rolling upgrade - [x] send out an announcement about the upgrade of node.sh - [x] prepare release build - [x] update all node.sh on mainnet nodes - [x] rolling upgrade on mainnet - [x] build release build during rolling upgrade or after?
1.0
[mainnet] mainnet rolling upgrade - Mainnet rolling upgrade, https://github.com/harmony-one/harmony/releases/tag/v1.2.7 - [x] send out an announcement about the rolling upgrade - [x] send out an announcement about the upgrade of node.sh - [x] prepare release build - [x] update all node.sh on mainnet nodes - [x] rolling upgrade on mainnet - [x] build release build during rolling upgrade or after?
priority
mainnet rolling upgrade mainnet rolling upgrade send out an announcement about the rolling upgrade send out an announcement about the upgrade of node sh prepare release build update all node sh on mainnet nodes rolling upgrade on mainnet build release build during rolling upgrade or after
1
605,381
18,734,412,954
IssuesEvent
2021-11-04 04:20:27
AY2122S1-CS2103T-T11-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-T11-2/tp
closed
Incorrect error message displayed when importing CSV file with additional field
type.Bug priority.High
![image](https://user-images.githubusercontent.com/69179233/140253376-f0814648-cd26-4fdd-a924-7d5456012635.png) ![image](https://user-images.githubusercontent.com/69179233/140253403-2a9769fd-c12b-4a0d-a4e1-a7c97b247000.png) Should tell users that the CSV file contains extraneous values instead of the headers not being in the correct format.
1.0
Incorrect error message displayed when importing CSV file with additional field - ![image](https://user-images.githubusercontent.com/69179233/140253376-f0814648-cd26-4fdd-a924-7d5456012635.png) ![image](https://user-images.githubusercontent.com/69179233/140253403-2a9769fd-c12b-4a0d-a4e1-a7c97b247000.png) Should tell users that the CSV file contains extraneous values instead of the headers not being in the correct format.
priority
incorrect error message displayed when importing csv file with additional field should tell users that the csv file contains extraneous values instead of the headers not being in the correct format
1
669,828
22,642,368,655
IssuesEvent
2022-07-01 04:24:18
MSRevive/MSCScripts
https://api.github.com/repos/MSRevive/MSCScripts
closed
Skeleton-type Enemies (Tested in Thornlands) Reset Their HP to Full When Resurrecting.
bug high priority
- For some reason, whenever I kill a skeleton, the skeleton will fall down and lay on the ground with a health value of 35/70 hp. However, before the skeleton stands up, their health value is reset to 70/70. Skeleton revival behavior is not operating properly, and heals skeletons to full health when they get up again. - How does this behave with other skeleton types? (Notice: This has been tested in Thornlands on a new character. I am going through standard progression.)
1.0
Skeleton-type Enemies (Tested in Thornlands) Reset Their HP to Full When Resurrecting. - - For some reason, whenever I kill a skeleton, the skeleton will fall down and lay on the ground with a health value of 35/70 hp. However, before the skeleton stands up, their health value is reset to 70/70. Skeleton revival behavior is not operating properly, and heals skeletons to full health when they get up again. - How does this behave with other skeleton types? (Notice: This has been tested in Thornlands on a new character. I am going through standard progression.)
priority
skeleton type enemies tested in thornlands reset their hp to full when resurrecting for some reason whenever i kill a skeleton the skeleton will fall down and lay on the ground with a health value of hp however before the skeleton stands up their health value is reset to skeleton revival behavior is not operating properly and heals skeletons to full health when they get up again how does this behave with other skeleton types notice this has been tested in thornlands on a new character i am going through standard progression
1
104,664
4,216,944,555
IssuesEvent
2016-06-30 11:10:02
steedos/apps
https://api.github.com/repos/steedos/apps
opened
选择部门和选择人员字段虽然可修改,但显示成只读的背景色
priority:High
- 控件只读时,增加一个 class 属性 readonly - 编写CSS .selectUser .selectUser.readonly .selectOrg .selectOrg.readonly
1.0
选择部门和选择人员字段虽然可修改,但显示成只读的背景色 - - 控件只读时,增加一个 class 属性 readonly - 编写CSS .selectUser .selectUser.readonly .selectOrg .selectOrg.readonly
priority
选择部门和选择人员字段虽然可修改,但显示成只读的背景色 控件只读时,增加一个 class 属性 readonly 编写css selectuser selectuser readonly selectorg selectorg readonly
1
223,755
7,460,378,482
IssuesEvent
2018-03-30 19:24:55
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: Farming issue
High Priority
**Version:** 0.7.1.2 beta **Steps to Reproduce:** 1. till the ground 2.plant crop **Expected behavior:** to see tilled ground and plants that were planted **Actual behavior:** was not able to see the tilled ground or the plants they were invisible but game still recognizes they are planted
1.0
USER ISSUE: Farming issue - **Version:** 0.7.1.2 beta **Steps to Reproduce:** 1. till the ground 2.plant crop **Expected behavior:** to see tilled ground and plants that were planted **Actual behavior:** was not able to see the tilled ground or the plants they were invisible but game still recognizes they are planted
priority
user issue farming issue version beta steps to reproduce till the ground plant crop expected behavior to see tilled ground and plants that were planted actual behavior was not able to see the tilled ground or the plants they were invisible but game still recognizes they are planted
1
425,174
12,336,756,329
IssuesEvent
2020-05-14 14:02:56
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
Clean up atomix and zeebe parallel classes and hierarchies
Priority: High Status: In Progress Type: Maintenance
**Description** Currently we have a RaftStateMachine implementation, which is used per default in atomix. It is also used in their tests. Furthermore there exist a bunch of classes and implementation of interfaces for snapshots etc, which are set per default but actually replaced by our implementation in the broker. We should remove the uncessary classes and migrate to our classes, such that the tests are also running with our classes. This will reduce some of the unnecessary complexity we currently have.
1.0
Clean up atomix and zeebe parallel classes and hierarchies - **Description** Currently we have a RaftStateMachine implementation, which is used per default in atomix. It is also used in their tests. Furthermore there exist a bunch of classes and implementation of interfaces for snapshots etc, which are set per default but actually replaced by our implementation in the broker. We should remove the uncessary classes and migrate to our classes, such that the tests are also running with our classes. This will reduce some of the unnecessary complexity we currently have.
priority
clean up atomix and zeebe parallel classes and hierarchies description currently we have a raftstatemachine implementation which is used per default in atomix it is also used in their tests furthermore there exist a bunch of classes and implementation of interfaces for snapshots etc which are set per default but actually replaced by our implementation in the broker we should remove the uncessary classes and migrate to our classes such that the tests are also running with our classes this will reduce some of the unnecessary complexity we currently have
1
755,412
26,428,433,143
IssuesEvent
2023-01-14 13:54:32
codersforcauses/wadl
https://api.github.com/repos/codersforcauses/wadl
closed
Set up Middleware
backend frontend difficulty::hard priority::high MVP
## Basic Information **[Give a brief summary of what the issue is about]** - Protect routes - Handle redirects Only admin users should be able to view any page under the admin folder Team Coordinators should only be able to view any pages under the coordinator folder (you'll need to create this folder and move coordinator.vue into it & rename to index.vue)
1.0
Set up Middleware - ## Basic Information **[Give a brief summary of what the issue is about]** - Protect routes - Handle redirects Only admin users should be able to view any page under the admin folder Team Coordinators should only be able to view any pages under the coordinator folder (you'll need to create this folder and move coordinator.vue into it & rename to index.vue)
priority
set up middleware basic information protect routes handle redirects only admin users should be able to view any page under the admin folder team coordinators should only be able to view any pages under the coordinator folder you ll need to create this folder and move coordinator vue into it rename to index vue
1
432,506
12,494,276,715
IssuesEvent
2020-06-01 10:52:42
sodafoundation/SIM
https://api.github.com/repos/sodafoundation/SIM
closed
Update storage Access Info
Feature High Priority
*@NajmudheenCT commented on May 11, 2020, 2:49 AM UTC:* Update already registered storage access info *This issue was moved by [kumarashit](https://github.com/kumarashit) from [sodafoundation/SIM-TempIssues#8](https://github.com/sodafoundation/SIM-TempIssues/issues/8).*
1.0
Update storage Access Info - *@NajmudheenCT commented on May 11, 2020, 2:49 AM UTC:* Update already registered storage access info *This issue was moved by [kumarashit](https://github.com/kumarashit) from [sodafoundation/SIM-TempIssues#8](https://github.com/sodafoundation/SIM-TempIssues/issues/8).*
priority
update storage access info najmudheenct commented on may am utc update already registered storage access info this issue was moved by from
1
542,044
15,837,896,727
IssuesEvent
2021-04-06 21:29:54
QuangTran304/Instagram-Clone
https://api.github.com/repos/QuangTran304/Instagram-Clone
closed
[TASK] Make a user's posts visible on their profile page.
High Priority Medium Priority Task
Description: Find all the posts that have the same username as the user. Make the posts available in the user's profile. Story Points: 5 Risk: Medium Priority: High
2.0
[TASK] Make a user's posts visible on their profile page. - Description: Find all the posts that have the same username as the user. Make the posts available in the user's profile. Story Points: 5 Risk: Medium Priority: High
priority
make a user s posts visible on their profile page description find all the posts that have the same username as the user make the posts available in the user s profile story points risk medium priority high
1
367,150
10,850,092,648
IssuesEvent
2019-11-13 08:00:43
pombase/canto
https://api.github.com/repos/pombase/canto
closed
attribution for primary approval
admin attribution high priority
Because approval takes a long time (you need to read the paper in much the same depth as primary curation), and add annotation, it would be good if individual curators could 'claim' the approved sessions for their personal totals. However at present the 'last approver' is recorded, so if we re-approve a session to fix something (often), tha attribution is lost. Can we have a way of recording the primary approver? In the meantime we will keep a record of our individual approved sessions and at some point we can give you a list of these for most of the ones we already did. Val
1.0
attribution for primary approval - Because approval takes a long time (you need to read the paper in much the same depth as primary curation), and add annotation, it would be good if individual curators could 'claim' the approved sessions for their personal totals. However at present the 'last approver' is recorded, so if we re-approve a session to fix something (often), tha attribution is lost. Can we have a way of recording the primary approver? In the meantime we will keep a record of our individual approved sessions and at some point we can give you a list of these for most of the ones we already did. Val
priority
attribution for primary approval because approval takes a long time you need to read the paper in much the same depth as primary curation and add annotation it would be good if individual curators could claim the approved sessions for their personal totals however at present the last approver is recorded so if we re approve a session to fix something often tha attribution is lost can we have a way of recording the primary approver in the meantime we will keep a record of our individual approved sessions and at some point we can give you a list of these for most of the ones we already did val
1
829,461
31,880,279,377
IssuesEvent
2023-09-16 09:39:07
puppypawpaw/Pawpaw_project_BE
https://api.github.com/repos/puppypawpaw/Pawpaw_project_BE
closed
[feat] 게시글 이미지 업로드, 조회, 삭제 기능 구현
For : API For : Backend Priority : High Status : In Progress Type : Feature
## Description(설명) 게시글의 이미지 업로드, 조회, 삭제 기능을 구현한다. ## Tasks(New feature) - [x] 이미지 업로드 - [x] 이미지 조회 - [x] 이미지 삭제
1.0
[feat] 게시글 이미지 업로드, 조회, 삭제 기능 구현 - ## Description(설명) 게시글의 이미지 업로드, 조회, 삭제 기능을 구현한다. ## Tasks(New feature) - [x] 이미지 업로드 - [x] 이미지 조회 - [x] 이미지 삭제
priority
게시글 이미지 업로드 조회 삭제 기능 구현 description 설명 게시글의 이미지 업로드 조회 삭제 기능을 구현한다 tasks new feature 이미지 업로드 이미지 조회 이미지 삭제
1
591,456
17,840,397,206
IssuesEvent
2021-09-03 09:19:23
francheska-vicente/cssweng
https://api.github.com/repos/francheska-vicente/cssweng
closed
User should be able to enter discounts for both pwd/senior and other discounts
bug priority: high issue: back-end severity: medium issue: validation
### Summary: - Pwd/senior gets removed when adding discount for 'other discounts' and vice versa. - User should be able to enter discounts for both pwd/senior and other discounts. The bigger of the two is the only discount to be honored. ### Steps to Reproduce: 1. login 2. pick any date 3. create booking 4. type 2 in number of pwd 5. select a field under 'other discounts' ### Expected Results: - User should be able to add information for both pwd/senior and other discounts field. The bigger of the two will be honored. ### Actual Results: - User can only choose to add discount to one of the fields at a time | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
1.0
User should be able to enter discounts for both pwd/senior and other discounts - ### Summary: - Pwd/senior gets removed when adding discount for 'other discounts' and vice versa. - User should be able to enter discounts for both pwd/senior and other discounts. The bigger of the two is the only discount to be honored. ### Steps to Reproduce: 1. login 2. pick any date 3. create booking 4. type 2 in number of pwd 5. select a field under 'other discounts' ### Expected Results: - User should be able to add information for both pwd/senior and other discounts field. The bigger of the two will be honored. ### Actual Results: - User can only choose to add discount to one of the fields at a time | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
priority
user should be able to enter discounts for both pwd senior and other discounts summary pwd senior gets removed when adding discount for other discounts and vice versa user should be able to enter discounts for both pwd senior and other discounts the bigger of the two is the only discount to be honored steps to reproduce login pick any date create booking type in number of pwd select a field under other discounts expected results user should be able to add information for both pwd senior and other discounts field the bigger of the two will be honored actual results user can only choose to add discount to one of the fields at a time additional information platform engine google operating system windows
1
792,839
27,975,507,923
IssuesEvent
2023-03-25 14:32:31
LiteLDev/LiteLoaderBDS
https://api.github.com/repos/LiteLDev/LiteLoaderBDS
closed
Crash on Actor::getAllEffects()
type: bug module: LiteLoader status: more info needed priority: high
### 异常模块 LiteLoader (本体) ### 操作系统 Windows Server 2022 ### LiteLoader 版本 2.11.0 ### BDS 版本 1.19.63 ### 发生了什么? MCAPI Actor::getAllEffects()无效且造成崩溃。 经过多次测试,崩溃均发生在Actor::getAllEffects()这个函数。 ### 复现此问题的步骤 获取任意有效的实体指针(非空指针) 执行Actor::getAllEffects()函数。 服务器崩溃。 ### 有关的日志/输出 _No response_ ### 插件列表 _No response_
1.0
Crash on Actor::getAllEffects() - ### 异常模块 LiteLoader (本体) ### 操作系统 Windows Server 2022 ### LiteLoader 版本 2.11.0 ### BDS 版本 1.19.63 ### 发生了什么? MCAPI Actor::getAllEffects()无效且造成崩溃。 经过多次测试,崩溃均发生在Actor::getAllEffects()这个函数。 ### 复现此问题的步骤 获取任意有效的实体指针(非空指针) 执行Actor::getAllEffects()函数。 服务器崩溃。 ### 有关的日志/输出 _No response_ ### 插件列表 _No response_
priority
crash on actor getalleffects 异常模块 liteloader 本体 操作系统 windows server liteloader 版本 bds 版本 发生了什么 mcapi actor getalleffects 无效且造成崩溃。 经过多次测试,崩溃均发生在actor getalleffects 这个函数。 复现此问题的步骤 获取任意有效的实体指针(非空指针) 执行actor getalleffects 函数。 服务器崩溃。 有关的日志 输出 no response 插件列表 no response
1
799,571
28,309,698,615
IssuesEvent
2023-04-10 14:21:20
scientist-softserv/adventist-dl
https://api.github.com/repos/scientist-softserv/adventist-dl
closed
Spike: OCR search not working
Phase3/4 iiif_print High Priority Data Related
# Story Per Katharine, " I searched for a word in the UV and got no results despite seeing it on the page. This work has split into pages. Can you determine if something is wrong with the OCR search in viewer, or are we still waiting on the work to process?" # Acceptance Criteria - [ ] Can we duplicate the issue - [ ] Can we define the resolution for the specific work that isn't correct # Screenshots / Video link to example: https://adl.s2.adventistdigitallibrary.org/concern/published_works/20214381_an_appeal_to_the_youth_funeral_address_of_henry_n_white?locale=en generated ocr: [8b-txt.txt](https://github.com/scientist-softserv/adventist-dl/files/11029761/8b-txt.txt) Search for "Youth" returned no results: <img width="1244" alt="image" src="https://user-images.githubusercontent.com/10081604/226649184-0a0945d9-7d09-4d66-9dcd-772efda2ff72.png"> When looking at this solr doc in rancher, I'm not seeing all_text_tsimv <img width="677" alt="image" src="https://user-images.githubusercontent.com/10081604/226649399-87d3dfda-a7ca-4e79-bd54-9cf2620f0048.png"> Confirmed that this work type has ocr turned on: https://github.com/scientist-softserv/adventist-dl/blob/main/app/models/published_work.rb#L31 # Testing Instructions and Sample Files - create a work via bulkrax or manually - wait for sidekiq to complete and verify that ocr gets generated. - You can see the words it found by clicking on the action dropdown. Select download as txt <img width="1226" alt="image" src="https://user-images.githubusercontent.com/10081604/226650391-b451a4b2-6ffd-4f44-b274-069b86205690.png"> - [ ] a user should be able to search the UV and catalog, for the generated ocr words # Notes This issue was reported before sidekiq completed its jobs, so it may be a timing issue. <img width="1321" alt="image" src="https://user-images.githubusercontent.com/10081604/226649845-8b349015-0197-4c28-b030-c092703342aa.png">
1.0
Spike: OCR search not working - # Story Per Katharine, " I searched for a word in the UV and got no results despite seeing it on the page. This work has split into pages. Can you determine if something is wrong with the OCR search in viewer, or are we still waiting on the work to process?" # Acceptance Criteria - [ ] Can we duplicate the issue - [ ] Can we define the resolution for the specific work that isn't correct # Screenshots / Video link to example: https://adl.s2.adventistdigitallibrary.org/concern/published_works/20214381_an_appeal_to_the_youth_funeral_address_of_henry_n_white?locale=en generated ocr: [8b-txt.txt](https://github.com/scientist-softserv/adventist-dl/files/11029761/8b-txt.txt) Search for "Youth" returned no results: <img width="1244" alt="image" src="https://user-images.githubusercontent.com/10081604/226649184-0a0945d9-7d09-4d66-9dcd-772efda2ff72.png"> When looking at this solr doc in rancher, I'm not seeing all_text_tsimv <img width="677" alt="image" src="https://user-images.githubusercontent.com/10081604/226649399-87d3dfda-a7ca-4e79-bd54-9cf2620f0048.png"> Confirmed that this work type has ocr turned on: https://github.com/scientist-softserv/adventist-dl/blob/main/app/models/published_work.rb#L31 # Testing Instructions and Sample Files - create a work via bulkrax or manually - wait for sidekiq to complete and verify that ocr gets generated. - You can see the words it found by clicking on the action dropdown. Select download as txt <img width="1226" alt="image" src="https://user-images.githubusercontent.com/10081604/226650391-b451a4b2-6ffd-4f44-b274-069b86205690.png"> - [ ] a user should be able to search the UV and catalog, for the generated ocr words # Notes This issue was reported before sidekiq completed its jobs, so it may be a timing issue. <img width="1321" alt="image" src="https://user-images.githubusercontent.com/10081604/226649845-8b349015-0197-4c28-b030-c092703342aa.png">
priority
spike ocr search not working story per katharine i searched for a word in the uv and got no results despite seeing it on the page this work has split into pages can you determine if something is wrong with the ocr search in viewer or are we still waiting on the work to process acceptance criteria can we duplicate the issue can we define the resolution for the specific work that isn t correct screenshots video link to example generated ocr search for youth returned no results img width alt image src when looking at this solr doc in rancher i m not seeing all text tsimv img width alt image src confirmed that this work type has ocr turned on testing instructions and sample files create a work via bulkrax or manually wait for sidekiq to complete and verify that ocr gets generated you can see the words it found by clicking on the action dropdown select download as txt img width alt image src a user should be able to search the uv and catalog for the generated ocr words notes this issue was reported before sidekiq completed its jobs so it may be a timing issue img width alt image src
1
651,810
21,511,109,164
IssuesEvent
2022-04-28 04:40:27
framgia/sph_unity-dkk
https://api.github.com/repos/framgia/sph_unity-dkk
closed
[Documentation][Task] Create Documentation for Loading Levels-Design Sub-Section
High Priority
# Description - Create a numbered step by step process for the section - Divide it by sub-section if needed - Add screenshots if needed # Definition of Done - Step by step detail guide for Loading Levels-Design Sub-Section
1.0
[Documentation][Task] Create Documentation for Loading Levels-Design Sub-Section - # Description - Create a numbered step by step process for the section - Divide it by sub-section if needed - Add screenshots if needed # Definition of Done - Step by step detail guide for Loading Levels-Design Sub-Section
priority
create documentation for loading levels design sub section description create a numbered step by step process for the section divide it by sub section if needed add screenshots if needed definition of done step by step detail guide for loading levels design sub section
1
617,078
19,341,759,727
IssuesEvent
2021-12-15 05:58:20
TeamNADA/NADA-iOS-ForRelease
https://api.github.com/repos/TeamNADA/NADA-iOS-ForRelease
closed
[FIX] 카카오톡 소셜 로그인과 스플래시 뷰 연결, 수정
Minjae 🐻‍❄️ P1 / Priority High Fix
# 👀 이슈 (issue) 카카오톡 소셜 로그인과 관련된 여러 코드를 구현, 수정합니다. # 🚀 to-do <!-- 진행할 작업에 대해 적어주세요 --> - [x] 카카오톡 어플이 설치되지 않았을 때 웹으로 카카오톡이 열리는 로직 수정 - [x] 스플래시 어플에 붙이기 - [x] 스플래시 상황에서 자동 로그인 처리 로직 추가하기 - [x] 토큰을 UserDefaults로 보내는걸 제대로 수정해야 함
1.0
[FIX] 카카오톡 소셜 로그인과 스플래시 뷰 연결, 수정 - # 👀 이슈 (issue) 카카오톡 소셜 로그인과 관련된 여러 코드를 구현, 수정합니다. # 🚀 to-do <!-- 진행할 작업에 대해 적어주세요 --> - [x] 카카오톡 어플이 설치되지 않았을 때 웹으로 카카오톡이 열리는 로직 수정 - [x] 스플래시 어플에 붙이기 - [x] 스플래시 상황에서 자동 로그인 처리 로직 추가하기 - [x] 토큰을 UserDefaults로 보내는걸 제대로 수정해야 함
priority
카카오톡 소셜 로그인과 스플래시 뷰 연결 수정 👀 이슈 issue 카카오톡 소셜 로그인과 관련된 여러 코드를 구현 수정합니다 🚀 to do 카카오톡 어플이 설치되지 않았을 때 웹으로 카카오톡이 열리는 로직 수정 스플래시 어플에 붙이기 스플래시 상황에서 자동 로그인 처리 로직 추가하기 토큰을 userdefaults로 보내는걸 제대로 수정해야 함
1
34,866
2,789,096,637
IssuesEvent
2015-05-08 17:22:21
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
opened
File upload: Consider creating configurable file upload and ingest upload size limits.
Component: File Upload & Handling Priority: High Status: Dev Type: Suggestion
As we gain more experience with our new platform, we may need to set practical limits on uploads and ingests. It should be configurable so we can adjust as we learn more. One suggestion was to present a popup or some message to a user if their upload exceeds a limit, do not upload of course, then give them the option of requesting permission. For ingest same but maybe also allow a choice of upload without ingest in case of ingest size limit?
1.0
File upload: Consider creating configurable file upload and ingest upload size limits. - As we gain more experience with our new platform, we may need to set practical limits on uploads and ingests. It should be configurable so we can adjust as we learn more. One suggestion was to present a popup or some message to a user if their upload exceeds a limit, do not upload of course, then give them the option of requesting permission. For ingest same but maybe also allow a choice of upload without ingest in case of ingest size limit?
priority
file upload consider creating configurable file upload and ingest upload size limits as we gain more experience with our new platform we may need to set practical limits on uploads and ingests it should be configurable so we can adjust as we learn more one suggestion was to present a popup or some message to a user if their upload exceeds a limit do not upload of course then give them the option of requesting permission for ingest same but maybe also allow a choice of upload without ingest in case of ingest size limit
1
308,395
9,438,581,630
IssuesEvent
2019-04-14 01:26:07
code-helix/slatekit
https://api.github.com/repos/code-helix/slatekit
closed
[Alerts] - New alert component for diagnostics
effort_medium feature priority_high
## Overview Create a new Alerts component that can send out generalized alerts to various targets/destinations such as a **Slack Web Hook | PagerDuty | OpsGenie**. ## Details An Alerts component that can send out generalized a generalized Alert or Notification to various destinations. This is allow for notifications/alerts of important events in an application. One of the easiest integrations is to send an alert to a **Slack Web Hook**. Other potential integrations are sending alerts/incidents to **PagerDuty** or **OpsGenie**. A simple alert can have the following structure: ```javascript alert : { area: 'app.registration.new', name: 'NEW_DEVICE_REGISTRATION', uuid: 'abc-123-xyz', desc: 'User registration via mobile', success: true, code: 200, tag : "a1b2c3", fields: [ { title: 'region' , value:'usa' , pii: false }, { title: 'device' , value:'android', pii: false } ] } ``` ## Area Diagnostics and monitoring ## Problem Currently, in Slate Kit ( and our applications using Slate Kit ), there is no way to easily send out structured events/alerts to external systems to notify recipients of important activity occurring in an application. This is needed as a diagnostics / monitoring feature. ## Solution A simple alerts component that for the time being can integrate with **PagerDuty** and/or **Slack Web Hooks** for a start. ## Alternatives Potential integration with **OpsGenie** and **PagerDuty** are idea. However, there should be a distinction between severity level of the alerts so as to not trigger **Incidents**. Aslo, note that structured logs can serve the purpose to some degree as they can be set up to go to loggly and alerts in loggly can be set up to check for specific log messages. However, there should still be some direct alerting component available in Slate Kit and be used at the source of an important event in an application.
1.0
[Alerts] - New alert component for diagnostics - ## Overview Create a new Alerts component that can send out generalized alerts to various targets/destinations such as a **Slack Web Hook | PagerDuty | OpsGenie**. ## Details An Alerts component that can send out generalized a generalized Alert or Notification to various destinations. This is allow for notifications/alerts of important events in an application. One of the easiest integrations is to send an alert to a **Slack Web Hook**. Other potential integrations are sending alerts/incidents to **PagerDuty** or **OpsGenie**. A simple alert can have the following structure: ```javascript alert : { area: 'app.registration.new', name: 'NEW_DEVICE_REGISTRATION', uuid: 'abc-123-xyz', desc: 'User registration via mobile', success: true, code: 200, tag : "a1b2c3", fields: [ { title: 'region' , value:'usa' , pii: false }, { title: 'device' , value:'android', pii: false } ] } ``` ## Area Diagnostics and monitoring ## Problem Currently, in Slate Kit ( and our applications using Slate Kit ), there is no way to easily send out structured events/alerts to external systems to notify recipients of important activity occurring in an application. This is needed as a diagnostics / monitoring feature. ## Solution A simple alerts component that for the time being can integrate with **PagerDuty** and/or **Slack Web Hooks** for a start. ## Alternatives Potential integration with **OpsGenie** and **PagerDuty** are idea. However, there should be a distinction between severity level of the alerts so as to not trigger **Incidents**. Aslo, note that structured logs can serve the purpose to some degree as they can be set up to go to loggly and alerts in loggly can be set up to check for specific log messages. However, there should still be some direct alerting component available in Slate Kit and be used at the source of an important event in an application.
priority
new alert component for diagnostics overview create a new alerts component that can send out generalized alerts to various targets destinations such as a slack web hook pagerduty opsgenie details an alerts component that can send out generalized a generalized alert or notification to various destinations this is allow for notifications alerts of important events in an application one of the easiest integrations is to send an alert to a slack web hook other potential integrations are sending alerts incidents to pagerduty or opsgenie a simple alert can have the following structure javascript alert area app registration new name new device registration uuid abc xyz desc user registration via mobile success true code tag fields title region value usa pii false title device value android pii false area diagnostics and monitoring problem currently in slate kit and our applications using slate kit there is no way to easily send out structured events alerts to external systems to notify recipients of important activity occurring in an application this is needed as a diagnostics monitoring feature solution a simple alerts component that for the time being can integrate with pagerduty and or slack web hooks for a start alternatives potential integration with opsgenie and pagerduty are idea however there should be a distinction between severity level of the alerts so as to not trigger incidents aslo note that structured logs can serve the purpose to some degree as they can be set up to go to loggly and alerts in loggly can be set up to check for specific log messages however there should still be some direct alerting component available in slate kit and be used at the source of an important event in an application
1
157,808
6,016,236,381
IssuesEvent
2017-06-07 06:09:43
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
failing tests for window.open
priority/high QA/no-qa-needed regression
- Did you search for similar issues before submitting this one? yes - Describe the issue you encountered: https://travis-ci.org/brave/browser-laptop/jobs/239785625 - Platform (Win7, 8, 10? macOS? Linux distro?): All - Brave Version (revision SHA): https://github.com/brave/browser-laptop/commit/c531f2b7cbc4b59ff9e163c164e60e7ce68cc629
1.0
failing tests for window.open - - Did you search for similar issues before submitting this one? yes - Describe the issue you encountered: https://travis-ci.org/brave/browser-laptop/jobs/239785625 - Platform (Win7, 8, 10? macOS? Linux distro?): All - Brave Version (revision SHA): https://github.com/brave/browser-laptop/commit/c531f2b7cbc4b59ff9e163c164e60e7ce68cc629
priority
failing tests for window open did you search for similar issues before submitting this one yes describe the issue you encountered platform macos linux distro all brave version revision sha
1
71,137
3,352,411,348
IssuesEvent
2015-11-17 22:37:35
jgirald/ES2015C
https://api.github.com/repos/jgirald/ES2015C
closed
Implement the game initialization condition (create initial buildings and units depending on civ.)
Development High Priority Team C
Currently, the initial elements of the game are the ones that are initially located in the scene file, and thus are fixed for all configurations. This add several limitation, e.g. the number of players is fixed, and the initial buildings are the same regardless of the chosen civilization. The purpose of this issue is to implement a script, which should run immediately after the game is started, which will spawn the initial player settings, buildings and units according to the parameters chosen in the menu, and taking into account the different particularities of each civilization. For example, if the player chooses the Yamato civilization, a yamato civil and a yamato town center must be created. **Product backlog item:** As a user, I want to select one civilization to play. [High priority] **Estimated effort:** 3 hours **Ending condition:** The scene file should be initially devoid of any building or unit. When the game is started, a town center building, a civil unit, and a military unit will be created, corresponding to the civilization that the player has chosen in the menu. All those elements must have their correct scripts associated, including AI scripts for the enemy civilization if available.
1.0
Implement the game initialization condition (create initial buildings and units depending on civ.) - Currently, the initial elements of the game are the ones that are initially located in the scene file, and thus are fixed for all configurations. This add several limitation, e.g. the number of players is fixed, and the initial buildings are the same regardless of the chosen civilization. The purpose of this issue is to implement a script, which should run immediately after the game is started, which will spawn the initial player settings, buildings and units according to the parameters chosen in the menu, and taking into account the different particularities of each civilization. For example, if the player chooses the Yamato civilization, a yamato civil and a yamato town center must be created. **Product backlog item:** As a user, I want to select one civilization to play. [High priority] **Estimated effort:** 3 hours **Ending condition:** The scene file should be initially devoid of any building or unit. When the game is started, a town center building, a civil unit, and a military unit will be created, corresponding to the civilization that the player has chosen in the menu. All those elements must have their correct scripts associated, including AI scripts for the enemy civilization if available.
priority
implement the game initialization condition create initial buildings and units depending on civ currently the initial elements of the game are the ones that are initially located in the scene file and thus are fixed for all configurations this add several limitation e g the number of players is fixed and the initial buildings are the same regardless of the chosen civilization the purpose of this issue is to implement a script which should run immediately after the game is started which will spawn the initial player settings buildings and units according to the parameters chosen in the menu and taking into account the different particularities of each civilization for example if the player chooses the yamato civilization a yamato civil and a yamato town center must be created product backlog item as a user i want to select one civilization to play estimated effort hours ending condition the scene file should be initially devoid of any building or unit when the game is started a town center building a civil unit and a military unit will be created corresponding to the civilization that the player has chosen in the menu all those elements must have their correct scripts associated including ai scripts for the enemy civilization if available
1
584,223
17,408,931,616
IssuesEvent
2021-08-03 09:44:29
turbot/steampipe-plugin-aws
https://api.github.com/repos/turbot/steampipe-plugin-aws
closed
Add table aws_emr_cluster_metric_isidle
enhancement new table priority:high
**References** https://aws.amazon.com/blogs/big-data/optimize-amazon-emr-costs-with-idle-checks-and[…]tion-using-advanced-amazon-cloudwatch-metrics-and-aws-lambda/ Use case: Detect clusters that are idle . The IsIdle metric tracks whether a cluster is live, but not currently running tasks. You can set an alarm to fire when the cluster has been idle for a given period of time, such as thirty minutes.
1.0
Add table aws_emr_cluster_metric_isidle - **References** https://aws.amazon.com/blogs/big-data/optimize-amazon-emr-costs-with-idle-checks-and[…]tion-using-advanced-amazon-cloudwatch-metrics-and-aws-lambda/ Use case: Detect clusters that are idle . The IsIdle metric tracks whether a cluster is live, but not currently running tasks. You can set an alarm to fire when the cluster has been idle for a given period of time, such as thirty minutes.
priority
add table aws emr cluster metric isidle references tion using advanced amazon cloudwatch metrics and aws lambda use case detect clusters that are idle the isidle metric tracks whether a cluster is live but not currently running tasks you can set an alarm to fire when the cluster has been idle for a given period of time such as thirty minutes
1
439,310
12,681,193,362
IssuesEvent
2020-06-19 14:58:43
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
make text more readable
chore high priority
Chuck says: > One thing that would help me with the current text-based interface would be if the text were somehow “darker”. I don’t mean bold…but it to me it looks faded compared to things around it. For instance this email is easy to read on my Mac but mb2 isn’t as easy. (Notice the difference between “Tournament of Champions style 2019” and what’s in the chat box.) I'll have to bring up an inspector to see all the text attributes. One big difference is that the text in the "log" boxes is that it's fixed width. Chat doesn't need to be (and shoudn't be!) fixed width, but the various llnes that display cards need to be, because that's how I currently line things up (even though it's problematic -- see #104). The obvious fix is to get all the lines that require alignment to be in an area that is separate from the chat and then just use standard HTML/CSS techniques to do the alignment. Although I don't have an issue for that, it is implicit to some of the other issues (e.g., but not limited to, #104). Unfortunately, the obvious fix requires a little time that I don't have now, but Chuck is Chuck and therefore I'll look for a stop-gap that improves the situation. If I don't find one, I'll at least document what I looked at.
1.0
make text more readable - Chuck says: > One thing that would help me with the current text-based interface would be if the text were somehow “darker”. I don’t mean bold…but it to me it looks faded compared to things around it. For instance this email is easy to read on my Mac but mb2 isn’t as easy. (Notice the difference between “Tournament of Champions style 2019” and what’s in the chat box.) I'll have to bring up an inspector to see all the text attributes. One big difference is that the text in the "log" boxes is that it's fixed width. Chat doesn't need to be (and shoudn't be!) fixed width, but the various llnes that display cards need to be, because that's how I currently line things up (even though it's problematic -- see #104). The obvious fix is to get all the lines that require alignment to be in an area that is separate from the chat and then just use standard HTML/CSS techniques to do the alignment. Although I don't have an issue for that, it is implicit to some of the other issues (e.g., but not limited to, #104). Unfortunately, the obvious fix requires a little time that I don't have now, but Chuck is Chuck and therefore I'll look for a stop-gap that improves the situation. If I don't find one, I'll at least document what I looked at.
priority
make text more readable chuck says one thing that would help me with the current text based interface would be if the text were somehow “darker” i don’t mean bold…but it to me it looks faded compared to things around it for instance this email is easy to read on my mac but isn’t as easy notice the difference between “tournament of champions style ” and what’s in the chat box i ll have to bring up an inspector to see all the text attributes one big difference is that the text in the log boxes is that it s fixed width chat doesn t need to be and shoudn t be fixed width but the various llnes that display cards need to be because that s how i currently line things up even though it s problematic see the obvious fix is to get all the lines that require alignment to be in an area that is separate from the chat and then just use standard html css techniques to do the alignment although i don t have an issue for that it is implicit to some of the other issues e g but not limited to unfortunately the obvious fix requires a little time that i don t have now but chuck is chuck and therefore i ll look for a stop gap that improves the situation if i don t find one i ll at least document what i looked at
1
71,165
3,352,769,850
IssuesEvent
2015-11-18 00:28:50
pufexi/multiorder
https://api.github.com/repos/pufexi/multiorder
opened
OV-AT a OV = ruzne msg Pripraveno k vyzvednuti
high priority
Budeme mit nove vydejni misto, a tak se bude ridit rozesilani zprav Pripraveno k vyzvednuti a Upozorneni k osobnimu vyzvednuti, viz sablony nize, ridit dle typu OV. Udelal bych format, ze bude OV-XX, tj. tohle puvodni OV muzeme prejmenovat klidne na OV-GK, to nove, co budem pouzivat je jiz pridane jako OV-AT, takze by se to mohlo automaticky propisovat i do sablon napr.pickinfo_cs_ov-at.latte atd... pickinfo_cs.latte unpicked_ov_cs.latte
1.0
OV-AT a OV = ruzne msg Pripraveno k vyzvednuti - Budeme mit nove vydejni misto, a tak se bude ridit rozesilani zprav Pripraveno k vyzvednuti a Upozorneni k osobnimu vyzvednuti, viz sablony nize, ridit dle typu OV. Udelal bych format, ze bude OV-XX, tj. tohle puvodni OV muzeme prejmenovat klidne na OV-GK, to nove, co budem pouzivat je jiz pridane jako OV-AT, takze by se to mohlo automaticky propisovat i do sablon napr.pickinfo_cs_ov-at.latte atd... pickinfo_cs.latte unpicked_ov_cs.latte
priority
ov at a ov ruzne msg pripraveno k vyzvednuti budeme mit nove vydejni misto a tak se bude ridit rozesilani zprav pripraveno k vyzvednuti a upozorneni k osobnimu vyzvednuti viz sablony nize ridit dle typu ov udelal bych format ze bude ov xx tj tohle puvodni ov muzeme prejmenovat klidne na ov gk to nove co budem pouzivat je jiz pridane jako ov at takze by se to mohlo automaticky propisovat i do sablon napr pickinfo cs ov at latte atd pickinfo cs latte unpicked ov cs latte
1
322,771
9,828,437,884
IssuesEvent
2019-06-15 11:37:49
dankelley/oce
https://api.github.com/repos/dankelley/oce
closed
"package" to "PACKAGE" arg error (requires CRAN resubmission)
high priority
(I already emailed this to @richardsc, but it will be new to @clayton33.) Below is a quote from an email sent to several package maintainers by a CRAN organizer. The fix is very easy, and I'll do it right away. As the last line of the email says, oce will need to be resubmitted to CRAN by the end of the present month. I don't want to get tempted into trying to fix outstanding bugs in a rush, because that might cement bad new code for a year. However, it might be worth trying to address other issues that seem (a) important, (b) easy to fix and (c) easy to verify. Q: do either of the two of you see such issues? PS. I don't want to leave this for too long, in case something else comes up. It always seems that a submission finds new problems that the checking programs don't identify, and I don't want a terrible panic in the last few days of June. where PACKAGE= seems intended. BayesSingleSub/R/trendAR.R: out = .Call("MCAR_trend", y, N, Nmiss, as.integer(distMat), alphaTheta, betaTheta, r.scaleInt^2, r.scaleSlp^2, X[,c(1,3)], X[,c(2,4)], iterations, progress, pbFun, new.env(), package="BayesSingleSub") BayesSingleSub/R/trendAR.R: ret = .Call("MCnullMargLogLikeAR_trend",theta, Nmiss, as.integer(distMat), y,N,alphaTheta,betaTheta,X0,package="BayesFactorPCL") BayesSingleSub/R/trendAR.R: ret = .Call("MCmargLogLikeAR_trendR",theta, Nmiss, as.integer(distMat), g1,g2,y,N,alphaTheta,betaTheta,rInt,rSlp,X0,X1,package="BayesFactorPCL") BayesSingleSub/R/utility.R: .Call("RLogMeanExpLogs", as.numeric(v), N, package="BayesSingleSub") BayesSingleSub/R/notrendAR.R: logbf = .Call("RMCTwoSampleAR", as.numeric(y), N, as.integer(distMat), treat, r.scale, alphaTheta, betaTheta, iterations, package="BayesSingleSub") - mlike0.gq BayesSingleSub/R/multilevelTS.R: chains = .Call("RgibbsTwoSampleARmulti", y, N, Nobs, treat, r.scale, betaTheta, iterations, sdMet, progress, pbFun, new.env(), package="BayesFactorPCL") BayesSingleSub/R/multilevelTS.R: .Call("RthetaLogLikeARmulti", theta, mu0, beta, sig2, g, y, as.integer(N), t, as.integer(Nobs), as.integer(cumobs), as.integer(maxobs), betaTheta, package="BayesFactorPCL") BoutrosLab.plotting.general/R/dist.R: .Call("Cdist", x, method, attrs, p, package="BoutrosLab.plotting.general") DSL/R/utils.R: .Call("_collector2", x, y, package = "DSL") HapEstXXR/R/powerset.R: psl <- .Call ("powerset", x, package = "HapEstXXR") SCCI/R/sc.R: ret = .Call("conditionalShannonEntropy", x, data.matrix(y), package="SCCI") SCCI/R/sc.R: ret = .Call("shannonEntropy", x, package="SCCI") SCCI/R/sc.R: ret = .Call(fCall, data.matrix(x), data.matrix(y), package="SCCI") SCCI/R/sc.R: ret = .Call(fcall, data.matrix(x), data.matrix(y), data.matrix(data.frame(x,y)), data.matrix(Z), package="SCCI") SCCI/R/sc.R: cxgy = .Call(fcall, data.matrix(x), data.matrix(y), data.matrix(Z), package="SCCI") SCCI/R/sc.R: cygx = .Call(fcall, data.matrix(y), data.matrix(x), data.matrix(Z), package="SCCI") cpgen/R/cSSBR.R: .Call("get_generation", pede$sire, pede$dam, pede$id, pede$gene, as.integer(verbose), package="cpgen") geometry/R/tsearch.R: out <- .Call("C_tsearch_orig", x, y, t, xi, yi, bary, package="geometry") growcurves/R/ddpgrow.R: res <- .Call("DDP", y,X,Z,subject,dosemat,numt,typet,Omega,omegaplus,n.iter,n.burn,n.thin,shapealph,ratebeta,M.init, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmCplusDP", y, X, Z, Wcase, Wsubject, Omega, omegaplus, groups, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmIgroupDP", y, X, Z, Wcase, Wsubject, M, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmIplusDP", y, X, Z, Wcase, Wsubject, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmCmvplusDP", y, X, Z, H, Wcase, Wsubject, Omega, omegaplus, groups, subjects, niter, nburn, nthin, strength.mm, corsess, shapealph, ratebeta, typemm, package = "growcurves") growcurves/R/dpgrowmult.R: res <- .Call("mmmult", y, X, Z, Wcases, Mmats, Omegas, omegapluses, ngs, subjects, typet, niter, nburn, nthin, shapealph, ratebeta, strength.mm, package = "growcurves") growcurves/R/dpgrow.R: res <- .Call("DPre", y, X, Z, subjects, niter, nburn, nthin, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrow.R: res <- .Call("lgm", y, X, Z, subjects, niter, nburn, nthin, package = "growcurves") simPop/R/calibVars.R: res <- .Call("simPop_binary_representation", levels=1:length(levels(x)), values=as.integer(x), package="simPop") simPop/R/ipu.r: w <- .Call("simPop_ipu_work", inp, con, w, eps=eps, verbose=ifelse(verbose, 1L, 0L), package="simPop") DynamicGP/R/gp_sep.r: .C("deleteGPseps_R", package="DynamicGP") DynamicGP/R/gp_sep.r: .C("getmGPsep_R", gpsepi = as.integer(gpsepi), m = integer(1), package="DynamicGP")$m DynamicGP/R/gp_sep.r: .C("getdGPsep_R", gpsepi = as.integer(gpsepi), d = double(m), package="DynamicGP")$d DynamicGP/R/gp_sep.r: .C("getgGPsep_R", gpsepi = as.integer(gpsepi), g = double(1), package="DynamicGP")$g DynamicGP/R/gpseplm.r: .C("getmGPsep_R", gpsepi = as.integer(gpsepi), m = integer(1), package="DynamicGP")$m DynamicGP/R/gpseplm.r: .C("getmGPsepLm_R", gplmi = as.integer(gplmi), m = integer(1), package = "DynamicGP")$m HellCor/R/HellCor.R: res <- .C("hellcorC", as.double(cbind(x, y)), as.integer(2 * length(x)), statistic = as.double(0.0), as.integer(pvalue), pvalue = as.double(0.0), as.integer(KLmax), as.double(alpha), package = "HellCor") HellCor/R/HellCor.R: pval <- mean(replicate(10000, .C("hellcorC", as.double(cbind(runif(n), y)), as.integer(twon), statistic = as.double(0.0), as.integer(0), as.double(0.0), as.integer(2), as.double(alpha), package = "HellCor")$statistic) > res$statistic) SAM/R/samHL.R: out = .C("grpSVM", Z = as.double(Z), lambda = as.double(lambda), nnlambda = as.integer(nlambda), LL0 = as.double(L0), nn = as.integer(n), dd = as.integer(d), pp = as.integer(p),aa0 = as.double(a0), xx = as.double(matrix(0,m+1,nlambda)), mmu = as.double(mu), mmax_ite = as.integer(max.ite), tthol = as.double(thol),aalpha = as.double(0.5),df=as.double(rep(0,nlambda)),func_norm=as.double(matrix(0,d,nlambda)),package="SAM") SAM/R/samQL.R: out = .C("grplasso",y = as.double(y), X = as.double(Z), lambda = as.double(lambda), nnlambda = as.integer(nlambda), nn = as.integer(n), dd = as.integer(d), pp = as.integer(p), ww = as.double(matrix(0,m,nlambda)), mmax_ite = as.integer(max.ite), tthol = as.double(thol), regfunc = as.character(regfunc), iinput = as.integer(lambda_input), df=as.integer(rep(0,nlambda)), sse=as.double(rep(0,nlambda)), func_norm = as.double(matrix(0,d,nlambda)), package="SAM") flare/R/sugm.generator.R: out = .C("SFGen",dd0=as.integer(2),dd=as.integer(d),G=as.integer(theta),seed=as.integer(seed),package="flare") genMOSS/R/findLogMargLik.R: .C("findLogMargLik", as.integer(tData[model,]), as.integer(dim(tData)[2]), as.integer(length(model)), as.integer(dimens[model]), as.double(alpha), logMargLik = as.double(logMargLik), package = "c_code")$logMargLik genMOSSplus/R/findLogMargLik.R: .C("findLogMargLik", as.integer(tData[model,]), as.integer(dim(tData)[2]), as.integer(length(model)), as.integer(dimens[model]), as.double(alpha), logMargLik = as.double(logMargLik), package = "c_code")$logMargLik [ What id "c_code"?] longclust/R/t_chol_new.R: res <- .C("tchol_entry", as.double(x), as.integer(class), as.integer(N), as.integer(p), as.integer(Gmin), as.integer(Gmax), as.integer(df_flag), as.integer(gaussian_flag), as.integer(models_selected), as.integer(use_kmeans), as.double(kmeansz), Gbest = as.integer(0), Gbest_icl = as.integer(0), zbest = double(N*(Gmax+1)), zbest_icl = double(N*(Gmax+1)), nubest = double(Gmax+1), nubest_icl = double(Gmax+1), mubest = double((Gmax+1)*p), mubest_icl = double((Gmax+1)*p), Tbest = double((Gmax+1)*p*p), Tbest_icl = double((Gmax+1)*p*p), Dbest = double((Gmax+1)*p*p), Dbest_icl = double((Gmax+1)*p*p), llres = double(8*(Gmax+1)), bicres = double(8*(Gmax+1)), iclres = double(8*(Gmax+1)), package="longclust"); oce/R/echosounder.R: .C("biosonics_free_storage", package="oce") sparc/R/sparc.R: out.SGLM = .C("SGLM", A = as.double(rep(0,N*d)), X = as.double(X), y = as.double(y), NN = as.integer(N), nn = as.integer(n), dd = as.integer(d),lambda = as.double(rep(0,d)),package="sparc") sparc/R/sparc.R: out.SLRM = .C("SLRM", A = as.double(A), lambda = as.double(lambda), nnlambda = as.integer(nlambda),LL0 = as.double(L0), nn = as.integer(N), dd = as.integer(d), ww = as.double(rep(0,d*nlambda)), mmax_ite = as.integer(max.ite), tthol = as.double(thol), aalpha = as.double(alpha),package="sparc") This will show up as the wrong number of arguments in routine registration, so you will need to revise that. Please correct ASAP and before July 1 to safely retain the package on CRAN.
1.0
"package" to "PACKAGE" arg error (requires CRAN resubmission) - (I already emailed this to @richardsc, but it will be new to @clayton33.) Below is a quote from an email sent to several package maintainers by a CRAN organizer. The fix is very easy, and I'll do it right away. As the last line of the email says, oce will need to be resubmitted to CRAN by the end of the present month. I don't want to get tempted into trying to fix outstanding bugs in a rush, because that might cement bad new code for a year. However, it might be worth trying to address other issues that seem (a) important, (b) easy to fix and (c) easy to verify. Q: do either of the two of you see such issues? PS. I don't want to leave this for too long, in case something else comes up. It always seems that a submission finds new problems that the checking programs don't identify, and I don't want a terrible panic in the last few days of June. where PACKAGE= seems intended. BayesSingleSub/R/trendAR.R: out = .Call("MCAR_trend", y, N, Nmiss, as.integer(distMat), alphaTheta, betaTheta, r.scaleInt^2, r.scaleSlp^2, X[,c(1,3)], X[,c(2,4)], iterations, progress, pbFun, new.env(), package="BayesSingleSub") BayesSingleSub/R/trendAR.R: ret = .Call("MCnullMargLogLikeAR_trend",theta, Nmiss, as.integer(distMat), y,N,alphaTheta,betaTheta,X0,package="BayesFactorPCL") BayesSingleSub/R/trendAR.R: ret = .Call("MCmargLogLikeAR_trendR",theta, Nmiss, as.integer(distMat), g1,g2,y,N,alphaTheta,betaTheta,rInt,rSlp,X0,X1,package="BayesFactorPCL") BayesSingleSub/R/utility.R: .Call("RLogMeanExpLogs", as.numeric(v), N, package="BayesSingleSub") BayesSingleSub/R/notrendAR.R: logbf = .Call("RMCTwoSampleAR", as.numeric(y), N, as.integer(distMat), treat, r.scale, alphaTheta, betaTheta, iterations, package="BayesSingleSub") - mlike0.gq BayesSingleSub/R/multilevelTS.R: chains = .Call("RgibbsTwoSampleARmulti", y, N, Nobs, treat, r.scale, betaTheta, iterations, sdMet, progress, pbFun, new.env(), package="BayesFactorPCL") BayesSingleSub/R/multilevelTS.R: .Call("RthetaLogLikeARmulti", theta, mu0, beta, sig2, g, y, as.integer(N), t, as.integer(Nobs), as.integer(cumobs), as.integer(maxobs), betaTheta, package="BayesFactorPCL") BoutrosLab.plotting.general/R/dist.R: .Call("Cdist", x, method, attrs, p, package="BoutrosLab.plotting.general") DSL/R/utils.R: .Call("_collector2", x, y, package = "DSL") HapEstXXR/R/powerset.R: psl <- .Call ("powerset", x, package = "HapEstXXR") SCCI/R/sc.R: ret = .Call("conditionalShannonEntropy", x, data.matrix(y), package="SCCI") SCCI/R/sc.R: ret = .Call("shannonEntropy", x, package="SCCI") SCCI/R/sc.R: ret = .Call(fCall, data.matrix(x), data.matrix(y), package="SCCI") SCCI/R/sc.R: ret = .Call(fcall, data.matrix(x), data.matrix(y), data.matrix(data.frame(x,y)), data.matrix(Z), package="SCCI") SCCI/R/sc.R: cxgy = .Call(fcall, data.matrix(x), data.matrix(y), data.matrix(Z), package="SCCI") SCCI/R/sc.R: cygx = .Call(fcall, data.matrix(y), data.matrix(x), data.matrix(Z), package="SCCI") cpgen/R/cSSBR.R: .Call("get_generation", pede$sire, pede$dam, pede$id, pede$gene, as.integer(verbose), package="cpgen") geometry/R/tsearch.R: out <- .Call("C_tsearch_orig", x, y, t, xi, yi, bary, package="geometry") growcurves/R/ddpgrow.R: res <- .Call("DDP", y,X,Z,subject,dosemat,numt,typet,Omega,omegaplus,n.iter,n.burn,n.thin,shapealph,ratebeta,M.init, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmCplusDP", y, X, Z, Wcase, Wsubject, Omega, omegaplus, groups, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmIgroupDP", y, X, Z, Wcase, Wsubject, M, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmIplusDP", y, X, Z, Wcase, Wsubject, subjects, niter, nburn, nthin, strength.mm, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrowmm.R: res <- .Call("mmCmvplusDP", y, X, Z, H, Wcase, Wsubject, Omega, omegaplus, groups, subjects, niter, nburn, nthin, strength.mm, corsess, shapealph, ratebeta, typemm, package = "growcurves") growcurves/R/dpgrowmult.R: res <- .Call("mmmult", y, X, Z, Wcases, Mmats, Omegas, omegapluses, ngs, subjects, typet, niter, nburn, nthin, shapealph, ratebeta, strength.mm, package = "growcurves") growcurves/R/dpgrow.R: res <- .Call("DPre", y, X, Z, subjects, niter, nburn, nthin, shapealph, ratebeta, package = "growcurves") growcurves/R/dpgrow.R: res <- .Call("lgm", y, X, Z, subjects, niter, nburn, nthin, package = "growcurves") simPop/R/calibVars.R: res <- .Call("simPop_binary_representation", levels=1:length(levels(x)), values=as.integer(x), package="simPop") simPop/R/ipu.r: w <- .Call("simPop_ipu_work", inp, con, w, eps=eps, verbose=ifelse(verbose, 1L, 0L), package="simPop") DynamicGP/R/gp_sep.r: .C("deleteGPseps_R", package="DynamicGP") DynamicGP/R/gp_sep.r: .C("getmGPsep_R", gpsepi = as.integer(gpsepi), m = integer(1), package="DynamicGP")$m DynamicGP/R/gp_sep.r: .C("getdGPsep_R", gpsepi = as.integer(gpsepi), d = double(m), package="DynamicGP")$d DynamicGP/R/gp_sep.r: .C("getgGPsep_R", gpsepi = as.integer(gpsepi), g = double(1), package="DynamicGP")$g DynamicGP/R/gpseplm.r: .C("getmGPsep_R", gpsepi = as.integer(gpsepi), m = integer(1), package="DynamicGP")$m DynamicGP/R/gpseplm.r: .C("getmGPsepLm_R", gplmi = as.integer(gplmi), m = integer(1), package = "DynamicGP")$m HellCor/R/HellCor.R: res <- .C("hellcorC", as.double(cbind(x, y)), as.integer(2 * length(x)), statistic = as.double(0.0), as.integer(pvalue), pvalue = as.double(0.0), as.integer(KLmax), as.double(alpha), package = "HellCor") HellCor/R/HellCor.R: pval <- mean(replicate(10000, .C("hellcorC", as.double(cbind(runif(n), y)), as.integer(twon), statistic = as.double(0.0), as.integer(0), as.double(0.0), as.integer(2), as.double(alpha), package = "HellCor")$statistic) > res$statistic) SAM/R/samHL.R: out = .C("grpSVM", Z = as.double(Z), lambda = as.double(lambda), nnlambda = as.integer(nlambda), LL0 = as.double(L0), nn = as.integer(n), dd = as.integer(d), pp = as.integer(p),aa0 = as.double(a0), xx = as.double(matrix(0,m+1,nlambda)), mmu = as.double(mu), mmax_ite = as.integer(max.ite), tthol = as.double(thol),aalpha = as.double(0.5),df=as.double(rep(0,nlambda)),func_norm=as.double(matrix(0,d,nlambda)),package="SAM") SAM/R/samQL.R: out = .C("grplasso",y = as.double(y), X = as.double(Z), lambda = as.double(lambda), nnlambda = as.integer(nlambda), nn = as.integer(n), dd = as.integer(d), pp = as.integer(p), ww = as.double(matrix(0,m,nlambda)), mmax_ite = as.integer(max.ite), tthol = as.double(thol), regfunc = as.character(regfunc), iinput = as.integer(lambda_input), df=as.integer(rep(0,nlambda)), sse=as.double(rep(0,nlambda)), func_norm = as.double(matrix(0,d,nlambda)), package="SAM") flare/R/sugm.generator.R: out = .C("SFGen",dd0=as.integer(2),dd=as.integer(d),G=as.integer(theta),seed=as.integer(seed),package="flare") genMOSS/R/findLogMargLik.R: .C("findLogMargLik", as.integer(tData[model,]), as.integer(dim(tData)[2]), as.integer(length(model)), as.integer(dimens[model]), as.double(alpha), logMargLik = as.double(logMargLik), package = "c_code")$logMargLik genMOSSplus/R/findLogMargLik.R: .C("findLogMargLik", as.integer(tData[model,]), as.integer(dim(tData)[2]), as.integer(length(model)), as.integer(dimens[model]), as.double(alpha), logMargLik = as.double(logMargLik), package = "c_code")$logMargLik [ What id "c_code"?] longclust/R/t_chol_new.R: res <- .C("tchol_entry", as.double(x), as.integer(class), as.integer(N), as.integer(p), as.integer(Gmin), as.integer(Gmax), as.integer(df_flag), as.integer(gaussian_flag), as.integer(models_selected), as.integer(use_kmeans), as.double(kmeansz), Gbest = as.integer(0), Gbest_icl = as.integer(0), zbest = double(N*(Gmax+1)), zbest_icl = double(N*(Gmax+1)), nubest = double(Gmax+1), nubest_icl = double(Gmax+1), mubest = double((Gmax+1)*p), mubest_icl = double((Gmax+1)*p), Tbest = double((Gmax+1)*p*p), Tbest_icl = double((Gmax+1)*p*p), Dbest = double((Gmax+1)*p*p), Dbest_icl = double((Gmax+1)*p*p), llres = double(8*(Gmax+1)), bicres = double(8*(Gmax+1)), iclres = double(8*(Gmax+1)), package="longclust"); oce/R/echosounder.R: .C("biosonics_free_storage", package="oce") sparc/R/sparc.R: out.SGLM = .C("SGLM", A = as.double(rep(0,N*d)), X = as.double(X), y = as.double(y), NN = as.integer(N), nn = as.integer(n), dd = as.integer(d),lambda = as.double(rep(0,d)),package="sparc") sparc/R/sparc.R: out.SLRM = .C("SLRM", A = as.double(A), lambda = as.double(lambda), nnlambda = as.integer(nlambda),LL0 = as.double(L0), nn = as.integer(N), dd = as.integer(d), ww = as.double(rep(0,d*nlambda)), mmax_ite = as.integer(max.ite), tthol = as.double(thol), aalpha = as.double(alpha),package="sparc") This will show up as the wrong number of arguments in routine registration, so you will need to revise that. Please correct ASAP and before July 1 to safely retain the package on CRAN.
priority
package to package arg error requires cran resubmission i already emailed this to richardsc but it will be new to below is a quote from an email sent to several package maintainers by a cran organizer the fix is very easy and i ll do it right away as the last line of the email says oce will need to be resubmitted to cran by the end of the present month i don t want to get tempted into trying to fix outstanding bugs in a rush because that might cement bad new code for a year however it might be worth trying to address other issues that seem a important b easy to fix and c easy to verify q do either of the two of you see such issues ps i don t want to leave this for too long in case something else comes up it always seems that a submission finds new problems that the checking programs don t identify and i don t want a terrible panic in the last few days of june where package seems intended bayessinglesub r trendar r out call mcar trend y n nmiss as integer distmat alphatheta betatheta r scaleint r scaleslp x x iterations progress pbfun new env package bayessinglesub bayessinglesub r trendar r ret call mcnullmargloglikear trend theta nmiss as integer distmat y n alphatheta betatheta package bayesfactorpcl bayessinglesub r trendar r ret call mcmargloglikear trendr theta nmiss as integer distmat y n alphatheta betatheta rint rslp package bayesfactorpcl bayessinglesub r utility r call rlogmeanexplogs as numeric v n package bayessinglesub bayessinglesub r notrendar r logbf call rmctwosamplear as numeric y n as integer distmat treat r scale alphatheta betatheta iterations package bayessinglesub gq bayessinglesub r multilevelts r chains call rgibbstwosamplearmulti y n nobs treat r scale betatheta iterations sdmet progress pbfun new env package bayesfactorpcl bayessinglesub r multilevelts r call rthetaloglikearmulti theta beta g y as integer n t as integer nobs as integer cumobs as integer maxobs betatheta package bayesfactorpcl boutroslab plotting general r dist r call cdist x method attrs p package boutroslab plotting general dsl r utils r call x y package dsl hapestxxr r powerset r psl call powerset x package hapestxxr scci r sc r ret call conditionalshannonentropy x data matrix y package scci scci r sc r ret call shannonentropy x package scci scci r sc r ret call fcall data matrix x data matrix y package scci scci r sc r ret call fcall data matrix x data matrix y data matrix data frame x y data matrix z package scci scci r sc r cxgy call fcall data matrix x data matrix y data matrix z package scci scci r sc r cygx call fcall data matrix y data matrix x data matrix z package scci cpgen r cssbr r call get generation pede sire pede dam pede id pede gene as integer verbose package cpgen geometry r tsearch r out call c tsearch orig x y t xi yi bary package geometry growcurves r ddpgrow r res call ddp y x z subject dosemat numt typet omega omegaplus n iter n burn n thin shapealph ratebeta m init package growcurves growcurves r dpgrowmm r res call mmcplusdp y x z wcase wsubject omega omegaplus groups subjects niter nburn nthin strength mm shapealph ratebeta package growcurves growcurves r dpgrowmm r res call mmigroupdp y x z wcase wsubject m subjects niter nburn nthin strength mm shapealph ratebeta package growcurves growcurves r dpgrowmm r res call mmiplusdp y x z wcase wsubject subjects niter nburn nthin strength mm shapealph ratebeta package growcurves growcurves r dpgrowmm r res call mmcmvplusdp y x z h wcase wsubject omega omegaplus groups subjects niter nburn nthin strength mm corsess shapealph ratebeta typemm package growcurves growcurves r dpgrowmult r res call mmmult y x z wcases mmats omegas omegapluses ngs subjects typet niter nburn nthin shapealph ratebeta strength mm package growcurves growcurves r dpgrow r res call dpre y x z subjects niter nburn nthin shapealph ratebeta package growcurves growcurves r dpgrow r res call lgm y x z subjects niter nburn nthin package growcurves simpop r calibvars r res call simpop binary representation levels length levels x values as integer x package simpop simpop r ipu r w call simpop ipu work inp con w eps eps verbose ifelse verbose package simpop dynamicgp r gp sep r c deletegpseps r package dynamicgp dynamicgp r gp sep r c getmgpsep r gpsepi as integer gpsepi m integer package dynamicgp m dynamicgp r gp sep r c getdgpsep r gpsepi as integer gpsepi d double m package dynamicgp d dynamicgp r gp sep r c getggpsep r gpsepi as integer gpsepi g double package dynamicgp g dynamicgp r gpseplm r c getmgpsep r gpsepi as integer gpsepi m integer package dynamicgp m dynamicgp r gpseplm r c getmgpseplm r gplmi as integer gplmi m integer package dynamicgp m hellcor r hellcor r res c hellcorc as double cbind x y as integer length x statistic as double as integer pvalue pvalue as double as integer klmax as double alpha package hellcor hellcor r hellcor r pval res statistic sam r samhl r out c grpsvm z as double z lambda as double lambda nnlambda as integer nlambda as double nn as integer n dd as integer d pp as integer p as double xx as double matrix m nlambda mmu as double mu mmax ite as integer max ite tthol as double thol aalpha as double df as double rep nlambda func norm as double matrix d nlambda package sam sam r samql r out c grplasso y as double y x as double z lambda as double lambda nnlambda as integer nlambda nn as integer n dd as integer d pp as integer p ww as double matrix m nlambda mmax ite as integer max ite tthol as double thol regfunc as character regfunc iinput as integer lambda input df as integer rep nlambda sse as double rep nlambda func norm as double matrix d nlambda package sam flare r sugm generator r out c sfgen as integer dd as integer d g as integer theta seed as integer seed package flare genmoss r findlogmarglik r c findlogmarglik as integer tdata as integer dim tdata as integer length model as integer dimens as double alpha logmarglik as double logmarglik package c code logmarglik genmossplus r findlogmarglik r c findlogmarglik as integer tdata as integer dim tdata as integer length model as integer dimens as double alpha logmarglik as double logmarglik package c code logmarglik longclust r t chol new r res c tchol entry as double x as integer class as integer n as integer p as integer gmin as integer gmax as integer df flag as integer gaussian flag as integer models selected as integer use kmeans as double kmeansz gbest as integer gbest icl as integer zbest double n gmax zbest icl double n gmax nubest double gmax nubest icl double gmax mubest double gmax p mubest icl double gmax p tbest double gmax p p tbest icl double gmax p p dbest double gmax p p dbest icl double gmax p p llres double gmax bicres double gmax iclres double gmax package longclust oce r echosounder r c biosonics free storage package oce sparc r sparc r out sglm c sglm a as double rep n d x as double x y as double y nn as integer n nn as integer n dd as integer d lambda as double rep d package sparc sparc r sparc r out slrm c slrm a as double a lambda as double lambda nnlambda as integer nlambda as double nn as integer n dd as integer d ww as double rep d nlambda mmax ite as integer max ite tthol as double thol aalpha as double alpha package sparc this will show up as the wrong number of arguments in routine registration so you will need to revise that please correct asap and before july to safely retain the package on cran
1
387,270
11,459,172,550
IssuesEvent
2020-02-07 06:22:41
MolSnoo/Alter-Ego
https://api.github.com/repos/MolSnoo/Alter-Ego
closed
Make equip command
feature high priority
Equip command should allow a player (or moderator) to equip an item from either of their hands to the specified equipment slot.
1.0
Make equip command - Equip command should allow a player (or moderator) to equip an item from either of their hands to the specified equipment slot.
priority
make equip command equip command should allow a player or moderator to equip an item from either of their hands to the specified equipment slot
1
606,327
18,759,862,599
IssuesEvent
2021-11-05 15:15:55
margaritahumanitarian/helpafamily
https://api.github.com/repos/margaritahumanitarian/helpafamily
closed
🌐 ✨ We ran Lighthouse and we have some performance issues
good first issue high priority easy idea hacktoberfest css
**What should we change and why?** ### Lighthouse [Our Lighthouse score](https://lighthouse-dot-webdotdevsite.appspot.com//lh/html?url=https%3A%2F%2Fhelpafamily.margaritahumanitarian.org%2F) isssss, okay. We can use some performance and accessibility adjustments. It's import for web vitals that we implement these changes. Help us out! ### Some examples <img width="751" alt="Screen Shot 2021-10-26 at 12 50 29 PM" src="https://user-images.githubusercontent.com/68655112/138950873-29bd97ac-4bb2-442a-9f78-aed2c555397d.png"> <img width="914" alt="Screen Shot 2021-10-26 at 12 50 07 PM" src="https://user-images.githubusercontent.com/68655112/138950877-63950485-484f-4fb7-a315-313a8934cd40.png"> <img width="920" alt="Screen Shot 2021-10-26 at 12 50 00 PM" src="https://user-images.githubusercontent.com/68655112/138950881-647bcb42-22da-46a5-890a-8f19cbcf8dd2.png">
1.0
🌐 ✨ We ran Lighthouse and we have some performance issues - **What should we change and why?** ### Lighthouse [Our Lighthouse score](https://lighthouse-dot-webdotdevsite.appspot.com//lh/html?url=https%3A%2F%2Fhelpafamily.margaritahumanitarian.org%2F) isssss, okay. We can use some performance and accessibility adjustments. It's import for web vitals that we implement these changes. Help us out! ### Some examples <img width="751" alt="Screen Shot 2021-10-26 at 12 50 29 PM" src="https://user-images.githubusercontent.com/68655112/138950873-29bd97ac-4bb2-442a-9f78-aed2c555397d.png"> <img width="914" alt="Screen Shot 2021-10-26 at 12 50 07 PM" src="https://user-images.githubusercontent.com/68655112/138950877-63950485-484f-4fb7-a315-313a8934cd40.png"> <img width="920" alt="Screen Shot 2021-10-26 at 12 50 00 PM" src="https://user-images.githubusercontent.com/68655112/138950881-647bcb42-22da-46a5-890a-8f19cbcf8dd2.png">
priority
🌐 ✨ we ran lighthouse and we have some performance issues what should we change and why lighthouse isssss okay we can use some performance and accessibility adjustments it s import for web vitals that we implement these changes help us out some examples img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src
1
274,424
8,561,036,227
IssuesEvent
2018-11-09 04:31:04
QuantEcon/lecture-source-jl
https://api.github.com/repos/QuantEcon/lecture-source-jl
closed
Interpolations pass
high-priority
Change everything to use Interpolations.jl and remove references to the QuantEcon
1.0
Interpolations pass - Change everything to use Interpolations.jl and remove references to the QuantEcon
priority
interpolations pass change everything to use interpolations jl and remove references to the quantecon
1
592,482
17,908,929,990
IssuesEvent
2021-09-09 00:37:36
ClinGen/gene-and-variant-curation-tools
https://api.github.com/repos/ClinGen/gene-and-variant-curation-tools
closed
API for GCEPs to access their data
backend feature GCI priority: high milestone
Each GCEP affiliation will have their own unique key that will provide them access to their own GCEP’s data within the GCI. They can choose time frames and which data to export (currently in JSON format only). This API output is intended for GCEPs to get full access to all their own data within the GCI for their own external analysis of their own data. It is a way for GCEPs to have full access to their own data. First use case for specification was working with May Flowers at UNC.
1.0
API for GCEPs to access their data - Each GCEP affiliation will have their own unique key that will provide them access to their own GCEP’s data within the GCI. They can choose time frames and which data to export (currently in JSON format only). This API output is intended for GCEPs to get full access to all their own data within the GCI for their own external analysis of their own data. It is a way for GCEPs to have full access to their own data. First use case for specification was working with May Flowers at UNC.
priority
api for gceps to access their data each gcep affiliation will have their own unique key that will provide them access to their own gcep’s data within the gci they can choose time frames and which data to export currently in json format only this api output is intended for gceps to get full access to all their own data within the gci for their own external analysis of their own data it is a way for gceps to have full access to their own data first use case for specification was working with may flowers at unc
1
277,799
8,632,811,949
IssuesEvent
2018-11-22 11:57:05
canonical-websites/snapcraft.io
https://api.github.com/repos/canonical-websites/snapcraft.io
closed
Revisions available table needs a reload to reflect change just made
Priority: High
### Expected behaviour Revisions available table is immediately updated after making changes. ### Steps to reproduce the problem * Navigate to the releases and revision history page e.g. https://snapcraft.io/snap-store-proxy/release. * Promote a revision (e.g. 20) from edge to stable, closing candidate. * Note that the table doesn't show that the given revision in stable. * Reload the page. * Now it does. ### Specs - _URL:_ https://snapcraft.io/snap-store-proxy/release - _Operating system:_ Ubuntu 18.04 LTS - _Browser:_ Firefox 63.0.1
1.0
Revisions available table needs a reload to reflect change just made - ### Expected behaviour Revisions available table is immediately updated after making changes. ### Steps to reproduce the problem * Navigate to the releases and revision history page e.g. https://snapcraft.io/snap-store-proxy/release. * Promote a revision (e.g. 20) from edge to stable, closing candidate. * Note that the table doesn't show that the given revision in stable. * Reload the page. * Now it does. ### Specs - _URL:_ https://snapcraft.io/snap-store-proxy/release - _Operating system:_ Ubuntu 18.04 LTS - _Browser:_ Firefox 63.0.1
priority
revisions available table needs a reload to reflect change just made expected behaviour revisions available table is immediately updated after making changes steps to reproduce the problem navigate to the releases and revision history page e g promote a revision e g from edge to stable closing candidate note that the table doesn t show that the given revision in stable reload the page now it does specs url operating system ubuntu lts browser firefox
1
295,633
9,099,361,777
IssuesEvent
2019-02-20 04:05:22
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
closed
thread race in lazy loader _simple_lazy_clause + baked
bug high priority orm
the baked query gets a hold of _simple_lazy_clause, if another thread re-runs it, the new state is incompatible. POC below. ```python import random import threading import time from sqlalchemy import Column from sqlalchemy import create_engine from sqlalchemy import ForeignKey from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy.orm import Session Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) data = Column(String(10)) bs = relationship("B", ) #bake_queries=False) # turn off baked to fix class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(ForeignKey("a.id")) data = Column(String(10)) e = create_engine("mysql://scott:tiger@localhost/test", echo="debug") Base.metadata.create_all(e) s = Session(e) s.add(A(bs=[B()])) s.commit() def worker(): s = Session(e) a1 = s.query(A).first() while True: time.sleep(.001) bs = a1.bs assert len(bs) == 1 s.expire(a1, ['bs']) def chaos(): while True: time.sleep(random.random()) try: del A.bs.property._lazy_strategy._simple_lazy_clause except AttributeError: pass threads = [threading.Thread(target=chaos)] # or remove chaos to fix threads.extend(threading.Thread(target=worker) for i in range(20)) for t in threads: t.start() for t in threads: t.join() ```
1.0
thread race in lazy loader _simple_lazy_clause + baked - the baked query gets a hold of _simple_lazy_clause, if another thread re-runs it, the new state is incompatible. POC below. ```python import random import threading import time from sqlalchemy import Column from sqlalchemy import create_engine from sqlalchemy import ForeignKey from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy.orm import Session Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) data = Column(String(10)) bs = relationship("B", ) #bake_queries=False) # turn off baked to fix class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(ForeignKey("a.id")) data = Column(String(10)) e = create_engine("mysql://scott:tiger@localhost/test", echo="debug") Base.metadata.create_all(e) s = Session(e) s.add(A(bs=[B()])) s.commit() def worker(): s = Session(e) a1 = s.query(A).first() while True: time.sleep(.001) bs = a1.bs assert len(bs) == 1 s.expire(a1, ['bs']) def chaos(): while True: time.sleep(random.random()) try: del A.bs.property._lazy_strategy._simple_lazy_clause except AttributeError: pass threads = [threading.Thread(target=chaos)] # or remove chaos to fix threads.extend(threading.Thread(target=worker) for i in range(20)) for t in threads: t.start() for t in threads: t.join() ```
priority
thread race in lazy loader simple lazy clause baked the baked query gets a hold of simple lazy clause if another thread re runs it the new state is incompatible poc below python import random import threading import time from sqlalchemy import column from sqlalchemy import create engine from sqlalchemy import foreignkey from sqlalchemy import integer from sqlalchemy import string from sqlalchemy ext declarative import declarative base from sqlalchemy orm import relationship from sqlalchemy orm import session base declarative base class a base tablename a id column integer primary key true data column string bs relationship b bake queries false turn off baked to fix class b base tablename b id column integer primary key true a id column foreignkey a id data column string e create engine mysql scott tiger localhost test echo debug base metadata create all e s session e s add a bs s commit def worker s session e s query a first while true time sleep bs bs assert len bs s expire def chaos while true time sleep random random try del a bs property lazy strategy simple lazy clause except attributeerror pass threads or remove chaos to fix threads extend threading thread target worker for i in range for t in threads t start for t in threads t join
1
479,890
13,807,044,288
IssuesEvent
2020-10-11 20:15:40
vitreo12/AlgaSC
https://api.github.com/repos/vitreo12/AlgaSC
closed
Clock sync for interpolation execution
high priority
This is an easy fix, just sync the execution to a clock
1.0
Clock sync for interpolation execution - This is an easy fix, just sync the execution to a clock
priority
clock sync for interpolation execution this is an easy fix just sync the execution to a clock
1
739,569
25,602,820,490
IssuesEvent
2022-12-01 21:55:22
MmgTools/mmg
https://api.github.com/repos/MmgTools/mmg
opened
Allow to disable the check of memory consumption
kind: enhancement priority: high
Maximal available memory is evaluated and stored in `memMax` field of the mesh. Then, at each allocation/unallocation, Mmg computes the used memory and check that it doesn't overflow the maximal authorized one. It is useful to keep local computer running smoothly or to ensure that Mmg will save a mesh if it fails due to lack of memory. But it can also be annoying: - the system optimizes memory allocation, thus, Mmg checks may detect a lack of memory when in fact it is possible to end the remeshing process - inside ParMmg it has no interests: - it is almost never possible to save the mesh (because groups merging step needs available memory) - it forces to compute and distribute remaining memory between the each Mmg process and ParMmg which is tricky, source of errors (erroneous detection of lack of memory) and quite wordy Allowing disabling of memory checks in Mmg can be useful.
1.0
Allow to disable the check of memory consumption - Maximal available memory is evaluated and stored in `memMax` field of the mesh. Then, at each allocation/unallocation, Mmg computes the used memory and check that it doesn't overflow the maximal authorized one. It is useful to keep local computer running smoothly or to ensure that Mmg will save a mesh if it fails due to lack of memory. But it can also be annoying: - the system optimizes memory allocation, thus, Mmg checks may detect a lack of memory when in fact it is possible to end the remeshing process - inside ParMmg it has no interests: - it is almost never possible to save the mesh (because groups merging step needs available memory) - it forces to compute and distribute remaining memory between the each Mmg process and ParMmg which is tricky, source of errors (erroneous detection of lack of memory) and quite wordy Allowing disabling of memory checks in Mmg can be useful.
priority
allow to disable the check of memory consumption maximal available memory is evaluated and stored in memmax field of the mesh then at each allocation unallocation mmg computes the used memory and check that it doesn t overflow the maximal authorized one it is useful to keep local computer running smoothly or to ensure that mmg will save a mesh if it fails due to lack of memory but it can also be annoying the system optimizes memory allocation thus mmg checks may detect a lack of memory when in fact it is possible to end the remeshing process inside parmmg it has no interests it is almost never possible to save the mesh because groups merging step needs available memory it forces to compute and distribute remaining memory between the each mmg process and parmmg which is tricky source of errors erroneous detection of lack of memory and quite wordy allowing disabling of memory checks in mmg can be useful
1
5,060
2,570,948,427
IssuesEvent
2015-02-10 13:38:07
chocolatey/choco
https://api.github.com/repos/chocolatey/choco
closed
CS0200 - Serialization and internal/private sets
3 - Done Bug Priority_HIGH
This seems to affect .NET 4.0, where an in place upgrade of .NET 4.5 works around it. Reference: http://stackoverflow.com/q/23809092/18475 ``` Error serializing type chocolatey.infrastructure.app.domain.Registry: System.InvalidOperationException: Unable to generate a temporary class (result=1 ). error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only at System.Xml.Serialization.Compiler.Compile(Assembly parent, String ns, XmlS erializerCompilerParameters xmlParameters, Evidence evidence) at System.Xml.Serialization.TempAssembly.GenerateAssembly(XmlMapping[] xmlMap pings, Type[] types, String defaultNamespace, Evidence evidence, XmlSerializerCo mpilerParameters parameters, Assembly assembly, Hashtable assemblies) at System.Xml.Serialization.XmlSerializer.GenerateTempAssembly(XmlMapping xml Mapping, Type type, String defaultNamespace) at System.Xml.Serialization.XmlSerializer..ctor(Type type, String defaultName space) at chocolatey.infrastructure.services.XmlService.serialize[XmlType](XmlType x mlType, String xmlFilePath) Unable to generate a temporary class (result=1). error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only ``` It does appear that it may have something to do with permissions as well from the OP.
1.0
CS0200 - Serialization and internal/private sets - This seems to affect .NET 4.0, where an in place upgrade of .NET 4.5 works around it. Reference: http://stackoverflow.com/q/23809092/18475 ``` Error serializing type chocolatey.infrastructure.app.domain.Registry: System.InvalidOperationException: Unable to generate a temporary class (result=1 ). error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only at System.Xml.Serialization.Compiler.Compile(Assembly parent, String ns, XmlS erializerCompilerParameters xmlParameters, Evidence evidence) at System.Xml.Serialization.TempAssembly.GenerateAssembly(XmlMapping[] xmlMap pings, Type[] types, String defaultNamespace, Evidence evidence, XmlSerializerCo mpilerParameters parameters, Assembly assembly, Hashtable assemblies) at System.Xml.Serialization.XmlSerializer.GenerateTempAssembly(XmlMapping xml Mapping, Type type, String defaultNamespace) at System.Xml.Serialization.XmlSerializer..ctor(Type type, String defaultName space) at chocolatey.infrastructure.services.XmlService.serialize[XmlType](XmlType x mlType, String xmlFilePath) Unable to generate a temporary class (result=1). error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only error CS0200: Property or indexer 'chocolatey.infrastructure.app.domain.Registry .RegistryKeys' cannot be assigned to -- it is read only ``` It does appear that it may have something to do with permissions as well from the OP.
priority
serialization and internal private sets this seems to affect net where an in place upgrade of net works around it reference error serializing type chocolatey infrastructure app domain registry system invalidoperationexception unable to generate a temporary class result error property or indexer chocolatey infrastructure app domain registry registrykeys cannot be assigned to it is read only error property or indexer chocolatey infrastructure app domain registry registrykeys cannot be assigned to it is read only at system xml serialization compiler compile assembly parent string ns xmls erializercompilerparameters xmlparameters evidence evidence at system xml serialization tempassembly generateassembly xmlmapping xmlmap pings type types string defaultnamespace evidence evidence xmlserializerco mpilerparameters parameters assembly assembly hashtable assemblies at system xml serialization xmlserializer generatetempassembly xmlmapping xml mapping type type string defaultnamespace at system xml serialization xmlserializer ctor type type string defaultname space at chocolatey infrastructure services xmlservice serialize xmltype x mltype string xmlfilepath unable to generate a temporary class result error property or indexer chocolatey infrastructure app domain registry registrykeys cannot be assigned to it is read only error property or indexer chocolatey infrastructure app domain registry registrykeys cannot be assigned to it is read only it does appear that it may have something to do with permissions as well from the op
1
25,318
2,679,200,753
IssuesEvent
2015-03-26 15:32:58
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
Downloads are not visible
Bug: Production Community Priority - High
The downloads for projects are all 0. Here's a screen shot from the APS 2014 presentation service that has multiple downloads associated with them. ![screen shot 2015-03-26 at 10 33 43 am](https://cloud.githubusercontent.com/assets/5816601/6848453/b609b442-d3a3-11e4-8cd4-eb5397af14bd.png)
1.0
Downloads are not visible - The downloads for projects are all 0. Here's a screen shot from the APS 2014 presentation service that has multiple downloads associated with them. ![screen shot 2015-03-26 at 10 33 43 am](https://cloud.githubusercontent.com/assets/5816601/6848453/b609b442-d3a3-11e4-8cd4-eb5397af14bd.png)
priority
downloads are not visible the downloads for projects are all here s a screen shot from the aps presentation service that has multiple downloads associated with them
1
648,976
21,214,744,980
IssuesEvent
2022-04-11 05:54:05
darealstyl/TeamJ-soen341project2022
https://api.github.com/repos/darealstyl/TeamJ-soen341project2022
closed
User should be able to submit a review by clicking the submit button.
Subtask High-priority
**High Priority**: The addition of a review by the user is an important feature. The submit button has to be successfully implemented to have access to the review.
1.0
User should be able to submit a review by clicking the submit button. - **High Priority**: The addition of a review by the user is an important feature. The submit button has to be successfully implemented to have access to the review.
priority
user should be able to submit a review by clicking the submit button high priority the addition of a review by the user is an important feature the submit button has to be successfully implemented to have access to the review
1
214,230
7,268,023,786
IssuesEvent
2018-02-20 08:33:38
kowala-tech/kUSD
https://api.github.com/repos/kowala-tech/kUSD
closed
Measure comparative work between proposer and validators
High priority blocked
We need to know how much 'work' the proposer does in comparison to a validator. Ideally the proposer would do more 'work' than all validators combined in any scenario. How do we define and measure 'work'? How much 'work' do validators and proposers do, on average?
1.0
Measure comparative work between proposer and validators - We need to know how much 'work' the proposer does in comparison to a validator. Ideally the proposer would do more 'work' than all validators combined in any scenario. How do we define and measure 'work'? How much 'work' do validators and proposers do, on average?
priority
measure comparative work between proposer and validators we need to know how much work the proposer does in comparison to a validator ideally the proposer would do more work than all validators combined in any scenario how do we define and measure work how much work do validators and proposers do on average
1
655,734
21,706,973,966
IssuesEvent
2022-05-10 10:29:42
canonical-web-and-design/multipass.run
https://api.github.com/repos/canonical-web-and-design/multipass.run
closed
important keyword missing in cheatsheet, as offered on multipass site
Priority: High
In the cheatsheet [offered on the site](https://assets.ubuntu.com/v1/f401c3f4-Ubuntu_Server_CLI_pro_tips_2020-04.pdf) as "_Ubuntu Server CLI pro tips_", the two-page PDF has a section on the first page showing a group of 3 commands for "_Which package provides this file?_". The last of the 3 shows `apt-file <filename or command>`, but that just causes display of the command's help (without pointing out what mistake was made). Reading the help, we can see that the issue is that the command as offered should have used a "find" keyword, as in: `apt-file find <filename or command>` Perhaps there was a time in the past when that was a default, but since I experienced it in finding the cheatsheet today (on the front page at multipass.run) and in following the steps, I wanted to pass it along for you to consider fixing, for the sake of future readers. Thanks for the cheatsheet, otherwise.
1.0
important keyword missing in cheatsheet, as offered on multipass site - In the cheatsheet [offered on the site](https://assets.ubuntu.com/v1/f401c3f4-Ubuntu_Server_CLI_pro_tips_2020-04.pdf) as "_Ubuntu Server CLI pro tips_", the two-page PDF has a section on the first page showing a group of 3 commands for "_Which package provides this file?_". The last of the 3 shows `apt-file <filename or command>`, but that just causes display of the command's help (without pointing out what mistake was made). Reading the help, we can see that the issue is that the command as offered should have used a "find" keyword, as in: `apt-file find <filename or command>` Perhaps there was a time in the past when that was a default, but since I experienced it in finding the cheatsheet today (on the front page at multipass.run) and in following the steps, I wanted to pass it along for you to consider fixing, for the sake of future readers. Thanks for the cheatsheet, otherwise.
priority
important keyword missing in cheatsheet as offered on multipass site in the cheatsheet as ubuntu server cli pro tips the two page pdf has a section on the first page showing a group of commands for which package provides this file the last of the shows apt file but that just causes display of the command s help without pointing out what mistake was made reading the help we can see that the issue is that the command as offered should have used a find keyword as in apt file find perhaps there was a time in the past when that was a default but since i experienced it in finding the cheatsheet today on the front page at multipass run and in following the steps i wanted to pass it along for you to consider fixing for the sake of future readers thanks for the cheatsheet otherwise
1
99,326
4,052,881,776
IssuesEvent
2016-05-24 05:57:39
damlaren/ogle
https://api.github.com/repos/damlaren/ogle
closed
what if make_unique returns null?
priority:very high
Want to know when this happens. One of the cases where it's desirable to stop the world.
1.0
what if make_unique returns null? - Want to know when this happens. One of the cases where it's desirable to stop the world.
priority
what if make unique returns null want to know when this happens one of the cases where it s desirable to stop the world
1
223,038
7,445,877,245
IssuesEvent
2018-03-28 07:06:30
bedita/bedita
https://api.github.com/repos/bedita/bedita
closed
Password and login policy
Priority - High Topic - Authentication Topic - Security Type - Task
Implement basic password and login policy. In BE3 we had this configuration defaults: ```php $config['loginPolicy'] = array ( "maxLoginAttempts" => 10, "maxNumDaysInactivity" => 180, "maxNumDaysValidity" => 60, "passwordRule" => "/\w{3,}/", // regexp to match for valid passwords (empty => no regexp) "passwordErrorMessage" => "Password must contain at least 3 valid alphanumeric characters", // error message for passwrds not matching given regexp ); ``` Similar settings could go in an `Auth` configuration key Those settings should be available to client applications in order to perform some client validation.
1.0
Password and login policy - Implement basic password and login policy. In BE3 we had this configuration defaults: ```php $config['loginPolicy'] = array ( "maxLoginAttempts" => 10, "maxNumDaysInactivity" => 180, "maxNumDaysValidity" => 60, "passwordRule" => "/\w{3,}/", // regexp to match for valid passwords (empty => no regexp) "passwordErrorMessage" => "Password must contain at least 3 valid alphanumeric characters", // error message for passwrds not matching given regexp ); ``` Similar settings could go in an `Auth` configuration key Those settings should be available to client applications in order to perform some client validation.
priority
password and login policy implement basic password and login policy in we had this configuration defaults php config array maxloginattempts maxnumdaysinactivity maxnumdaysvalidity passwordrule w regexp to match for valid passwords empty no regexp passworderrormessage password must contain at least valid alphanumeric characters error message for passwrds not matching given regexp similar settings could go in an auth configuration key those settings should be available to client applications in order to perform some client validation
1
241,775
7,834,335,101
IssuesEvent
2018-06-16 12:30:11
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
opened
3.0 Reports: New stats class
Enhancement Priority: High Reports
We need to introduce a new stats class that uses Carbon to calculate the dates and caches results. The stats class will be used to calculate the data for the reports to avoid duplication of code. The new class will also be used to refactor the calculations for the dashboard widget. Adding the calculations in a single class allows us to keep all the calculations together whilst reusing code rather than duplicating it. Once the class is complete, it will make the reports implementation significantly easier. ### To-Do - [ ] Order earnings for given time period - [ ] Number of orders for given time period - [ ] Earnings for specific order items - [ ] Total amount saved for a specific discount code - [ ] Number of refunds for a given time period - [ ] Average customer value - [ ] Average refund amount - [ ] Average discount amount - [ ] Sales/refunds/earnings per gateway - [ ] Average value per gateway - [ ] Total tax collected
1.0
3.0 Reports: New stats class - We need to introduce a new stats class that uses Carbon to calculate the dates and caches results. The stats class will be used to calculate the data for the reports to avoid duplication of code. The new class will also be used to refactor the calculations for the dashboard widget. Adding the calculations in a single class allows us to keep all the calculations together whilst reusing code rather than duplicating it. Once the class is complete, it will make the reports implementation significantly easier. ### To-Do - [ ] Order earnings for given time period - [ ] Number of orders for given time period - [ ] Earnings for specific order items - [ ] Total amount saved for a specific discount code - [ ] Number of refunds for a given time period - [ ] Average customer value - [ ] Average refund amount - [ ] Average discount amount - [ ] Sales/refunds/earnings per gateway - [ ] Average value per gateway - [ ] Total tax collected
priority
reports new stats class we need to introduce a new stats class that uses carbon to calculate the dates and caches results the stats class will be used to calculate the data for the reports to avoid duplication of code the new class will also be used to refactor the calculations for the dashboard widget adding the calculations in a single class allows us to keep all the calculations together whilst reusing code rather than duplicating it once the class is complete it will make the reports implementation significantly easier to do order earnings for given time period number of orders for given time period earnings for specific order items total amount saved for a specific discount code number of refunds for a given time period average customer value average refund amount average discount amount sales refunds earnings per gateway average value per gateway total tax collected
1
264,829
8,320,014,733
IssuesEvent
2018-09-25 18:53:53
MARKETProtocol/dApp
https://api.github.com/repos/MARKETProtocol/dApp
closed
[BUG] My Open Positions are displaying opposite types
Priority: High Type: Bug
<!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### User Story [comment]: # (As a <user type>, I want to <task> so that <goal>.) ### Why Is this Needed? [comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen) *Summary*: ### Description Currently my long positions are showing up as short positions and vice versa. [comment]: # (Feature or Bug? i.e Type: Bug) *Type*: ### Current Behavior [comment]: # (Describe what actually happened.) ### Expected Behavior [comment]: # (Describe what you expected to happen.) ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) ### Solution [comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.) *Summary*: ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) ### Additional Information [comment]: # (Any other information that would be useful, content, screenshots, etc.) speak w/ @travisdmathis for more information
1.0
[BUG] My Open Positions are displaying opposite types - <!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### User Story [comment]: # (As a <user type>, I want to <task> so that <goal>.) ### Why Is this Needed? [comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen) *Summary*: ### Description Currently my long positions are showing up as short positions and vice versa. [comment]: # (Feature or Bug? i.e Type: Bug) *Type*: ### Current Behavior [comment]: # (Describe what actually happened.) ### Expected Behavior [comment]: # (Describe what you expected to happen.) ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) ### Solution [comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.) *Summary*: ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) ### Additional Information [comment]: # (Any other information that would be useful, content, screenshots, etc.) speak w/ @travisdmathis for more information
priority
my open positions are displaying opposite types hello please use the template below for issue ideas or bugs found within market protocol if it is general support you need reach out to us at provide a general summary of the issue in the title above and use relevant fields below to define the problem before you start work please read our contribution and if there is a bounty involved please also see if you have ongoing work from other bounties with us where funding has not been released please do not pick up a new issue we would like to involve as many contributors as possible and parallelize the work flow as much as possible please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline please also note that in order for work to be accepted all code must be accompanied by test cases as well user story as a i want to so that why is this needed describe the problem and why this task is needed provide description of the current state what you would like to happen and what actually happen summary description currently my long positions are showing up as short positions and vice versa feature or bug i e type bug type current behavior describe what actually happened expected behavior describe what you expected to happen reproduction describe how we can replicate the bug step by step solution provide a summary of the solution and a task list on what needs to be fixed summary definition of done any other information that would be useful bullets are helpful additional information any other information that would be useful content screenshots etc speak w travisdmathis for more information
1
345,995
10,382,830,861
IssuesEvent
2019-09-10 08:22:38
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Limit the zone where draggable dialogs can be dragged or change them
Accepted Priority: High bug
### Description Opening the ABOUT window from the BURGER MENU' (menù options) you can move this window also outside the screen but after this you can't close the window using the X. ### In case of Bug (otherwise remove this paragraph): *Browser Affected* - [x] Internet Explorer - [x] Chrome - [x] Firefox - [ ] Safari *Steps to reproduce* - Open a new map - open the burger menù - open the ABOUT option - move the window outside the screen - try to close using the X *Expected Result* - you can't move the window outside the screen and you can close the window using the X *Current Result* - you can move the window also outside the screen but after this you can't close the window using the X. ![immagine](https://user-images.githubusercontent.com/33256901/41764956-e95daf92-7602-11e8-8b7f-ccb8a48e5d45.png)
1.0
Limit the zone where draggable dialogs can be dragged or change them - ### Description Opening the ABOUT window from the BURGER MENU' (menù options) you can move this window also outside the screen but after this you can't close the window using the X. ### In case of Bug (otherwise remove this paragraph): *Browser Affected* - [x] Internet Explorer - [x] Chrome - [x] Firefox - [ ] Safari *Steps to reproduce* - Open a new map - open the burger menù - open the ABOUT option - move the window outside the screen - try to close using the X *Expected Result* - you can't move the window outside the screen and you can close the window using the X *Current Result* - you can move the window also outside the screen but after this you can't close the window using the X. ![immagine](https://user-images.githubusercontent.com/33256901/41764956-e95daf92-7602-11e8-8b7f-ccb8a48e5d45.png)
priority
limit the zone where draggable dialogs can be dragged or change them description opening the about window from the burger menu menù options you can move this window also outside the screen but after this you can t close the window using the x in case of bug otherwise remove this paragraph browser affected internet explorer chrome firefox safari steps to reproduce open a new map open the burger menù open the about option move the window outside the screen try to close using the x expected result you can t move the window outside the screen and you can close the window using the x current result you can move the window also outside the screen but after this you can t close the window using the x
1
795,579
28,078,321,827
IssuesEvent
2023-03-30 02:54:26
gmlc-dispatches/dispatches
https://api.github.com/repos/gmlc-dispatches/dispatches
closed
March 2023 Release Goals
Priority:High
## Migrated from #132 - [x] Integrate TEAL with DISPATCHES workflow: - Issue #86 - Issue #90 addressed by - PRs #177 (replacing ~PR #150~) - [ ] #192 - PR #142 - [x] (**D**) Any changes connected with IDAES/idaes-pse#1002 - Issue #176
1.0
March 2023 Release Goals - ## Migrated from #132 - [x] Integrate TEAL with DISPATCHES workflow: - Issue #86 - Issue #90 addressed by - PRs #177 (replacing ~PR #150~) - [ ] #192 - PR #142 - [x] (**D**) Any changes connected with IDAES/idaes-pse#1002 - Issue #176
priority
march release goals migrated from integrate teal with dispatches workflow issue issue addressed by prs replacing pr pr d any changes connected with idaes idaes pse issue
1
185,102
6,718,966,143
IssuesEvent
2017-10-15 18:51:37
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
strange runtime behavior on macOS
High Priority OS/Arch specific Stdlib
```nim import posix proc f1() = discard newSeq[int]() proc f2(): seq[Pid] = var ret = newSeq[Pid]() ret.add(1) ret.add(2) echo "in :", ret result = ret echo "out:", f2() ``` output: ``` in :@[1, 2] out:@[1, 4] ``` **Removing the proc `f1` will get correct output.** There are two issues behind: - definition of `f1` will cause `ret` in f2 got wrong type inference; the `ret` in f2 was inferred as type `seq[int]`(inspected from the output C code) ```c ... struct TY_qwqHTkRvwhrRyENtudHQ7g { TGenericSeq Sup; NI data[SEQ_DECL_SIZE]; }; struct TY_UW5n9c6ePSIwvorQes4mRDw { TGenericSeq Sup; pid_t data[SEQ_DECL_SIZE]; }; ... N_NIMCALL(TY_UW5n9c6ePSIwvorQes4mRDw*, f2_886jPBEFp8TW0B9b3ZevLBA)(void) { TY_UW5n9c6ePSIwvorQes4mRDw* result; TY_qwqHTkRvwhrRyENtudHQ7g* ret; // wrong type inference here ``` - Pid in `posix_other.nim` defined as int and `import c` as pid_t `Pid* {.importc: "pid_t", header: "<sys/types.h>".} = int` on macOS `pid_t` and `NI` has different size (4 byte vs 8 byte)
1.0
strange runtime behavior on macOS - ```nim import posix proc f1() = discard newSeq[int]() proc f2(): seq[Pid] = var ret = newSeq[Pid]() ret.add(1) ret.add(2) echo "in :", ret result = ret echo "out:", f2() ``` output: ``` in :@[1, 2] out:@[1, 4] ``` **Removing the proc `f1` will get correct output.** There are two issues behind: - definition of `f1` will cause `ret` in f2 got wrong type inference; the `ret` in f2 was inferred as type `seq[int]`(inspected from the output C code) ```c ... struct TY_qwqHTkRvwhrRyENtudHQ7g { TGenericSeq Sup; NI data[SEQ_DECL_SIZE]; }; struct TY_UW5n9c6ePSIwvorQes4mRDw { TGenericSeq Sup; pid_t data[SEQ_DECL_SIZE]; }; ... N_NIMCALL(TY_UW5n9c6ePSIwvorQes4mRDw*, f2_886jPBEFp8TW0B9b3ZevLBA)(void) { TY_UW5n9c6ePSIwvorQes4mRDw* result; TY_qwqHTkRvwhrRyENtudHQ7g* ret; // wrong type inference here ``` - Pid in `posix_other.nim` defined as int and `import c` as pid_t `Pid* {.importc: "pid_t", header: "<sys/types.h>".} = int` on macOS `pid_t` and `NI` has different size (4 byte vs 8 byte)
priority
strange runtime behavior on macos nim import posix proc discard newseq proc seq var ret newseq ret add ret add echo in ret result ret echo out output in out removing the proc will get correct output there are two issues behind definition of will cause ret in got wrong type inference the ret in was inferred as type seq inspected from the output c code c struct ty tgenericseq sup ni data struct ty tgenericseq sup pid t data n nimcall ty void ty result ty ret wrong type inference here pid in posix other nim defined as int and import c as pid t pid importc pid t header int on macos pid t and ni has different size byte vs byte
1
225,880
7,495,952,097
IssuesEvent
2018-04-08 03:25:09
EvictionLab/eviction-lab-website
https://api.github.com/repos/EvictionLab/eviction-lab-website
closed
Add FAQ question for Maryland
high priority
Maryland filings are significantly higher than everywhere else. Matt says this is because Maryland has a different filing process than everywhere else. Eviction process in most states: 1. Eviction notice is issued by landlord (not tracked) 2. Person does not leave, eviction court date is set (tracked) 3. Eviction judgement is made (tracked) In Maryland, step number 1 in the process **is** tracked and counted as an eviction filing.
1.0
Add FAQ question for Maryland - Maryland filings are significantly higher than everywhere else. Matt says this is because Maryland has a different filing process than everywhere else. Eviction process in most states: 1. Eviction notice is issued by landlord (not tracked) 2. Person does not leave, eviction court date is set (tracked) 3. Eviction judgement is made (tracked) In Maryland, step number 1 in the process **is** tracked and counted as an eviction filing.
priority
add faq question for maryland maryland filings are significantly higher than everywhere else matt says this is because maryland has a different filing process than everywhere else eviction process in most states eviction notice is issued by landlord not tracked person does not leave eviction court date is set tracked eviction judgement is made tracked in maryland step number in the process is tracked and counted as an eviction filing
1
495,144
14,272,650,689
IssuesEvent
2020-11-21 17:57:27
bounswe/bounswe2020group4
https://api.github.com/repos/bounswe/bounswe2020group4
closed
(AND) Log-in/Sign-up
Android Android-Logic Android-UI Coding Effort: High Priority: High Status: Needs Review
* UI implementation of the Login/Signup page for Android app. This page will be reached from the My Account button which is on the homepage. * In this page user will sign up or log in with password and username. * User can sign up using his/her google account. (This issue does not contain this feature, it will be implemented later.) ### Simple Page Layout ![Screen Shot 2020-11-04 at 17 33 07](https://user-images.githubusercontent.com/35741925/98124893-cefc5a00-1ec4-11eb-8496-2b7bb6585774.png) Deadline: 22.11.2020 18:30 (updated)
1.0
(AND) Log-in/Sign-up - * UI implementation of the Login/Signup page for Android app. This page will be reached from the My Account button which is on the homepage. * In this page user will sign up or log in with password and username. * User can sign up using his/her google account. (This issue does not contain this feature, it will be implemented later.) ### Simple Page Layout ![Screen Shot 2020-11-04 at 17 33 07](https://user-images.githubusercontent.com/35741925/98124893-cefc5a00-1ec4-11eb-8496-2b7bb6585774.png) Deadline: 22.11.2020 18:30 (updated)
priority
and log in sign up ui implementation of the login signup page for android app this page will be reached from the my account button which is on the homepage in this page user will sign up or log in with password and username user can sign up using his her google account this issue does not contain this feature it will be implemented later simple page layout deadline updated
1
205,336
7,097,109,363
IssuesEvent
2018-01-14 15:49:15
wso2/carbon-apimgt
https://api.github.com/repos/wso2/carbon-apimgt
closed
Store Publisher OIDC login permission not working
APIM 2.1.0 APIM 2.2.0 Priority/High Severity/Critical Type/Bug
**Description:** Users without create/publish permission can access to the publisher and users without subscription permission can access the store when OIDC SSO is configured. {isUserPermitted is hardcoded to true in both store and publisher jaggery_oidc_acs.jag} file **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** **OS, DB, other environments details and versions:** **Steps to reproduce:** 1. Configure OIDC SSO for API Manager store and publisher 2. Create a user without publihser or store login access 3. Login to publisher/store User is allowed to log in. **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
1.0
Store Publisher OIDC login permission not working - **Description:** Users without create/publish permission can access to the publisher and users without subscription permission can access the store when OIDC SSO is configured. {isUserPermitted is hardcoded to true in both store and publisher jaggery_oidc_acs.jag} file **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** **OS, DB, other environments details and versions:** **Steps to reproduce:** 1. Configure OIDC SSO for API Manager store and publisher 2. Create a user without publihser or store login access 3. Login to publisher/store User is allowed to log in. **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
priority
store publisher oidc login permission not working description users without create publish permission can access to the publisher and users without subscription permission can access the store when oidc sso is configured isuserpermitted is hardcoded to true in both store and publisher jaggery oidc acs jag file suggested labels suggested assignees affected product version os db other environments details and versions steps to reproduce configure oidc sso for api manager store and publisher create a user without publihser or store login access login to publisher store user is allowed to log in related issues
1
596,951
18,151,304,864
IssuesEvent
2021-09-26 10:06:15
teambit/bit
https://api.github.com/repos/teambit/bit
opened
make capsule commands work on bare scope
type/feature priority/high
I want to see capsules created for bare scopes using the capsule list command. I think we should move the capsule commands out of the workspace into the isolator. And use the component host instead of workspace directly.
1.0
make capsule commands work on bare scope - I want to see capsules created for bare scopes using the capsule list command. I think we should move the capsule commands out of the workspace into the isolator. And use the component host instead of workspace directly.
priority
make capsule commands work on bare scope i want to see capsules created for bare scopes using the capsule list command i think we should move the capsule commands out of the workspace into the isolator and use the component host instead of workspace directly
1
340,057
10,266,158,132
IssuesEvent
2019-08-22 20:41:50
webkom/lego
https://api.github.com/repos/webkom/lego
closed
Add proper scope support for oauth!
backend bug enhancement priority:high
Add proper support for scopes in oauth. Current impl. will give full access, regardless of what the scope is provided.
1.0
Add proper scope support for oauth! - Add proper support for scopes in oauth. Current impl. will give full access, regardless of what the scope is provided.
priority
add proper scope support for oauth add proper support for scopes in oauth current impl will give full access regardless of what the scope is provided
1