Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
42,293
| 10,933,871,006
|
IssuesEvent
|
2019-11-24 06:35:51
|
davidjamesca/ctypesgen
|
https://api.github.com/repos/davidjamesca/ctypesgen
|
closed
|
Preserve #ifdef #else #endif
|
Priority-Medium Type-Defect auto-migrated
|
```
Some codes could be optionally included into the binary through #ifdef like
directives. Currently there are not means to deal with that.
Especially when some data types are #ifdef to different size like 'double' and
'long doubt'. These could cause segfault when used in struct.
If the #ifdef is turned into
if USE_LONG_DOUBLE == 1:
real = ctypes.longdouble
else
real = ctypes.double
Then USE_LONG_DOUBLE can be set in the python code before import the ctypes
wrapped module.
```
Original issue reported on code.google.com by `mozbug...@gmail.com` on 23 May 2011 at 7:29
|
1.0
|
Preserve #ifdef #else #endif - ```
Some codes could be optionally included into the binary through #ifdef like
directives. Currently there are not means to deal with that.
Especially when some data types are #ifdef to different size like 'double' and
'long doubt'. These could cause segfault when used in struct.
If the #ifdef is turned into
if USE_LONG_DOUBLE == 1:
real = ctypes.longdouble
else
real = ctypes.double
Then USE_LONG_DOUBLE can be set in the python code before import the ctypes
wrapped module.
```
Original issue reported on code.google.com by `mozbug...@gmail.com` on 23 May 2011 at 7:29
|
defect
|
preserve ifdef else endif some codes could be optionally included into the binary through ifdef like directives currently there are not means to deal with that especially when some data types are ifdef to different size like double and long doubt these could cause segfault when used in struct if the ifdef is turned into if use long double real ctypes longdouble else real ctypes double then use long double can be set in the python code before import the ctypes wrapped module original issue reported on code google com by mozbug gmail com on may at
| 1
|
21,191
| 3,471,081,486
|
IssuesEvent
|
2015-12-23 13:10:22
|
lpechacek/cpuset
|
https://api.github.com/repos/lpechacek/cpuset
|
closed
|
cset crashes if /proc/<pid>/stat contains spaces in process name
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Call "cset shield --unshield -v" on a plain Ubuntu
What is the expected output? What do you see instead?
Traceback (most recent call last):
File "/usr/bin/cset", line 47, in <module>
main()
File "/usr/lib/pymodules/python2.7/cpuset/main.py", line 228, in main
command.func(parser, options, args)
File "/usr/lib/pymodules/python2.7/cpuset/commands/shield.py", line 291, in func
print_stats()
File "/usr/lib/pymodules/python2.7/cpuset/commands/shield.py", line 315, in print_sys_stats
proc.log_detailed_task_table(cset.unique_set(SYS_SET), ' ', 76)
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 787, in log_detailed_task_table
l.extend(task_detail_table(set.tasks, indent, width))
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 776, in task_detail_table
if width: l.append(istr + task_detail(task, width))
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 730, in task_detail
out2.append(policy[int(stat[statdef['rtpolicy']])])
IndexError: list index out of range
The problem is that some processes have spaces in their process name, e.g.:
# cat /proc/2845/stat
2845 (alsa-sink-HDMI ) S 2634 2833 2833 0 -1 4218944 3621 0 4 0 13 12 0 0 -6
-11 4 0 6386 463695872 1657 18446744073709551615 4194304 4273956
140722199213344 139661618646368 139661677494767 0 0 3674112 19011
18446744073709551615 0 0 -1 7 5 2 0 0 0 6375016 6377584 36630528
140722199217975 140722199218023 140722199218023 140722199220196 0
# cat /proc/26638/stat
26638 (DNS Res~ver #16) S 2863 2781 2781 0 -1 1077960768 536802 3650 138 0
11390 905 1 0 20 0 51 0 771645 1537323008 136670 18446744073709551615
139959562293248 139959562399028 140732783909232 139958028811552 139959557875662
0 0 4096 33572079 18446744073709551615 0 0 -1 7 0 0 0 0 0 139959564499304
139959564501112 139959593127936 140732783916821 140732783916846 140732783916846
140732783919071 0
This is not handled by the "stat = stat.split()" in task_detail() in proc.py.
What version of the product are you using? On what operating system?
cset 1.5.6, Linux 3.13.0-49 on Ubuntu 14.04, x86_64
```
Original issue reported on code.google.com by `boris1...@nurfuerspam.de` on 30 Apr 2015 at 9:57
|
1.0
|
cset crashes if /proc/<pid>/stat contains spaces in process name - ```
What steps will reproduce the problem?
1. Call "cset shield --unshield -v" on a plain Ubuntu
What is the expected output? What do you see instead?
Traceback (most recent call last):
File "/usr/bin/cset", line 47, in <module>
main()
File "/usr/lib/pymodules/python2.7/cpuset/main.py", line 228, in main
command.func(parser, options, args)
File "/usr/lib/pymodules/python2.7/cpuset/commands/shield.py", line 291, in func
print_stats()
File "/usr/lib/pymodules/python2.7/cpuset/commands/shield.py", line 315, in print_sys_stats
proc.log_detailed_task_table(cset.unique_set(SYS_SET), ' ', 76)
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 787, in log_detailed_task_table
l.extend(task_detail_table(set.tasks, indent, width))
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 776, in task_detail_table
if width: l.append(istr + task_detail(task, width))
File "/usr/lib/pymodules/python2.7/cpuset/commands/proc.py", line 730, in task_detail
out2.append(policy[int(stat[statdef['rtpolicy']])])
IndexError: list index out of range
The problem is that some processes have spaces in their process name, e.g.:
# cat /proc/2845/stat
2845 (alsa-sink-HDMI ) S 2634 2833 2833 0 -1 4218944 3621 0 4 0 13 12 0 0 -6
-11 4 0 6386 463695872 1657 18446744073709551615 4194304 4273956
140722199213344 139661618646368 139661677494767 0 0 3674112 19011
18446744073709551615 0 0 -1 7 5 2 0 0 0 6375016 6377584 36630528
140722199217975 140722199218023 140722199218023 140722199220196 0
# cat /proc/26638/stat
26638 (DNS Res~ver #16) S 2863 2781 2781 0 -1 1077960768 536802 3650 138 0
11390 905 1 0 20 0 51 0 771645 1537323008 136670 18446744073709551615
139959562293248 139959562399028 140732783909232 139958028811552 139959557875662
0 0 4096 33572079 18446744073709551615 0 0 -1 7 0 0 0 0 0 139959564499304
139959564501112 139959593127936 140732783916821 140732783916846 140732783916846
140732783919071 0
This is not handled by the "stat = stat.split()" in task_detail() in proc.py.
What version of the product are you using? On what operating system?
cset 1.5.6, Linux 3.13.0-49 on Ubuntu 14.04, x86_64
```
Original issue reported on code.google.com by `boris1...@nurfuerspam.de` on 30 Apr 2015 at 9:57
|
defect
|
cset crashes if proc stat contains spaces in process name what steps will reproduce the problem call cset shield unshield v on a plain ubuntu what is the expected output what do you see instead traceback most recent call last file usr bin cset line in main file usr lib pymodules cpuset main py line in main command func parser options args file usr lib pymodules cpuset commands shield py line in func print stats file usr lib pymodules cpuset commands shield py line in print sys stats proc log detailed task table cset unique set sys set file usr lib pymodules cpuset commands proc py line in log detailed task table l extend task detail table set tasks indent width file usr lib pymodules cpuset commands proc py line in task detail table if width l append istr task detail task width file usr lib pymodules cpuset commands proc py line in task detail append policy indexerror list index out of range the problem is that some processes have spaces in their process name e g cat proc stat alsa sink hdmi s cat proc stat dns res ver s this is not handled by the stat stat split in task detail in proc py what version of the product are you using on what operating system cset linux on ubuntu original issue reported on code google com by nurfuerspam de on apr at
| 1
|
74,175
| 7,388,130,114
|
IssuesEvent
|
2018-03-16 00:42:37
|
mbari-media-management/vars-anno-sync
|
https://api.github.com/repos/mbari-media-management/vars-anno-sync
|
opened
|
Create test plan/integration test
|
test
|
Integration plan:
1. Spin up m3-microservices (master -- using derby database)
2. Add some annotations using a given concept name
3. Send change message to rabbitMQ changing concept name to different one.
4. Verify that changes occurred in annosaurus -- use web api.
|
1.0
|
Create test plan/integration test - Integration plan:
1. Spin up m3-microservices (master -- using derby database)
2. Add some annotations using a given concept name
3. Send change message to rabbitMQ changing concept name to different one.
4. Verify that changes occurred in annosaurus -- use web api.
|
non_defect
|
create test plan integration test integration plan spin up microservices master using derby database add some annotations using a given concept name send change message to rabbitmq changing concept name to different one verify that changes occurred in annosaurus use web api
| 0
|
8,291
| 2,611,486,286
|
IssuesEvent
|
2015-02-27 05:27:30
|
chrsmith/switchlist
|
https://api.github.com/repos/chrsmith/switchlist
|
closed
|
Help link when opening example layout doesn't lead to example docs
|
auto-migrated Priority-Medium Type-Defect
|
```
Run switchlist.
Open an example layout
The layout loads, and a dialog appears explaning you're looking at an example.
Click the question mark in the lower left of the dialog box.
Expected: Help page for that layout appears.
What actually happens: usually an error about being unable to load a page.
```
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 2 Aug 2012 at 6:08
|
1.0
|
Help link when opening example layout doesn't lead to example docs - ```
Run switchlist.
Open an example layout
The layout loads, and a dialog appears explaning you're looking at an example.
Click the question mark in the lower left of the dialog box.
Expected: Help page for that layout appears.
What actually happens: usually an error about being unable to load a page.
```
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 2 Aug 2012 at 6:08
|
defect
|
help link when opening example layout doesn t lead to example docs run switchlist open an example layout the layout loads and a dialog appears explaning you re looking at an example click the question mark in the lower left of the dialog box expected help page for that layout appears what actually happens usually an error about being unable to load a page original issue reported on code google com by rwbowdi gmail com on aug at
| 1
|
444,653
| 12,815,382,462
|
IssuesEvent
|
2020-07-05 02:05:41
|
Reckue/post-ui
|
https://api.github.com/repos/Reckue/post-ui
|
closed
|
Adaptation css styles
|
Priority:Critical Task
|
Hello, need to redesigne pages to dynamically display CSS styles. The page should be able to fit the size of the window.
|
1.0
|
Adaptation css styles - Hello, need to redesigne pages to dynamically display CSS styles. The page should be able to fit the size of the window.
|
non_defect
|
adaptation css styles hello need to redesigne pages to dynamically display css styles the page should be able to fit the size of the window
| 0
|
60,979
| 17,023,572,109
|
IssuesEvent
|
2021-07-03 02:43:08
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Why cannot I report a bug about josm here?
|
Component: admin Priority: major Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 6.51pm, Wednesday, 31st March 2010]**
is there a reason why we dont have josm tracked here?
|
1.0
|
Why cannot I report a bug about josm here? - **[Submitted to the original trac issue database at 6.51pm, Wednesday, 31st March 2010]**
is there a reason why we dont have josm tracked here?
|
defect
|
why cannot i report a bug about josm here is there a reason why we dont have josm tracked here
| 1
|
20,248
| 3,321,133,297
|
IssuesEvent
|
2015-11-09 06:25:27
|
Tetz95/linux-g13-driver
|
https://api.github.com/repos/Tetz95/linux-g13-driver
|
closed
|
Mouse over description text is hard to read
|
auto-migrated Priority-Medium Type-Defect
|
```
This might be due to my colour vision deficiency but the mouseover text is very
hard to read against the palm rest background. A background window for the text
or a changing the colour of the text can solve this.
```
Original issue reported on code.google.com by `azzc...@gmail.com` on 23 Nov 2011 at 8:18
|
1.0
|
Mouse over description text is hard to read - ```
This might be due to my colour vision deficiency but the mouseover text is very
hard to read against the palm rest background. A background window for the text
or a changing the colour of the text can solve this.
```
Original issue reported on code.google.com by `azzc...@gmail.com` on 23 Nov 2011 at 8:18
|
defect
|
mouse over description text is hard to read this might be due to my colour vision deficiency but the mouseover text is very hard to read against the palm rest background a background window for the text or a changing the colour of the text can solve this original issue reported on code google com by azzc gmail com on nov at
| 1
|
81,584
| 7,787,034,235
|
IssuesEvent
|
2018-06-06 20:56:48
|
brave-intl/publishers
|
https://api.github.com/repos/brave-intl/publishers
|
closed
|
Automated tests for TOTP
|
Automated tests P1
|
TOTP affects account access, and testing it would help reassure us that TOTP users won't get locked out of their account due to a small JS bug.
|
1.0
|
Automated tests for TOTP - TOTP affects account access, and testing it would help reassure us that TOTP users won't get locked out of their account due to a small JS bug.
|
non_defect
|
automated tests for totp totp affects account access and testing it would help reassure us that totp users won t get locked out of their account due to a small js bug
| 0
|
41,278
| 10,353,957,918
|
IssuesEvent
|
2019-09-05 12:48:31
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Error in tree.css file
|
defect
|
I am working with PrimeNG tree and got this build error.

it's from css file of tree component which have two css class with incorrect image path.
Current path is _'./images/line.gif'_ and it should be _'../../images/line.gif'_.
I already fixed it on my code and its working... :)
I am using "primeng": "8.0.0".
|
1.0
|
Error in tree.css file - I am working with PrimeNG tree and got this build error.

it's from css file of tree component which have two css class with incorrect image path.
Current path is _'./images/line.gif'_ and it should be _'../../images/line.gif'_.
I already fixed it on my code and its working... :)
I am using "primeng": "8.0.0".
|
defect
|
error in tree css file i am working with primeng tree and got this build error it s from css file of tree component which have two css class with incorrect image path current path is images line gif and it should be images line gif i already fixed it on my code and its working i am using primeng
| 1
|
61,328
| 17,023,667,920
|
IssuesEvent
|
2021-07-03 03:11:55
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Autocomplete is not ignorable
|
Component: potlatch2 Priority: major Resolution: worksforme Type: defect
|
**[Submitted to the original trac issue database at 12.24pm, Friday, 7th January 2011]**
When adding a building, I tried to add amenity=pub but the autocomplete insisted on turning it to amenity=public_building, even when I escaped out, or pasted the text in.
I had to go to Potlatch 1 to complete my edit and change amenity=public_building to amenity=pub. The autocomplete there still suggests public_building but can at least be escaped out.
This was on an area object.
|
1.0
|
Autocomplete is not ignorable - **[Submitted to the original trac issue database at 12.24pm, Friday, 7th January 2011]**
When adding a building, I tried to add amenity=pub but the autocomplete insisted on turning it to amenity=public_building, even when I escaped out, or pasted the text in.
I had to go to Potlatch 1 to complete my edit and change amenity=public_building to amenity=pub. The autocomplete there still suggests public_building but can at least be escaped out.
This was on an area object.
|
defect
|
autocomplete is not ignorable when adding a building i tried to add amenity pub but the autocomplete insisted on turning it to amenity public building even when i escaped out or pasted the text in i had to go to potlatch to complete my edit and change amenity public building to amenity pub the autocomplete there still suggests public building but can at least be escaped out this was on an area object
| 1
|
46,887
| 13,055,997,604
|
IssuesEvent
|
2020-07-30 03:20:41
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
PPC should not attempt to build with OpenCL unless detection succeeds (Trac #2088)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/2088
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:38",
"description": "It appears that PPC unconditionally sets a flag to build with OpenCL:\n\n\tset(BUILD_OPENCL_PPC ON CACHE BOOL \"BUILD_OPENCL_PPC is ON by default\")\n\nand only gives a warning if OpenCL is not, in fact available:\n\n\tmessage(WARNING \"BUILD_OPENCL_PPC is ON, but OpenCL was not found\")\n\nThis causes compile failures on almost all buildbots, and will likely create problems for users as well. Unless there is some way that this helps with something (which I certainly cannot see) `BUILD_OPENCL_PPC ` should be turned off if detection has failed. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1550067278746682",
"component": "combo simulation",
"summary": "PPC should not attempt to build with OpenCL unless detection succeeds",
"priority": "major",
"keywords": "",
"time": "2017-09-06T04:28:12",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
|
1.0
|
PPC should not attempt to build with OpenCL unless detection succeeds (Trac #2088) - Migrated from https://code.icecube.wisc.edu/ticket/2088
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:38",
"description": "It appears that PPC unconditionally sets a flag to build with OpenCL:\n\n\tset(BUILD_OPENCL_PPC ON CACHE BOOL \"BUILD_OPENCL_PPC is ON by default\")\n\nand only gives a warning if OpenCL is not, in fact available:\n\n\tmessage(WARNING \"BUILD_OPENCL_PPC is ON, but OpenCL was not found\")\n\nThis causes compile failures on almost all buildbots, and will likely create problems for users as well. Unless there is some way that this helps with something (which I certainly cannot see) `BUILD_OPENCL_PPC ` should be turned off if detection has failed. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1550067278746682",
"component": "combo simulation",
"summary": "PPC should not attempt to build with OpenCL unless detection succeeds",
"priority": "major",
"keywords": "",
"time": "2017-09-06T04:28:12",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
|
defect
|
ppc should not attempt to build with opencl unless detection succeeds trac migrated from json status closed changetime description it appears that ppc unconditionally sets a flag to build with opencl n n tset build opencl ppc on cache bool build opencl ppc is on by default n nand only gives a warning if opencl is not in fact available n n tmessage warning build opencl ppc is on but opencl was not found n nthis causes compile failures on almost all buildbots and will likely create problems for users as well unless there is some way that this helps with something which i certainly cannot see build opencl ppc should be turned off if detection has failed reporter cweaver cc resolution fixed ts component combo simulation summary ppc should not attempt to build with opencl unless detection succeeds priority major keywords time milestone owner dima type defect
| 1
|
63,181
| 17,409,860,563
|
IssuesEvent
|
2021-08-03 10:52:54
|
AtlasOfLivingAustralia/collectory
|
https://api.github.com/repos/AtlasOfLivingAustralia/collectory
|
closed
|
Collectory: "email this contact" links are broken
|
bug priority-medium status-new type-defect
|
_migrated from:_ https://code.google.com/p/ala/issues/detail?id=257
_date:_ Wed Sep 4 00:20:09 2013
_author:_ nickdos
---
Noticed on collection pages, e.g.
[http://collections.ala.org.au/public/show/co9](http://collections.ala.org.au/public/show/co9)
Clicking on the "email this contact" link simply refreshes the page.
|
1.0
|
Collectory: "email this contact" links are broken - _migrated from:_ https://code.google.com/p/ala/issues/detail?id=257
_date:_ Wed Sep 4 00:20:09 2013
_author:_ nickdos
---
Noticed on collection pages, e.g.
[http://collections.ala.org.au/public/show/co9](http://collections.ala.org.au/public/show/co9)
Clicking on the "email this contact" link simply refreshes the page.
|
defect
|
collectory email this contact links are broken migrated from date wed sep author nickdos noticed on collection pages e g clicking on the email this contact link simply refreshes the page
| 1
|
79,428
| 28,240,354,423
|
IssuesEvent
|
2023-04-06 06:33:51
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
opened
|
Calendar: Change event not fired after change when pattern and mask is applied
|
:lady_beetle: defect :bangbang: needs-triage
|
### Describe the bug
A p:calendar with a pattern and mask doesn't fire a change event when leaving the input after changing it. Related to #9953 as reverting the fix for it fixes this issue
### Reproducer
See this test (add it to Calendar002Test):
```java
@Test
@Order(2)
@DisplayName("Calendar: AJAX on leaving input after changes")
public void testLeaveInputWithChange(Page page) {
// Arrange
Calendar datePicker = page.datePicker;
WebElement outside = page.outside;
Messages messages = page.messages;
// Act
datePicker.getInput().click();
datePicker.getInput().sendKeys(Keys.chord(PrimeSelenium.isSafari() ? Keys.COMMAND : Keys.CONTROL, "a"));
datePicker.getInput().sendKeys("01/01/2023");
PrimeSelenium.guardAjax(outside).click();
// Assert
assertDisplayed(messages);
}
```
### Expected behavior
Change event is fired again when the input is changed
### PrimeFaces edition
None
### PrimeFaces version
13.0.0-SNAPSHOT
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_
|
1.0
|
Calendar: Change event not fired after change when pattern and mask is applied - ### Describe the bug
A p:calendar with a pattern and mask doesn't fire a change event when leaving the input after changing it. Related to #9953 as reverting the fix for it fixes this issue
### Reproducer
See this test (add it to Calendar002Test):
```java
@Test
@Order(2)
@DisplayName("Calendar: AJAX on leaving input after changes")
public void testLeaveInputWithChange(Page page) {
// Arrange
Calendar datePicker = page.datePicker;
WebElement outside = page.outside;
Messages messages = page.messages;
// Act
datePicker.getInput().click();
datePicker.getInput().sendKeys(Keys.chord(PrimeSelenium.isSafari() ? Keys.COMMAND : Keys.CONTROL, "a"));
datePicker.getInput().sendKeys("01/01/2023");
PrimeSelenium.guardAjax(outside).click();
// Assert
assertDisplayed(messages);
}
```
### Expected behavior
Change event is fired again when the input is changed
### PrimeFaces edition
None
### PrimeFaces version
13.0.0-SNAPSHOT
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_
|
defect
|
calendar change event not fired after change when pattern and mask is applied describe the bug a p calendar with a pattern and mask doesn t fire a change event when leaving the input after changing it related to as reverting the fix for it fixes this issue reproducer see this test add it to java test order displayname calendar ajax on leaving input after changes public void testleaveinputwithchange page page arrange calendar datepicker page datepicker webelement outside page outside messages messages page messages act datepicker getinput click datepicker getinput sendkeys keys chord primeselenium issafari keys command keys control a datepicker getinput sendkeys primeselenium guardajax outside click assert assertdisplayed messages expected behavior change event is fired again when the input is changed primefaces edition none primefaces version snapshot theme no response jsf implementation mojarra jsf version java version browser s no response
| 1
|
12,387
| 2,694,264,556
|
IssuesEvent
|
2015-04-01 19:16:28
|
google/google-api-go-client
|
https://api.github.com/repos/google/google-api-go-client
|
closed
|
storage: can't operate on files with spaces in the name
|
fixed priority-medium type-defect
|
**nickcw** on 14 Jul 2014 at 6:06:
```
What steps will reproduce the problem?
1. Create a file with a space in the name, eg "bad file"
2. Run svc.Objects.Delete("bucket", "bad file").Do()
3. It returns `googleapi: Error 404: Not Found, notFound`
What is the expected output? What do you see instead?
Expected output is no error
What version of the product are you using? On what operating system?
Using version 116:0260180239b8
Please provide any additional information below.
The problem appears to be that in (for example)
func (c *ObjectsDeleteCall) Do() error {
In storage/v1/storage-gen.go we find these lines
req.URL.Path = strings.Replace(req.URL.Path, "{object}", url.QueryEscape(c.object), 1)
replacing it with this causes the Delete function to work
req.URL.Path = strings.Replace(req.URL.Path, "{object}", strings.Replace(url.QueryEscape(c.object), "+", "%20", -1), 1)
So it appears that the storage API doesn't understand `+` in paths to mean ` `
but is quite willing to accept `%20`
Other methods are affected including `Get` but I haven't made a comprehensive
survey.
I'll propose a CL in a moment for discussion as to a fix
```
|
1.0
|
storage: can't operate on files with spaces in the name -
**nickcw** on 14 Jul 2014 at 6:06:
```
What steps will reproduce the problem?
1. Create a file with a space in the name, eg "bad file"
2. Run svc.Objects.Delete("bucket", "bad file").Do()
3. It returns `googleapi: Error 404: Not Found, notFound`
What is the expected output? What do you see instead?
Expected output is no error
What version of the product are you using? On what operating system?
Using version 116:0260180239b8
Please provide any additional information below.
The problem appears to be that in (for example)
func (c *ObjectsDeleteCall) Do() error {
In storage/v1/storage-gen.go we find these lines
req.URL.Path = strings.Replace(req.URL.Path, "{object}", url.QueryEscape(c.object), 1)
replacing it with this causes the Delete function to work
req.URL.Path = strings.Replace(req.URL.Path, "{object}", strings.Replace(url.QueryEscape(c.object), "+", "%20", -1), 1)
So it appears that the storage API doesn't understand `+` in paths to mean ` `
but is quite willing to accept `%20`
Other methods are affected including `Get` but I haven't made a comprehensive
survey.
I'll propose a CL in a moment for discussion as to a fix
```
|
defect
|
storage can t operate on files with spaces in the name nickcw on jul at what steps will reproduce the problem create a file with a space in the name eg bad file run svc objects delete bucket bad file do it returns googleapi error not found notfound what is the expected output what do you see instead expected output is no error what version of the product are you using on what operating system using version please provide any additional information below the problem appears to be that in for example func c objectsdeletecall do error in storage storage gen go we find these lines req url path strings replace req url path object url queryescape c object replacing it with this causes the delete function to work req url path strings replace req url path object strings replace url queryescape c object so it appears that the storage api doesn t understand in paths to mean but is quite willing to accept other methods are affected including get but i haven t made a comprehensive survey i ll propose a cl in a moment for discussion as to a fix
| 1
|
47,953
| 13,067,325,624
|
IssuesEvent
|
2020-07-31 00:06:18
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
[dataio] dataio-pyshovel goto frame offset (Trac #1508)
|
Migrated from Trac combo core defect
|
goto frame uses frame index starting at 0, but the display uses frame index starting at 1. this leads to an offset when trying to go to a specific frame number.
Migrated from https://code.icecube.wisc.edu/ticket/1508
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:22",
"description": "goto frame uses frame index starting at 0, but the display uses frame index starting at 1. this leads to an offset when trying to go to a specific frame number.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067082284240",
"component": "combo core",
"summary": "[dataio] dataio-pyshovel goto frame offset",
"priority": "major",
"keywords": "",
"time": "2016-01-13T16:29:48",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
1.0
|
[dataio] dataio-pyshovel goto frame offset (Trac #1508) - goto frame uses frame index starting at 0, but the display uses frame index starting at 1. this leads to an offset when trying to go to a specific frame number.
Migrated from https://code.icecube.wisc.edu/ticket/1508
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:22",
"description": "goto frame uses frame index starting at 0, but the display uses frame index starting at 1. this leads to an offset when trying to go to a specific frame number.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067082284240",
"component": "combo core",
"summary": "[dataio] dataio-pyshovel goto frame offset",
"priority": "major",
"keywords": "",
"time": "2016-01-13T16:29:48",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
defect
|
dataio pyshovel goto frame offset trac goto frame uses frame index starting at but the display uses frame index starting at this leads to an offset when trying to go to a specific frame number migrated from json status closed changetime description goto frame uses frame index starting at but the display uses frame index starting at this leads to an offset when trying to go to a specific frame number reporter david schultz cc resolution fixed ts component combo core summary dataio pyshovel goto frame offset priority major keywords time milestone owner david schultz type defect
| 1
|
69,607
| 22,574,708,774
|
IssuesEvent
|
2022-06-28 06:02:50
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Live Location Share - Tile in reply should link to original message, not open map
|
T-Defect S-Tolerable O-Uncommon A-Location-Sharing
|
### Steps to reproduce
- Share a live location
- Reply to the live location tile
- On the reply, click anywhere on the live location map
- Notice that the maximised map opens
### Outcome
#### What did you expect?
When you click the live location tile someone replied to, it doesn’t open maximised but instead jumps to the original message.
#### What happened instead?
When you click the live location tile someone replied to, it opens the maximised map.
### Operating system
Ubuntu 22.04 LTS
### Browser information
Firefox 102.0 (64-bit)
### URL for webapp
https://develop.element.io/
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Live Location Share - Tile in reply should link to original message, not open map - ### Steps to reproduce
- Share a live location
- Reply to the live location tile
- On the reply, click anywhere on the live location map
- Notice that the maximised map opens
### Outcome
#### What did you expect?
When you click the live location tile someone replied to, it doesn’t open maximised but instead jumps to the original message.
#### What happened instead?
When you click the live location tile someone replied to, it opens the maximised map.
### Operating system
Ubuntu 22.04 LTS
### Browser information
Firefox 102.0 (64-bit)
### URL for webapp
https://develop.element.io/
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
live location share tile in reply should link to original message not open map steps to reproduce share a live location reply to the live location tile on the reply click anywhere on the live location map notice that the maximised map opens outcome what did you expect when you click the live location tile someone replied to it doesn’t open maximised but instead jumps to the original message what happened instead when you click the live location tile someone replied to it opens the maximised map operating system ubuntu lts browser information firefox bit url for webapp application version no response homeserver no response will you send logs no
| 1
|
39,168
| 9,248,818,532
|
IssuesEvent
|
2019-03-15 07:36:35
|
hasse69/rar2fs
|
https://api.github.com/repos/hasse69/rar2fs
|
reopened
|
rar2fs dies with "Got signal 11, faulty address"
|
Defect Priority-Medium
|
I am using rar2fs on centos 7 and after a while the rar2fs mountpoint was dead.
`ls /rar2fs-mountpoint` prints
`ls: cannot access rar2fs: Transport endpoint is not connected` and the rar2fs process is gone. The systemd journal contains:
```
rar2fs[27787]: Got signal 11, faulty address is 0x20000000000, from 0x411c39
rar2fs[27787]: rar2fs() [0x411c39]
rar2fs[27787]: /lib64/libpthread.so.0(+0xf6d0) [0x7fe5c48366d0]
rar2fs[27787]: rar2fs(_Z8wcsncpyzPwPKwm+0x1a) [0x41c32a]
rar2fs[27787]: rar2fs(RARReadHeaderEx+0x2a1) [0x440e51]
rar2fs[27787]: rar2fs() [0x413ca2]
rar2fs[27787]: rar2fs() [0x417d47]
rar2fs[27787]: rar2fs() [0x41a4b0]
rar2fs[27787]: rar2fs() [0x41aaee]
rar2fs[27787]: rar2fs() [0x41aeea]
rar2fs[27787]: /lib64/libfuse.so.2(fuse_fs_readdir+0x92) [0x7fe5c537a922]
rar2fs[27787]: /lib64/libfuse.so.2(+0xeaab) [0x7fe5c537aaab]
rar2fs[27787]: /lib64/libfuse.so.2(+0x15696) [0x7fe5c5381696]
rar2fs[27787]: /lib64/libfuse.so.2(+0x16bcb) [0x7fe5c5382bcb]
rar2fs[27787]: /lib64/libfuse.so.2(+0x13461) [0x7fe5c537f461]
rar2fs[27787]: /lib64/libpthread.so.0(+0x7e25) [0x7fe5c482ee25]
rar2fs[27787]: /lib64/libc.so.6(clone+0x6d) [0x7fe5c4558bad]
```
Using:
* unrarsrc-5.6.4.tar.gz
* rar2fs git 2357751c167ec6e19c55bb87df0ba0064df1522e
Rar2fs was mounted using
`rar2fs --seek-length=1 -o noatime,kernel_cache,allow_other,big_writes,uid=$(id -u johndoe),gid=$(id -g johndoe) /rar-src /rar2fs-mountpoint`
@hasse69 Do you have any clue what's causing this? Does it have a problem with the rar headers (maybe due to `--seek-length=1`) or is it s.th. about `unrarsrc` version? I guess in any case it should also be a bug in rar2fs since it just crashed instead of ignoring the problematic rar file in a more graceful way right? Is there anything I can do to figure out on which file it's crashing without modifying the rar2fs source?
|
1.0
|
rar2fs dies with "Got signal 11, faulty address" - I am using rar2fs on centos 7 and after a while the rar2fs mountpoint was dead.
`ls /rar2fs-mountpoint` prints
`ls: cannot access rar2fs: Transport endpoint is not connected` and the rar2fs process is gone. The systemd journal contains:
```
rar2fs[27787]: Got signal 11, faulty address is 0x20000000000, from 0x411c39
rar2fs[27787]: rar2fs() [0x411c39]
rar2fs[27787]: /lib64/libpthread.so.0(+0xf6d0) [0x7fe5c48366d0]
rar2fs[27787]: rar2fs(_Z8wcsncpyzPwPKwm+0x1a) [0x41c32a]
rar2fs[27787]: rar2fs(RARReadHeaderEx+0x2a1) [0x440e51]
rar2fs[27787]: rar2fs() [0x413ca2]
rar2fs[27787]: rar2fs() [0x417d47]
rar2fs[27787]: rar2fs() [0x41a4b0]
rar2fs[27787]: rar2fs() [0x41aaee]
rar2fs[27787]: rar2fs() [0x41aeea]
rar2fs[27787]: /lib64/libfuse.so.2(fuse_fs_readdir+0x92) [0x7fe5c537a922]
rar2fs[27787]: /lib64/libfuse.so.2(+0xeaab) [0x7fe5c537aaab]
rar2fs[27787]: /lib64/libfuse.so.2(+0x15696) [0x7fe5c5381696]
rar2fs[27787]: /lib64/libfuse.so.2(+0x16bcb) [0x7fe5c5382bcb]
rar2fs[27787]: /lib64/libfuse.so.2(+0x13461) [0x7fe5c537f461]
rar2fs[27787]: /lib64/libpthread.so.0(+0x7e25) [0x7fe5c482ee25]
rar2fs[27787]: /lib64/libc.so.6(clone+0x6d) [0x7fe5c4558bad]
```
Using:
* unrarsrc-5.6.4.tar.gz
* rar2fs git 2357751c167ec6e19c55bb87df0ba0064df1522e
Rar2fs was mounted using
`rar2fs --seek-length=1 -o noatime,kernel_cache,allow_other,big_writes,uid=$(id -u johndoe),gid=$(id -g johndoe) /rar-src /rar2fs-mountpoint`
@hasse69 Do you have any clue what's causing this? Does it have a problem with the rar headers (maybe due to `--seek-length=1`) or is it s.th. about `unrarsrc` version? I guess in any case it should also be a bug in rar2fs since it just crashed instead of ignoring the problematic rar file in a more graceful way right? Is there anything I can do to figure out on which file it's crashing without modifying the rar2fs source?
|
defect
|
dies with got signal faulty address i am using on centos and after a while the mountpoint was dead ls mountpoint prints ls cannot access transport endpoint is not connected and the process is gone the systemd journal contains got signal faulty address is from libpthread so rarreadheaderex libfuse so fuse fs readdir libfuse so libfuse so libfuse so libfuse so libpthread so libc so clone using unrarsrc tar gz git was mounted using seek length o noatime kernel cache allow other big writes uid id u johndoe gid id g johndoe rar src mountpoint do you have any clue what s causing this does it have a problem with the rar headers maybe due to seek length or is it s th about unrarsrc version i guess in any case it should also be a bug in since it just crashed instead of ignoring the problematic rar file in a more graceful way right is there anything i can do to figure out on which file it s crashing without modifying the source
| 1
|
267,854
| 20,248,068,449
|
IssuesEvent
|
2022-02-14 15:25:43
|
Soucouyant/Snipe-Bot
|
https://api.github.com/repos/Soucouyant/Snipe-Bot
|
opened
|
Edit Snipe
|
documentation
|
Implement a feature in which the user can snipe edited messages.
Can be done with a new element in the schema, as well as edit detection. Another collection could work for edited messages, and then recall based on user sent or another metric.
|
1.0
|
Edit Snipe - Implement a feature in which the user can snipe edited messages.
Can be done with a new element in the schema, as well as edit detection. Another collection could work for edited messages, and then recall based on user sent or another metric.
|
non_defect
|
edit snipe implement a feature in which the user can snipe edited messages can be done with a new element in the schema as well as edit detection another collection could work for edited messages and then recall based on user sent or another metric
| 0
|
52,150
| 13,211,393,379
|
IssuesEvent
|
2020-08-15 22:49:15
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
ports install of lhapdf 5.8.7 fails, Ubuntu 16.04 (Trac #1752)
|
Incomplete Migration Migrated from Trac defect tools/ports
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1752">https://code.icecube.wisc.edu/projects/icecube/ticket/1752</a>, reported by jlanfranchiand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-21T15:22:02",
"_ts": "1466522522975275",
"description": "This dependency failing causes Genie 2.8.6 to fail to install as well.",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "fixed",
"time": "2016-06-20T20:57:58",
"component": "tools/ports",
"summary": "ports install of lhapdf 5.8.7 fails, Ubuntu 16.04",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
ports install of lhapdf 5.8.7 fails, Ubuntu 16.04 (Trac #1752) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1752">https://code.icecube.wisc.edu/projects/icecube/ticket/1752</a>, reported by jlanfranchiand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-21T15:22:02",
"_ts": "1466522522975275",
"description": "This dependency failing causes Genie 2.8.6 to fail to install as well.",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "fixed",
"time": "2016-06-20T20:57:58",
"component": "tools/ports",
"summary": "ports install of lhapdf 5.8.7 fails, Ubuntu 16.04",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
ports install of lhapdf fails ubuntu trac migrated from json status closed changetime ts description this dependency failing causes genie to fail to install as well reporter jlanfranchi cc resolution fixed time component tools ports summary ports install of lhapdf fails ubuntu priority normal keywords milestone owner nega type defect
| 1
|
29,104
| 5,535,644,912
|
IssuesEvent
|
2017-03-21 17:47:53
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
closed
|
nvcc / cuda clang issue due to a missing -DHPX_WITH_CUDA flag
|
category: CMake type: defect
|
Got a error due to empty definitions of `HPX_HOST`, `HPX_DEVICE`, and `HPX_HOST_DEVICE` when compiling example cuda codes that are in the build tree. The reason is a missing flag `-DHPX_WITH_CUDA` that is now required since commit 80277b9c3db29f5ff2bdfa7bac2196baf0b33876
|
1.0
|
nvcc / cuda clang issue due to a missing -DHPX_WITH_CUDA flag - Got a error due to empty definitions of `HPX_HOST`, `HPX_DEVICE`, and `HPX_HOST_DEVICE` when compiling example cuda codes that are in the build tree. The reason is a missing flag `-DHPX_WITH_CUDA` that is now required since commit 80277b9c3db29f5ff2bdfa7bac2196baf0b33876
|
defect
|
nvcc cuda clang issue due to a missing dhpx with cuda flag got a error due to empty definitions of hpx host hpx device and hpx host device when compiling example cuda codes that are in the build tree the reason is a missing flag dhpx with cuda that is now required since commit
| 1
|
366,368
| 25,580,428,091
|
IssuesEvent
|
2022-12-01 03:32:41
|
happynclient/happynwindows
|
https://api.github.com/repos/happynclient/happynwindows
|
closed
|
根本无法编译,你们真的是开源吗?还是只是想蹭开源的噱头圈钱?
|
documentation enhancement
|
`已启动生成...
1>------ 已启动生成: 项目: happynet, 配置: Debug Win32 ------
1>cl : 命令行 warning D9035: “Gm”选项已否决,并将在将来的版本中移除
1>utils.cpp
1>service.cpp
1>registry.cpp
1>process.cpp
1>netinterface.cpp
1>tray.cpp
1>G:\happynwindows\happynet\resource.h(1395,1): warning C4005: “_WIN32_WINNT”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(218,1): message : 参见“_WIN32_WINNT”的前一个定义
1>G:\happynwindows\happynet\resource.h(1419,1): warning C4005: “_WIN32_IE”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(263,1): message : 参见“_WIN32_IE”的前一个定义
1>systemsrv.cpp
1>netadapter.cpp
1>G:\happynwindows\happynet\netadapter.h(27,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(27,17): error C3646: “sIp”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(27,20): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(28,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(28,17): error C3646: “sSubnet”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(28,24): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(32,41): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(32,48): error C2065: “string”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2923: "std::vector": "string" 不是参数 "_Ty" 的有效 模板 类型参数
1>G:\happynwindows\happynet\netadapter.h(32,48): message : 参见“string”的声明
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2976: “std::vector'”: 模板 参数太少
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(32,50): error C2955: “std::vector”: 使用 类 模板 需要 模板 参数列表
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(52,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(52,13): error C3646: “GetAdapterName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(52,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(52,35): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(53,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(53,13): error C3646: “GetAdapterDescription”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(53,34): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(53,42): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(62,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(62,17): error C3646: “GetIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(62,26): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(62,45): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(63,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(63,17): error C3646: “GetSubnetForIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(63,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(63,54): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(64,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(64,17): error C3646: “GetDnsAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(64,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(64,47): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(65,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(65,17): error C3646: “GetCurrentIpAddress”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(65,36): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(65,44): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(69,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(69,13): error C3646: “GetDchpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(69,24): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(69,32): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(73,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(73,13): error C3646: “GetPrimaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(73,33): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(73,41): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(74,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(74,13): error C3646: “GetSecondaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(74,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(74,43): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(76,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(76,17): error C3646: “GetGatewayAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(76,31): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(76,74): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(79,12): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(79,20): error C3646: “GetAdapterTypeString”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(79,41): error C2275: “UINT”: 应为表达式而不是类型
1>G:\happynwindows\happynet\netadapter.h(79,46): error C2146: 语法错误: 缺少“)”(在标识符“nType”的前面)
1>G:\happynwindows\happynet\netadapter.h(79,52): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(87,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(87,17): error C3646: “GetStringFromArray”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(87,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(87,78): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(91,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(91,15): error C3646: “m_sName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(91,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(92,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(92,15): error C3646: “m_sDesc”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(92,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(93,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(93,15): error C3646: “m_sPriWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(93,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(94,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(94,15): error C3646: “m_sSecWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(94,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(95,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(95,15): error C3646: “m_sDefGateway”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(95,28): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(96,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(96,15): error C3646: “m_sDhcpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(96,26): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.cpp(91,20): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(93,19): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(112,20): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(114,19): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(141,9): error C2065: “m_sName”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(142,9): error C2065: “m_sDesc”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(143,9): error C2065: “m_sPriWins”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(144,9): error C2065: “m_sSecWins”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(145,9): error C2065: “m_sDhcpAddr”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(156,26): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(157,26): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(160,26): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(161,26): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(169,20): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(170,20): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(180,27): error C2039: "push_back": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(197,36): error C2039: "push_back": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(281,30): error C2039: "GetStringFromArray": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(281,92): error C2270: “GetStringFromArray”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(283,41): error C2039: "size": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(284,30): error C2676: 二进制“[”:“const StringArray”不定义该运算符或到预定义运算符可接收的类型的转换
1>G:\happynwindows\happynet\netadapter.cpp(295,26): error C2039: "GetAdapterTypeString": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(317,26): error C2039: "GetAdapterName": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(317,52): error C2270: “GetAdapterName”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(317,61): error C2065: “m_sName”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(318,26): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(318,57): error C2270: “GetAdapterDescription”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(318,66): error C2065: “m_sDesc”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(322,75): error C2039: "size": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(323,30): error C2039: "GetDnsAddr": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(323,58): error C2270: “GetDnsAddr”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(323,58): fatal error C1003: 错误计数超过 100;正在停止编译
1>happynwinclient.cpp
1>G:\happynwindows\happynet\netadapter.h(27,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(27,17): error C3646: “sIp”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(27,20): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(28,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(28,17): error C3646: “sSubnet”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(28,24): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(32,41): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(32,48): error C2065: “string”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2923: "std::vector": "string" 不是参数 "_Ty" 的有效 模板 类型参数
1>G:\happynwindows\happynet\netadapter.h(32,48): message : 参见“string”的声明
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2976: “std::vector'”: 模板 参数太少
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(32,50): error C2955: “std::vector”: 使用 类 模板 需要 模板 参数列表
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(52,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(52,13): error C3646: “GetAdapterName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(52,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(52,35): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(53,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(53,13): error C3646: “GetAdapterDescription”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(53,34): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(53,42): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(62,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(62,17): error C3646: “GetIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(62,26): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(62,45): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(63,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(63,17): error C3646: “GetSubnetForIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(63,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(63,54): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(64,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(64,17): error C3646: “GetDnsAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(64,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(64,47): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(65,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(65,17): error C3646: “GetCurrentIpAddress”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(65,36): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(65,44): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(69,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(69,13): error C3646: “GetDchpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(69,24): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(69,32): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(73,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(73,13): error C3646: “GetPrimaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(73,33): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(73,41): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(74,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(74,13): error C3646: “GetSecondaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(74,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(74,43): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(76,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(76,17): error C3646: “GetGatewayAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(76,31): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(76,74): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(79,12): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(79,20): error C3646: “GetAdapterTypeString”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(79,41): error C2275: “UINT”: 应为表达式而不是类型
1>G:\happynwindows\happynet\netadapter.h(79,46): error C2146: 语法错误: 缺少“)”(在标识符“nType”的前面)
1>G:\happynwindows\happynet\netadapter.h(79,52): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(87,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(87,17): error C3646: “GetStringFromArray”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(87,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(87,78): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(91,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(91,15): error C3646: “m_sName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(91,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(92,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(92,15): error C3646: “m_sDesc”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(92,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(93,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(93,15): error C3646: “m_sPriWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(93,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(94,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(94,15): error C3646: “m_sSecWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(94,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(95,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(95,15): error C3646: “m_sDefGateway”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(95,28): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(96,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(96,15): error C3646: “m_sDhcpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(96,26): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\resource.h(1395,1): warning C4005: “_WIN32_WINNT”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(218,1): message : 参见“_WIN32_WINNT”的前一个定义
1>G:\happynwindows\happynet\resource.h(1419,1): warning C4005: “_WIN32_IE”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(263,1): message : 参见“_WIN32_IE”的前一个定义
1>G:\happynwindows\happynet\happynwinclient.cpp(785,29): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,70): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,111): error C2039: "GetAdapterName": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): warning C4473: “swprintf_s”: 没有为格式字符串传递足够的参数
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): message : 占位符和其参数预计 2 可变参数,但提供的却是 0 参数
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): message : 缺失的可变参数 1 为格式字符串“%s”所需
1>正在生成代码...
1>已完成生成项目“happynet.vcxproj”的操作 - 失败。
========== 版本: 0 成功,1 失败,0 更新,0 跳过 ==========
========== 占用时间 00:05.944 ==========
`
|
1.0
|
根本无法编译,你们真的是开源吗?还是只是想蹭开源的噱头圈钱? - `已启动生成...
1>------ 已启动生成: 项目: happynet, 配置: Debug Win32 ------
1>cl : 命令行 warning D9035: “Gm”选项已否决,并将在将来的版本中移除
1>utils.cpp
1>service.cpp
1>registry.cpp
1>process.cpp
1>netinterface.cpp
1>tray.cpp
1>G:\happynwindows\happynet\resource.h(1395,1): warning C4005: “_WIN32_WINNT”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(218,1): message : 参见“_WIN32_WINNT”的前一个定义
1>G:\happynwindows\happynet\resource.h(1419,1): warning C4005: “_WIN32_IE”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(263,1): message : 参见“_WIN32_IE”的前一个定义
1>systemsrv.cpp
1>netadapter.cpp
1>G:\happynwindows\happynet\netadapter.h(27,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(27,17): error C3646: “sIp”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(27,20): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(28,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(28,17): error C3646: “sSubnet”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(28,24): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(32,41): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(32,48): error C2065: “string”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2923: "std::vector": "string" 不是参数 "_Ty" 的有效 模板 类型参数
1>G:\happynwindows\happynet\netadapter.h(32,48): message : 参见“string”的声明
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2976: “std::vector'”: 模板 参数太少
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(32,50): error C2955: “std::vector”: 使用 类 模板 需要 模板 参数列表
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(52,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(52,13): error C3646: “GetAdapterName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(52,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(52,35): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(53,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(53,13): error C3646: “GetAdapterDescription”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(53,34): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(53,42): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(62,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(62,17): error C3646: “GetIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(62,26): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(62,45): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(63,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(63,17): error C3646: “GetSubnetForIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(63,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(63,54): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(64,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(64,17): error C3646: “GetDnsAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(64,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(64,47): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(65,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(65,17): error C3646: “GetCurrentIpAddress”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(65,36): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(65,44): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(69,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(69,13): error C3646: “GetDchpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(69,24): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(69,32): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(73,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(73,13): error C3646: “GetPrimaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(73,33): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(73,41): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(74,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(74,13): error C3646: “GetSecondaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(74,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(74,43): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(76,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(76,17): error C3646: “GetGatewayAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(76,31): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(76,74): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(79,12): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(79,20): error C3646: “GetAdapterTypeString”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(79,41): error C2275: “UINT”: 应为表达式而不是类型
1>G:\happynwindows\happynet\netadapter.h(79,46): error C2146: 语法错误: 缺少“)”(在标识符“nType”的前面)
1>G:\happynwindows\happynet\netadapter.h(79,52): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(87,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(87,17): error C3646: “GetStringFromArray”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(87,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(87,78): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(91,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(91,15): error C3646: “m_sName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(91,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(92,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(92,15): error C3646: “m_sDesc”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(92,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(93,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(93,15): error C3646: “m_sPriWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(93,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(94,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(94,15): error C3646: “m_sSecWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(94,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(95,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(95,15): error C3646: “m_sDefGateway”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(95,28): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(96,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(96,15): error C3646: “m_sDhcpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(96,26): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.cpp(91,20): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(93,19): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(112,20): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(114,19): error C2039: "clear": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(141,9): error C2065: “m_sName”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(142,9): error C2065: “m_sDesc”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(143,9): error C2065: “m_sPriWins”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(144,9): error C2065: “m_sSecWins”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(145,9): error C2065: “m_sDhcpAddr”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(156,26): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(157,26): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(160,26): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(161,26): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(169,20): error C2039: "sIp": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(170,20): error C2039: "sSubnet": 不是 "_IPINFO" 的成员
1>G:\happynwindows\happynet\netadapter.h(26,8): message : 参见“_IPINFO”的声明
1>G:\happynwindows\happynet\netadapter.cpp(180,27): error C2039: "push_back": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(197,36): error C2039: "push_back": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(281,30): error C2039: "GetStringFromArray": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(281,92): error C2270: “GetStringFromArray”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(283,41): error C2039: "size": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(284,30): error C2676: 二进制“[”:“const StringArray”不定义该运算符或到预定义运算符可接收的类型的转换
1>G:\happynwindows\happynet\netadapter.cpp(295,26): error C2039: "GetAdapterTypeString": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(317,26): error C2039: "GetAdapterName": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(317,52): error C2270: “GetAdapterName”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(317,61): error C2065: “m_sName”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(318,26): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(318,57): error C2270: “GetAdapterDescription”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(318,66): error C2065: “m_sDesc”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.cpp(322,75): error C2039: "size": 不是 "StringArray" 的成员
1>G:\happynwindows\happynet\netadapter.h(32,7): message : 参见“StringArray”的声明
1>G:\happynwindows\happynet\netadapter.cpp(323,30): error C2039: "GetDnsAddr": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\netadapter.cpp(323,58): error C2270: “GetDnsAddr”: 非成员函数上不允许修饰符
1>G:\happynwindows\happynet\netadapter.cpp(323,58): fatal error C1003: 错误计数超过 100;正在停止编译
1>happynwinclient.cpp
1>G:\happynwindows\happynet\netadapter.h(27,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(27,17): error C3646: “sIp”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(27,20): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(28,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(28,17): error C3646: “sSubnet”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(28,24): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(32,41): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(32,48): error C2065: “string”: 未声明的标识符
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2923: "std::vector": "string" 不是参数 "_Ty" 的有效 模板 类型参数
1>G:\happynwindows\happynet\netadapter.h(32,48): message : 参见“string”的声明
1>G:\happynwindows\happynet\netadapter.h(32,28): error C2976: “std::vector'”: 模板 参数太少
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(32,50): error C2955: “std::vector”: 使用 类 模板 需要 模板 参数列表
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(494,14): message : 参见“std::vector”的声明
1>G:\happynwindows\happynet\netadapter.h(52,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(52,13): error C3646: “GetAdapterName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(52,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(52,35): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(53,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(53,13): error C3646: “GetAdapterDescription”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(53,34): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(53,42): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(62,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(62,17): error C3646: “GetIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(62,26): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(62,45): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(63,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(63,17): error C3646: “GetSubnetForIpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(63,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(63,54): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(64,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(64,17): error C3646: “GetDnsAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(64,27): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(64,47): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(65,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(65,17): error C3646: “GetCurrentIpAddress”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(65,36): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(65,44): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(69,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(69,13): error C3646: “GetDchpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(69,24): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(69,32): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(73,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(73,13): error C3646: “GetPrimaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(73,33): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(73,41): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(74,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(74,13): error C3646: “GetSecondaryWinsServer”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(74,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(74,43): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(76,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(76,17): error C3646: “GetGatewayAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(76,31): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(76,74): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(79,12): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(79,20): error C3646: “GetAdapterTypeString”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(79,41): error C2275: “UINT”: 应为表达式而不是类型
1>G:\happynwindows\happynet\netadapter.h(79,46): error C2146: 语法错误: 缺少“)”(在标识符“nType”的前面)
1>G:\happynwindows\happynet\netadapter.h(79,52): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(87,10): error C2039: "string": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(87,17): error C3646: “GetStringFromArray”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(87,35): error C2059: 语法错误:“(”
1>G:\happynwindows\happynet\netadapter.h(87,78): error C2238: 意外的标记位于“;”之前
1>G:\happynwindows\happynet\netadapter.h(91,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(91,15): error C3646: “m_sName”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(91,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(92,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(92,15): error C3646: “m_sDesc”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(92,22): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(93,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(93,15): error C3646: “m_sPriWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(93,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(94,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(94,15): error C3646: “m_sSecWins”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(94,25): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(95,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(95,15): error C3646: “m_sDefGateway”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(95,28): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\netadapter.h(96,5): error C2039: "wstring": 不是 "std" 的成员
1>G:\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector(24,1): message : 参见“std”的声明
1>G:\happynwindows\happynet\netadapter.h(96,15): error C3646: “m_sDhcpAddr”: 未知重写说明符
1>G:\happynwindows\happynet\netadapter.h(96,26): error C4430: 缺少类型说明符 - 假定为 int。注意: C++ 不支持默认 int
1>G:\happynwindows\happynet\resource.h(1395,1): warning C4005: “_WIN32_WINNT”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(218,1): message : 参见“_WIN32_WINNT”的前一个定义
1>G:\happynwindows\happynet\resource.h(1419,1): warning C4005: “_WIN32_IE”: 宏重定义
1>G:\Windows Kits\10\Include\10.0.19041.0\shared\sdkddkver.h(263,1): message : 参见“_WIN32_IE”的前一个定义
1>G:\happynwindows\happynet\happynwinclient.cpp(785,29): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,70): error C2039: "GetAdapterDescription": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,111): error C2039: "GetAdapterName": 不是 "CNetworkAdapter" 的成员
1>G:\happynwindows\happynet\netadapter.h(44,7): message : 参见“CNetworkAdapter”的声明
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): warning C4473: “swprintf_s”: 没有为格式字符串传递足够的参数
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): message : 占位符和其参数预计 2 可变参数,但提供的却是 0 参数
1>G:\happynwindows\happynet\happynwinclient.cpp(790,47): message : 缺失的可变参数 1 为格式字符串“%s”所需
1>正在生成代码...
1>已完成生成项目“happynet.vcxproj”的操作 - 失败。
========== 版本: 0 成功,1 失败,0 更新,0 跳过 ==========
========== 占用时间 00:05.944 ==========
`
|
non_defect
|
根本无法编译,你们真的是开源吗?还是只是想蹭开源的噱头圈钱? 已启动生成 已启动生成 项目 happynet 配置 debug cl 命令行 warning “gm”选项已否决,并将在将来的版本中移除 utils cpp service cpp registry cpp process cpp netinterface cpp tray cpp g happynwindows happynet resource h warning “ winnt” 宏重定义 g windows kits include shared sdkddkver h message 参见“ winnt”的前一个定义 g happynwindows happynet resource h warning “ ie” 宏重定义 g windows kits include shared sdkddkver h message 参见“ ie”的前一个定义 systemsrv cpp netadapter cpp g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “sip” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “ssubnet” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “string” 未声明的标识符 g happynwindows happynet netadapter h error std vector string 不是参数 ty 的有效 模板 类型参数 g happynwindows happynet netadapter h message 参见“string”的声明 g happynwindows happynet netadapter h error “std vector ” 模板 参数太少 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std vector”的声明 g happynwindows happynet netadapter h error “std vector” 使用 类 模板 需要 模板 参数列表 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std vector”的声明 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadaptername” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadapterdescription” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getipaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getsubnetforipaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getdnsaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getcurrentipaddress” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getdchpaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getprimarywinsserver” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getsecondarywinsserver” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getgatewayaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadaptertypestring” 未知重写说明符 g happynwindows happynet netadapter h error “uint” 应为表达式而不是类型 g happynwindows happynet netadapter h error 语法错误 缺少“ ” 在标识符“ntype”的前面 g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getstringfromarray” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sname” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdesc” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m spriwins” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m ssecwins” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdefgateway” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdhcpaddr” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter cpp error clear 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error clear 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error clear 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error clear 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error “m sname” 未声明的标识符 g happynwindows happynet netadapter cpp error “m sdesc” 未声明的标识符 g happynwindows happynet netadapter cpp error “m spriwins” 未声明的标识符 g happynwindows happynet netadapter cpp error “m ssecwins” 未声明的标识符 g happynwindows happynet netadapter cpp error “m sdhcpaddr” 未声明的标识符 g happynwindows happynet netadapter cpp error sip 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error ssubnet 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error sip 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error ssubnet 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error sip 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error ssubnet 不是 ipinfo 的成员 g happynwindows happynet netadapter h message 参见“ ipinfo”的声明 g happynwindows happynet netadapter cpp error push back 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error push back 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error getstringfromarray 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet netadapter cpp error “getstringfromarray” 非成员函数上不允许修饰符 g happynwindows happynet netadapter cpp error size 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error 二进制“ ” “const stringarray”不定义该运算符或到预定义运算符可接收的类型的转换 g happynwindows happynet netadapter cpp error getadaptertypestring 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet netadapter cpp error getadaptername 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet netadapter cpp error “getadaptername” 非成员函数上不允许修饰符 g happynwindows happynet netadapter cpp error “m sname” 未声明的标识符 g happynwindows happynet netadapter cpp error getadapterdescription 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet netadapter cpp error “getadapterdescription” 非成员函数上不允许修饰符 g happynwindows happynet netadapter cpp error “m sdesc” 未声明的标识符 g happynwindows happynet netadapter cpp error size 不是 stringarray 的成员 g happynwindows happynet netadapter h message 参见“stringarray”的声明 g happynwindows happynet netadapter cpp error getdnsaddr 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet netadapter cpp error “getdnsaddr” 非成员函数上不允许修饰符 g happynwindows happynet netadapter cpp fatal error 错误计数超过 ;正在停止编译 happynwinclient cpp g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “sip” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “ssubnet” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “string” 未声明的标识符 g happynwindows happynet netadapter h error std vector string 不是参数 ty 的有效 模板 类型参数 g happynwindows happynet netadapter h message 参见“string”的声明 g happynwindows happynet netadapter h error “std vector ” 模板 参数太少 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std vector”的声明 g happynwindows happynet netadapter h error “std vector” 使用 类 模板 需要 模板 参数列表 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std vector”的声明 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadaptername” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadapterdescription” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getipaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getsubnetforipaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getdnsaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getcurrentipaddress” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getdchpaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getprimarywinsserver” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getsecondarywinsserver” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getgatewayaddr” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getadaptertypestring” 未知重写说明符 g happynwindows happynet netadapter h error “uint” 应为表达式而不是类型 g happynwindows happynet netadapter h error 语法错误 缺少“ ” 在标识符“ntype”的前面 g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error string 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “getstringfromarray” 未知重写说明符 g happynwindows happynet netadapter h error 语法错误 “ ” g happynwindows happynet netadapter h error 意外的标记位于“ ”之前 g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sname” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdesc” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m spriwins” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m ssecwins” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdefgateway” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet netadapter h error wstring 不是 std 的成员 g microsoft visual studio enterprise vc tools msvc include vector message 参见“std”的声明 g happynwindows happynet netadapter h error “m sdhcpaddr” 未知重写说明符 g happynwindows happynet netadapter h error 缺少类型说明符 假定为 int。注意 c 不支持默认 int g happynwindows happynet resource h warning “ winnt” 宏重定义 g windows kits include shared sdkddkver h message 参见“ winnt”的前一个定义 g happynwindows happynet resource h warning “ ie” 宏重定义 g windows kits include shared sdkddkver h message 参见“ ie”的前一个定义 g happynwindows happynet happynwinclient cpp error getadapterdescription 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet happynwinclient cpp error getadapterdescription 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet happynwinclient cpp error getadaptername 不是 cnetworkadapter 的成员 g happynwindows happynet netadapter h message 参见“cnetworkadapter”的声明 g happynwindows happynet happynwinclient cpp warning “swprintf s” 没有为格式字符串传递足够的参数 g happynwindows happynet happynwinclient cpp message 占位符和其参数预计 可变参数,但提供的却是 参数 g happynwindows happynet happynwinclient cpp message 缺失的可变参数 为格式字符串“ s”所需 正在生成代码 已完成生成项目“happynet vcxproj”的操作 失败。 版本 成功, 失败, 更新, 跳过 占用时间
| 0
|
50,055
| 7,558,870,534
|
IssuesEvent
|
2018-04-20 00:32:09
|
canjs/canjs
|
https://api.github.com/repos/canjs/canjs
|
opened
|
Improve the content of the CanJS documentation
|
documentation
|
__TLDR:__ we propose reviewing our current docs to look for:
- pages that don’t have useful content
- broken links
- bad/broken code examples & demos
- other ways we can improve them to make them more useful
## Problem
Our docs have evolved over the years without a thorough review. Inevitably, some docs have deteriorated with broken links, out-of-date examples, and busted demos. Some pages lack good content or examples.
## Solution
Review the most popular pages (maybe top 50?) to:
- Fix broken links, code examples, demos, etc.
- Look for ways to improve the content
Additionally, we should tackle these existing issues:
- [Consistently use const/let in sample code](https://github.com/canjs/canjs/issues/3846)
- [Consistently use ES6 import syntax instead of CJS in sample code](https://github.com/canjs/canjs/issues/3847)
- [Improve the descriptions of (at least) the Core packages](https://github.com/canjs/canjs/issues/4057)
- [Improve docs on loading templates](https://github.com/canjs/can-stache/issues/429)
We should also look at [issues with the “documentation” label](https://github.com/search?o=desc&q=org%3Acanjs+is%3Aopen+is%3Aissue+label%3Adocumentation&s=created&type=Issues).
As a stretch goal, we should consider:
- [Adding a way for users to give us feedback on whether docs are good/useful](https://github.com/canjs/bit-docs-html-canjs/issues/451)
- Making it possible for our demos/examples to open in an online code editor
|
1.0
|
Improve the content of the CanJS documentation - __TLDR:__ we propose reviewing our current docs to look for:
- pages that don’t have useful content
- broken links
- bad/broken code examples & demos
- other ways we can improve them to make them more useful
## Problem
Our docs have evolved over the years without a thorough review. Inevitably, some docs have deteriorated with broken links, out-of-date examples, and busted demos. Some pages lack good content or examples.
## Solution
Review the most popular pages (maybe top 50?) to:
- Fix broken links, code examples, demos, etc.
- Look for ways to improve the content
Additionally, we should tackle these existing issues:
- [Consistently use const/let in sample code](https://github.com/canjs/canjs/issues/3846)
- [Consistently use ES6 import syntax instead of CJS in sample code](https://github.com/canjs/canjs/issues/3847)
- [Improve the descriptions of (at least) the Core packages](https://github.com/canjs/canjs/issues/4057)
- [Improve docs on loading templates](https://github.com/canjs/can-stache/issues/429)
We should also look at [issues with the “documentation” label](https://github.com/search?o=desc&q=org%3Acanjs+is%3Aopen+is%3Aissue+label%3Adocumentation&s=created&type=Issues).
As a stretch goal, we should consider:
- [Adding a way for users to give us feedback on whether docs are good/useful](https://github.com/canjs/bit-docs-html-canjs/issues/451)
- Making it possible for our demos/examples to open in an online code editor
|
non_defect
|
improve the content of the canjs documentation tldr we propose reviewing our current docs to look for pages that don’t have useful content broken links bad broken code examples demos other ways we can improve them to make them more useful problem our docs have evolved over the years without a thorough review inevitably some docs have deteriorated with broken links out of date examples and busted demos some pages lack good content or examples solution review the most popular pages maybe top to fix broken links code examples demos etc look for ways to improve the content additionally we should tackle these existing issues we should also look at as a stretch goal we should consider making it possible for our demos examples to open in an online code editor
| 0
|
7,115
| 2,610,327,757
|
IssuesEvent
|
2015-02-26 19:45:36
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Typo
|
auto-migrated Priority-Low Type-Defect
|
```
Need a period before "Try" on the Tatooine mission objective about Tusken
Raiders
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 Jun 2011 at 12:31
|
1.0
|
Typo - ```
Need a period before "Try" on the Tatooine mission objective about Tusken
Raiders
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 Jun 2011 at 12:31
|
defect
|
typo need a period before try on the tatooine mission objective about tusken raiders original issue reported on code google com by gmail com on jun at
| 1
|
35,110
| 7,573,256,959
|
IssuesEvent
|
2018-04-23 17:13:03
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Unexpected behavior for request->getRequestTarget()
|
Defect http
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6.1.
* Platform and Target: Apache/2.4.27, MySQL 5.6.37
### What you did
I upgraded my App from 3.5 to 3.6. and my app is located into a sub directory called myapp so the full URL is http://development.local/myapp/
### What happened
I got a warning error says `Accessing 'here' as a property will be removed in 4.0.0. Use request->getRequestTarget() instead. - /www/myapp/src/Template/Layout/front.ctp, Line 45` so I used `getRequestTarget() ` instead of `here`
### What you expected to happen
I expected that `request->getRequestTarget()` will return the same reult from `request->here` but this was not the case. As I told you above that my app is localed at http://development.local/myapp/
in 3.5 & 3.6, `request->here` returns `/myapp/` but on 3.6 `request->getRequestTarget()` returns `/` which is a different result and it should return `/myapp/`
Thanks
|
1.0
|
Unexpected behavior for request->getRequestTarget() - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6.1.
* Platform and Target: Apache/2.4.27, MySQL 5.6.37
### What you did
I upgraded my App from 3.5 to 3.6. and my app is located into a sub directory called myapp so the full URL is http://development.local/myapp/
### What happened
I got a warning error says `Accessing 'here' as a property will be removed in 4.0.0. Use request->getRequestTarget() instead. - /www/myapp/src/Template/Layout/front.ctp, Line 45` so I used `getRequestTarget() ` instead of `here`
### What you expected to happen
I expected that `request->getRequestTarget()` will return the same reult from `request->here` but this was not the case. As I told you above that my app is localed at http://development.local/myapp/
in 3.5 & 3.6, `request->here` returns `/myapp/` but on 3.6 `request->getRequestTarget()` returns `/` which is a different result and it should return `/myapp/`
Thanks
|
defect
|
unexpected behavior for request getrequesttarget this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target apache mysql what you did i upgraded my app from to and my app is located into a sub directory called myapp so the full url is what happened i got a warning error says accessing here as a property will be removed in use request getrequesttarget instead www myapp src template layout front ctp line so i used getrequesttarget instead of here what you expected to happen i expected that request getrequesttarget will return the same reult from request here but this was not the case as i told you above that my app is localed at in request here returns myapp but on request getrequesttarget returns which is a different result and it should return myapp thanks
| 1
|
23,399
| 3,812,640,247
|
IssuesEvent
|
2016-03-27 18:43:59
|
zealdocs/zeal
|
https://api.github.com/repos/zealdocs/zeal
|
closed
|
Chinese characters are corrupted after unpacking HTML docset
|
Platform: Windows Resolution: Upstream Problem Type: Defect
|
My system is Windows 7 zh_CN.
System default code is GBK.
In HTML.tgz\HTML.docset\Contents\Resources\Documents\mdn.mozillademos.org\zh-CN\docs\Web\HTML\Element\datalist$samples\ have a file is 示例426e.html.But file name is ê?ày426e.html in system.

Then std::ifstream can't find file and QIODevice::read: device not open.
|
1.0
|
Chinese characters are corrupted after unpacking HTML docset - My system is Windows 7 zh_CN.
System default code is GBK.
In HTML.tgz\HTML.docset\Contents\Resources\Documents\mdn.mozillademos.org\zh-CN\docs\Web\HTML\Element\datalist$samples\ have a file is 示例426e.html.But file name is ê?ày426e.html in system.

Then std::ifstream can't find file and QIODevice::read: device not open.
|
defect
|
chinese characters are corrupted after unpacking html docset my system is windows zh cn system default code is gbk in html tgz html docset contents resources documents mdn mozillademos org zh cn docs web html element datalist samples have a file is html but file name is ê html in system then std ifstream can t find file and qiodevice read device not open
| 1
|
55,563
| 14,546,094,399
|
IssuesEvent
|
2020-12-15 20:41:01
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
ColumnToggler: removes the Responsive priority of the selected columns
|
defect
|
Hello,
interaction with `<p:columnToggler>` removes the 'Responsive' priority of the selected columns.
ex:
uncheck then re-select the display of a column, it will lose its priority when resizing.
Thanks.
|
1.0
|
ColumnToggler: removes the Responsive priority of the selected columns - Hello,
interaction with `<p:columnToggler>` removes the 'Responsive' priority of the selected columns.
ex:
uncheck then re-select the display of a column, it will lose its priority when resizing.
Thanks.
|
defect
|
columntoggler removes the responsive priority of the selected columns hello interaction with removes the responsive priority of the selected columns ex uncheck then re select the display of a column it will lose its priority when resizing thanks
| 1
|
12,284
| 5,183,440,260
|
IssuesEvent
|
2017-01-20 00:51:31
|
docker/docker
|
https://api.github.com/repos/docker/docker
|
closed
|
Trying to run docker on armhf, "Operation not supported"
|
area/builder platform/arm
|
**Description**
Hello, i'm trying to run Docker on allwinner A10 SOC within Cubieboard HW. I compiled linux kernel v4.9.0, and enabled all available options, that docker requires, except for few.
Using this script: `https://raw.githubusercontent.com/docker/docker/master/contrib/check-config.sh` i'm checking what dependencies are missing, while compiling kernel. Here is the result after running script: [kernel-config.txt](https://github.com/docker/docker/files/697555/kernel-config.txt)
Missing modules marked "missing". Seems, that DEVPTS_MULTIPLE_INSTANCES is no more available in 4.9.0 kernel and other missing elements is not supported on arm architecture. I was not touch aufs support.
`uname -a`: [uname.txt](https://github.com/docker/docker/files/697567/uname.txt)
Docker log: [docker-log.txt](https://github.com/docker/docker/files/697569/docker-log.txt)
`docker info`: [docker-info.txt](https://github.com/docker/docker/files/697573/docker-info.txt)
`lsmod`: [lsmod.txt](https://github.com/docker/docker/files/697574/lsmod.txt)
`mount`: [mount.txt](https://github.com/docker/docker/files/697577/mount.txt)
`ifconfig`: [ifconfig.txt](https://github.com/docker/docker/files/697579/ifconfig.txt)
**Steps to reproduce the issue:**
1. edit /etc/apk/repositories - uncomment community repos
2. apk add docker
3. service docker start
4. docker run -it armhf/alpine sh
**Describe the results you received:**
`docker: Error response from daemon: operation not supported.`
**Describe the results you expected:**
Container running.
**Additional information you deem important (e.g. issue happens only occasionally):**
I may include the Dockerfile, which builds u-boot, patches and compiles linux-kernel, adds Alpine Linux rootFS and makes sdcard image, ready to burn onto SD if someone want to reproduce issue with same hardware (Cubietech Cubieboard 1 / A10 Allwinner SoC)
**Output of `docker version`:**
```
Client:
Version: 1.12.5
API: 1.24
Go: go1.7.3
Git: v1.12.5
Built: Jan 9 2017
OS/arch: linux/arm
Server: same
```
**Output of `docker info`:**
```
Containers: 0
Running: 0
Paused: 0
Stopped: 0
Images: 1
Server Version: 1.12.5
Storage Driver: overlay
Backing Filesystem: tmpfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: null host bridge overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options:
Kernel Version: 4.9.0
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: armv7l
CPUs: 1
Total Memory: 996.5 MiB
Name: aldock
ID: SAAM:KJKB:SGC3:Z7SM:OKLK:2S4Y:OQEC:T6XX:6L6J:KRUV:XWV6:H7UB
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
Cubietech Cubieboard 1 on A10 Alwinner SoC hardware
Linux Kernel 4.9.0, manually customized
Alpine linux 3.5.0
Docker downloaded from community repositories, not yet available in main repo (not considered stable on Alpine ?)
|
1.0
|
Trying to run docker on armhf, "Operation not supported" - **Description**
Hello, i'm trying to run Docker on allwinner A10 SOC within Cubieboard HW. I compiled linux kernel v4.9.0, and enabled all available options, that docker requires, except for few.
Using this script: `https://raw.githubusercontent.com/docker/docker/master/contrib/check-config.sh` i'm checking what dependencies are missing, while compiling kernel. Here is the result after running script: [kernel-config.txt](https://github.com/docker/docker/files/697555/kernel-config.txt)
Missing modules marked "missing". Seems, that DEVPTS_MULTIPLE_INSTANCES is no more available in 4.9.0 kernel and other missing elements is not supported on arm architecture. I was not touch aufs support.
`uname -a`: [uname.txt](https://github.com/docker/docker/files/697567/uname.txt)
Docker log: [docker-log.txt](https://github.com/docker/docker/files/697569/docker-log.txt)
`docker info`: [docker-info.txt](https://github.com/docker/docker/files/697573/docker-info.txt)
`lsmod`: [lsmod.txt](https://github.com/docker/docker/files/697574/lsmod.txt)
`mount`: [mount.txt](https://github.com/docker/docker/files/697577/mount.txt)
`ifconfig`: [ifconfig.txt](https://github.com/docker/docker/files/697579/ifconfig.txt)
**Steps to reproduce the issue:**
1. edit /etc/apk/repositories - uncomment community repos
2. apk add docker
3. service docker start
4. docker run -it armhf/alpine sh
**Describe the results you received:**
`docker: Error response from daemon: operation not supported.`
**Describe the results you expected:**
Container running.
**Additional information you deem important (e.g. issue happens only occasionally):**
I may include the Dockerfile, which builds u-boot, patches and compiles linux-kernel, adds Alpine Linux rootFS and makes sdcard image, ready to burn onto SD if someone want to reproduce issue with same hardware (Cubietech Cubieboard 1 / A10 Allwinner SoC)
**Output of `docker version`:**
```
Client:
Version: 1.12.5
API: 1.24
Go: go1.7.3
Git: v1.12.5
Built: Jan 9 2017
OS/arch: linux/arm
Server: same
```
**Output of `docker info`:**
```
Containers: 0
Running: 0
Paused: 0
Stopped: 0
Images: 1
Server Version: 1.12.5
Storage Driver: overlay
Backing Filesystem: tmpfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: null host bridge overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options:
Kernel Version: 4.9.0
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: armv7l
CPUs: 1
Total Memory: 996.5 MiB
Name: aldock
ID: SAAM:KJKB:SGC3:Z7SM:OKLK:2S4Y:OQEC:T6XX:6L6J:KRUV:XWV6:H7UB
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
Cubietech Cubieboard 1 on A10 Alwinner SoC hardware
Linux Kernel 4.9.0, manually customized
Alpine linux 3.5.0
Docker downloaded from community repositories, not yet available in main repo (not considered stable on Alpine ?)
|
non_defect
|
trying to run docker on armhf operation not supported description hello i m trying to run docker on allwinner soc within cubieboard hw i compiled linux kernel and enabled all available options that docker requires except for few using this script i m checking what dependencies are missing while compiling kernel here is the result after running script missing modules marked missing seems that devpts multiple instances is no more available in kernel and other missing elements is not supported on arm architecture i was not touch aufs support uname a docker log docker info lsmod mount ifconfig steps to reproduce the issue edit etc apk repositories uncomment community repos apk add docker service docker start docker run it armhf alpine sh describe the results you received docker error response from daemon operation not supported describe the results you expected container running additional information you deem important e g issue happens only occasionally i may include the dockerfile which builds u boot patches and compiles linux kernel adds alpine linux rootfs and makes sdcard image ready to burn onto sd if someone want to reproduce issue with same hardware cubietech cubieboard allwinner soc output of docker version client version api go git built jan os arch linux arm server same output of docker info containers running paused stopped images server version storage driver overlay backing filesystem tmpfs logging driver json file cgroup driver cgroupfs plugins volume local network null host bridge overlay swarm inactive runtimes runc default runtime runc security options kernel version operating system alpine linux ostype linux architecture cpus total memory mib name aldock id saam kjkb oklk oqec kruv docker root dir var lib docker debug mode client false debug mode server false registry insecure registries additional environment details aws virtualbox physical etc cubietech cubieboard on alwinner soc hardware linux kernel manually customized alpine linux docker downloaded from community repositories not yet available in main repo not considered stable on alpine
| 0
|
143,812
| 13,086,255,634
|
IssuesEvent
|
2020-08-02 05:20:04
|
mrinal-chandra/memorie_auto_mode
|
https://api.github.com/repos/mrinal-chandra/memorie_auto_mode
|
closed
|
short description
|
documentation
|
a longer description of the task and the project to organize and complete it.
|
1.0
|
short description - a longer description of the task and the project to organize and complete it.
|
non_defect
|
short description a longer description of the task and the project to organize and complete it
| 0
|
59,670
| 17,023,199,282
|
IssuesEvent
|
2021-07-03 00:49:28
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Allow nodes on the end and allow map calls to fetch them
|
Component: api Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 3.16pm, Wednesday, 23rd January 2008]**
Currently nodes are not permitted on the edges of the map (90N/S or 180E/W), this patch fixes that.
Secondly, it's currently next to impossible to catch those in a map call, because the API complains as soon as you go over the edge. So this patch changes the code to crop the user's bbox to the map area and only complain if what the user requested was completely outside.
|
1.0
|
Allow nodes on the end and allow map calls to fetch them - **[Submitted to the original trac issue database at 3.16pm, Wednesday, 23rd January 2008]**
Currently nodes are not permitted on the edges of the map (90N/S or 180E/W), this patch fixes that.
Secondly, it's currently next to impossible to catch those in a map call, because the API complains as soon as you go over the edge. So this patch changes the code to crop the user's bbox to the map area and only complain if what the user requested was completely outside.
|
defect
|
allow nodes on the end and allow map calls to fetch them currently nodes are not permitted on the edges of the map s or w this patch fixes that secondly it s currently next to impossible to catch those in a map call because the api complains as soon as you go over the edge so this patch changes the code to crop the user s bbox to the map area and only complain if what the user requested was completely outside
| 1
|
9,243
| 2,615,141,075
|
IssuesEvent
|
2015-03-01 06:15:41
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
closed
|
r65 tries second half of each PIN twice!
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Run reaver normally until it guesses the first half of the pin and jumps to
90+ %
2.
3.
What is the expected output?
-------------------------
[+] Trying pin 49387179
[+] Trying pin 49382341
[!] WARNING: Receive timeout occurred
[+] Trying pin 49382341
[+] Trying pin 49387605
[+] 91.15% complete @ 2012-01-04 23:49:00 (10 seconds/attempt)
[+] Trying pin 49384956
[+] Trying pin 49381337
[+] Trying pin 49381641
[+] 91.16% complete @ 2012-01-04 23:49:21 (10 seconds/attempt)
[+] Trying pin 49382136
-------------------------
What do you see instead?
-------------------------
[+] Trying pin 49387179
[+] Trying pin 49382341
[+] Trying pin 49382341
[!] WARNING: Receive timeout occurred
[+] Trying pin 49382341
[+] Trying pin 49387605
[+] Trying pin 49387605
[+] 91.15% complete @ 2012-01-04 23:49:00 (10 seconds/attempt)
[+] Trying pin 49384956
[+] Trying pin 49384956
[+] Trying pin 49381337
[+] Trying pin 49381337
[+] Trying pin 49381641
[+] 91.16% complete @ 2012-01-04 23:49:21 (10 seconds/attempt)
[+] Trying pin 49381641
[+] Trying pin 49382136
-------------------------
What version of the product are you using? On what operating system?
reaver r65 on ubuntu 11.10
Please provide any additional information below.
PINS were tried only once before completing the guess of the 1st half of the
PIN. Upon starting the 2nd half of the PIN each 2nd half of the PIN is tried
twice with no intervening timeouts or failures. This problem started occurring
on r65 and was not present on earlier versions. FYI. Thanks!
```
Original issue reported on code.google.com by `vsil...@gmail.com` on 5 Jan 2012 at 5:53
|
1.0
|
r65 tries second half of each PIN twice! - ```
What steps will reproduce the problem?
1. Run reaver normally until it guesses the first half of the pin and jumps to
90+ %
2.
3.
What is the expected output?
-------------------------
[+] Trying pin 49387179
[+] Trying pin 49382341
[!] WARNING: Receive timeout occurred
[+] Trying pin 49382341
[+] Trying pin 49387605
[+] 91.15% complete @ 2012-01-04 23:49:00 (10 seconds/attempt)
[+] Trying pin 49384956
[+] Trying pin 49381337
[+] Trying pin 49381641
[+] 91.16% complete @ 2012-01-04 23:49:21 (10 seconds/attempt)
[+] Trying pin 49382136
-------------------------
What do you see instead?
-------------------------
[+] Trying pin 49387179
[+] Trying pin 49382341
[+] Trying pin 49382341
[!] WARNING: Receive timeout occurred
[+] Trying pin 49382341
[+] Trying pin 49387605
[+] Trying pin 49387605
[+] 91.15% complete @ 2012-01-04 23:49:00 (10 seconds/attempt)
[+] Trying pin 49384956
[+] Trying pin 49384956
[+] Trying pin 49381337
[+] Trying pin 49381337
[+] Trying pin 49381641
[+] 91.16% complete @ 2012-01-04 23:49:21 (10 seconds/attempt)
[+] Trying pin 49381641
[+] Trying pin 49382136
-------------------------
What version of the product are you using? On what operating system?
reaver r65 on ubuntu 11.10
Please provide any additional information below.
PINS were tried only once before completing the guess of the 1st half of the
PIN. Upon starting the 2nd half of the PIN each 2nd half of the PIN is tried
twice with no intervening timeouts or failures. This problem started occurring
on r65 and was not present on earlier versions. FYI. Thanks!
```
Original issue reported on code.google.com by `vsil...@gmail.com` on 5 Jan 2012 at 5:53
|
defect
|
tries second half of each pin twice what steps will reproduce the problem run reaver normally until it guesses the first half of the pin and jumps to what is the expected output trying pin trying pin warning receive timeout occurred trying pin trying pin complete seconds attempt trying pin trying pin trying pin complete seconds attempt trying pin what do you see instead trying pin trying pin trying pin warning receive timeout occurred trying pin trying pin trying pin complete seconds attempt trying pin trying pin trying pin trying pin trying pin complete seconds attempt trying pin trying pin what version of the product are you using on what operating system reaver on ubuntu please provide any additional information below pins were tried only once before completing the guess of the half of the pin upon starting the half of the pin each half of the pin is tried twice with no intervening timeouts or failures this problem started occurring on and was not present on earlier versions fyi thanks original issue reported on code google com by vsil gmail com on jan at
| 1
|
187,795
| 6,761,185,862
|
IssuesEvent
|
2017-10-25 00:04:46
|
sul-dlss/preservation_core_catalog
|
https://api.github.com/repos/sul-dlss/preservation_core_catalog
|
closed
|
(M2C, PC) split out MoabToCatalog.check_existence into separate existence check and seed methods
|
high priority needs review refactor
|
we want two methods that will be called by rake tasks: one to walk the storage roots to seed the empty catalog, one to walk the storage roots on a regular basis and check the moab objects on disk against the catalog database.
currently, this is conflated into one method, `MoabToCatalog.check_moab_to_catalog_existence`. the code for the two new methods might be largely similar, but it's a short method, and you know, shameless green.
|
1.0
|
(M2C, PC) split out MoabToCatalog.check_existence into separate existence check and seed methods - we want two methods that will be called by rake tasks: one to walk the storage roots to seed the empty catalog, one to walk the storage roots on a regular basis and check the moab objects on disk against the catalog database.
currently, this is conflated into one method, `MoabToCatalog.check_moab_to_catalog_existence`. the code for the two new methods might be largely similar, but it's a short method, and you know, shameless green.
|
non_defect
|
pc split out moabtocatalog check existence into separate existence check and seed methods we want two methods that will be called by rake tasks one to walk the storage roots to seed the empty catalog one to walk the storage roots on a regular basis and check the moab objects on disk against the catalog database currently this is conflated into one method moabtocatalog check moab to catalog existence the code for the two new methods might be largely similar but it s a short method and you know shameless green
| 0
|
12,211
| 2,685,470,893
|
IssuesEvent
|
2015-03-30 01:20:18
|
IssueMigrationTest/Test5
|
https://api.github.com/repos/IssueMigrationTest/Test5
|
closed
|
libgcc_s_dw2-1.dll is never found no matter how many times you re-download
|
auto-migrated Priority-Medium Type-Defect
|
**Issue by JRonline...@gmail.com**
_14 May 2011 at 8:23 GMT_
_Originally opened on Google Code_
----
```
I always see a message that libgcc_s_dw2-1.dill cannot be found re-doanload to
fix this problem
What version of the product are you using? On what operating system?
i am using the newest shedskin avalible to windows an am using windows vista.
could you put libgcc_s_dw2-1.dill for download seperately and tell me where to
put it?
```
|
1.0
|
libgcc_s_dw2-1.dll is never found no matter how many times you re-download - **Issue by JRonline...@gmail.com**
_14 May 2011 at 8:23 GMT_
_Originally opened on Google Code_
----
```
I always see a message that libgcc_s_dw2-1.dill cannot be found re-doanload to
fix this problem
What version of the product are you using? On what operating system?
i am using the newest shedskin avalible to windows an am using windows vista.
could you put libgcc_s_dw2-1.dill for download seperately and tell me where to
put it?
```
|
defect
|
libgcc s dll is never found no matter how many times you re download issue by jronline gmail com may at gmt originally opened on google code i always see a message that libgcc s dill cannot be found re doanload to fix this problem what version of the product are you using on what operating system i am using the newest shedskin avalible to windows an am using windows vista could you put libgcc s dill for download seperately and tell me where to put it
| 1
|
60,310
| 17,023,393,804
|
IssuesEvent
|
2021-07-03 01:47:56
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Can't delete circular ways
|
Component: potlatch (flash editor) Priority: critical Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 3.30pm, Monday, 27th April 2009]**
Because it tries to delete the start/end node... twice.
PLEASE JUST SHOOT ME NOW
|
1.0
|
Can't delete circular ways - **[Submitted to the original trac issue database at 3.30pm, Monday, 27th April 2009]**
Because it tries to delete the start/end node... twice.
PLEASE JUST SHOOT ME NOW
|
defect
|
can t delete circular ways because it tries to delete the start end node twice please just shoot me now
| 1
|
83,014
| 10,316,717,950
|
IssuesEvent
|
2019-08-30 10:43:18
|
luniehq/lunie
|
https://api.github.com/repos/luniehq/lunie
|
closed
|
Vesting account and coins support
|
design-work-needed :art: epic stale
|
**Is your feature request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I'd like to know if any of my coins are vesting (locked) and the time they vest so I can choose those vesting coins over the regular ones to delegate.
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
- Clearly separate vesting coins from regular coins on the wallet balance
- Choose either vesting (`original_vesting`) or regular coins (`coins`) for staking (delegation, redelegation) and governance.
- If (re)delegate using my vesting coins, I want to see that the delegation was made with them and not with regular coins
- Show a message on Send action modal that the `original_vesting` coins are locked and I can't transfer them until `end_time` date
- (optional) show a time progress bar for when coins become vested (unlocked) on the Wallet Page
**Describe alternatives you've considered**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
- create a separate state for vesting balances (`original_vesting`)
- add a dropdown on action modals to choose which coin to use: regular or vesting (`original_vesting`)
**[BaseVestingAccount](https://github.com/cosmos/cosmos-sdk/blob/4b765b6ab4dfac38a896b66478b3b14b40d0c731/x/auth/account.go#L178-L191
)**
A vesting account has the following additional fields:
```
type BaseVestingAccount struct {
*BaseAccount
OriginalVesting sdk.Coins `json:"original_vesting"` // coins in account upon initialization
DelegatedFree sdk.Coins `json:"delegated_free"` // coins that are vested and delegated
DelegatedVesting sdk.Coins `json:"delegated_vesting"` // coins that vesting and delegated
EndTime int64 `json:"end_time"` // when the coins become unlocked
}
```
**[ContinuousVestingAccount](https://github.com/cosmos/cosmos-sdk/blob/4b765b6ab4dfac38a896b66478b3b14b40d0c731/x/auth/account.go#L347-L353)**
Same as the `BaseVestingAccount`, but it additionally has a **start_time** for `when the coins start to vest`:
```
type ContinuousVestingAccount struct {
*BaseVestingAccount
StartTime int64 `json:"start_time"` // when the coins start to vest
}
```
AC:
- [x] display vesting accounts tokens
- [ ] display vesting account information
|
1.0
|
Vesting account and coins support - **Is your feature request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I'd like to know if any of my coins are vesting (locked) and the time they vest so I can choose those vesting coins over the regular ones to delegate.
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
- Clearly separate vesting coins from regular coins on the wallet balance
- Choose either vesting (`original_vesting`) or regular coins (`coins`) for staking (delegation, redelegation) and governance.
- If (re)delegate using my vesting coins, I want to see that the delegation was made with them and not with regular coins
- Show a message on Send action modal that the `original_vesting` coins are locked and I can't transfer them until `end_time` date
- (optional) show a time progress bar for when coins become vested (unlocked) on the Wallet Page
**Describe alternatives you've considered**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
- create a separate state for vesting balances (`original_vesting`)
- add a dropdown on action modals to choose which coin to use: regular or vesting (`original_vesting`)
**[BaseVestingAccount](https://github.com/cosmos/cosmos-sdk/blob/4b765b6ab4dfac38a896b66478b3b14b40d0c731/x/auth/account.go#L178-L191
)**
A vesting account has the following additional fields:
```
type BaseVestingAccount struct {
*BaseAccount
OriginalVesting sdk.Coins `json:"original_vesting"` // coins in account upon initialization
DelegatedFree sdk.Coins `json:"delegated_free"` // coins that are vested and delegated
DelegatedVesting sdk.Coins `json:"delegated_vesting"` // coins that vesting and delegated
EndTime int64 `json:"end_time"` // when the coins become unlocked
}
```
**[ContinuousVestingAccount](https://github.com/cosmos/cosmos-sdk/blob/4b765b6ab4dfac38a896b66478b3b14b40d0c731/x/auth/account.go#L347-L353)**
Same as the `BaseVestingAccount`, but it additionally has a **start_time** for `when the coins start to vest`:
```
type ContinuousVestingAccount struct {
*BaseVestingAccount
StartTime int64 `json:"start_time"` // when the coins start to vest
}
```
AC:
- [x] display vesting accounts tokens
- [ ] display vesting account information
|
non_defect
|
vesting account and coins support is your feature request related to a problem please describe i d like to know if any of my coins are vesting locked and the time they vest so i can choose those vesting coins over the regular ones to delegate describe the solution you d like clearly separate vesting coins from regular coins on the wallet balance choose either vesting original vesting or regular coins coins for staking delegation redelegation and governance if re delegate using my vesting coins i want to see that the delegation was made with them and not with regular coins show a message on send action modal that the original vesting coins are locked and i can t transfer them until end time date optional show a time progress bar for when coins become vested unlocked on the wallet page describe alternatives you ve considered create a separate state for vesting balances original vesting add a dropdown on action modals to choose which coin to use regular or vesting original vesting a vesting account has the following additional fields type basevestingaccount struct baseaccount originalvesting sdk coins json original vesting coins in account upon initialization delegatedfree sdk coins json delegated free coins that are vested and delegated delegatedvesting sdk coins json delegated vesting coins that vesting and delegated endtime json end time when the coins become unlocked same as the basevestingaccount but it additionally has a start time for when the coins start to vest type continuousvestingaccount struct basevestingaccount starttime json start time when the coins start to vest ac display vesting accounts tokens display vesting account information
| 0
|
24,046
| 23,267,862,158
|
IssuesEvent
|
2022-08-04 19:17:39
|
bevyengine/bevy
|
https://api.github.com/repos/bevyengine/bevy
|
opened
|
Resolve `Val::Percent` into `Val::Px` on the Node
|
A-UI C-Usability
|
## What problem does this solve or what need does it fill?
- User should be able to get the fixed value equivalents of fields defined in `Val::Percent` on the `Node`.
- This in order so that you can add, subtract, multiply or divide these values
- This is useful for building higher level features on top and to increase ergonomics when composing and designing UI.
Would help solve #5525
Sizes, borders, margins are expressed in Val:
```rs
pub enum Val {
/// No value defined
#[default]
Undefined,
/// Automatically determine this value
Auto,
/// Set this value in pixels
Px(f32),
/// Set this value in percent
Percent(f32),
}
````
## What solution would you like?
- Add a function `calc()` or `resolve()` on the Node which is responsible for resolving a percent-value to a Px value
- This Px value can then be operated on
- The function should travel up the node-hierarchy until it reaches a fixed size value, then calculate the final value based on the hierarchy
## What alternative(s) have you considered?
- Not implement this but require the user to manually figure out the sizes in question.
## Additional context
See previous implementation attempt and discussion here: https://github.com/bevyengine/bevy/pull/5555#issuecomment-1204614042
|
True
|
Resolve `Val::Percent` into `Val::Px` on the Node - ## What problem does this solve or what need does it fill?
- User should be able to get the fixed value equivalents of fields defined in `Val::Percent` on the `Node`.
- This in order so that you can add, subtract, multiply or divide these values
- This is useful for building higher level features on top and to increase ergonomics when composing and designing UI.
Would help solve #5525
Sizes, borders, margins are expressed in Val:
```rs
pub enum Val {
/// No value defined
#[default]
Undefined,
/// Automatically determine this value
Auto,
/// Set this value in pixels
Px(f32),
/// Set this value in percent
Percent(f32),
}
````
## What solution would you like?
- Add a function `calc()` or `resolve()` on the Node which is responsible for resolving a percent-value to a Px value
- This Px value can then be operated on
- The function should travel up the node-hierarchy until it reaches a fixed size value, then calculate the final value based on the hierarchy
## What alternative(s) have you considered?
- Not implement this but require the user to manually figure out the sizes in question.
## Additional context
See previous implementation attempt and discussion here: https://github.com/bevyengine/bevy/pull/5555#issuecomment-1204614042
|
non_defect
|
resolve val percent into val px on the node what problem does this solve or what need does it fill user should be able to get the fixed value equivalents of fields defined in val percent on the node this in order so that you can add subtract multiply or divide these values this is useful for building higher level features on top and to increase ergonomics when composing and designing ui would help solve sizes borders margins are expressed in val rs pub enum val no value defined undefined automatically determine this value auto set this value in pixels px set this value in percent percent what solution would you like add a function calc or resolve on the node which is responsible for resolving a percent value to a px value this px value can then be operated on the function should travel up the node hierarchy until it reaches a fixed size value then calculate the final value based on the hierarchy what alternative s have you considered not implement this but require the user to manually figure out the sizes in question additional context see previous implementation attempt and discussion here
| 0
|
224,450
| 7,470,427,186
|
IssuesEvent
|
2018-04-03 04:56:34
|
fossasia/open-event-webapp
|
https://api.github.com/repos/fossasia/open-event-webapp
|
closed
|
Implement pdf download of schedule, especially the calendar view
|
Priority: High enhancement feature has-PR
|
Implement a pdf download of the schedule similar to the orga server.
|
1.0
|
Implement pdf download of schedule, especially the calendar view - Implement a pdf download of the schedule similar to the orga server.
|
non_defect
|
implement pdf download of schedule especially the calendar view implement a pdf download of the schedule similar to the orga server
| 0
|
312,617
| 26,873,403,996
|
IssuesEvent
|
2023-02-04 18:55:30
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Informações Institucionais - Leis Municipais - Alfenas
|
generalization test development template - Betha (26) tag - Informações Institucionais subtag - Leis Municipais
|
DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Leis Municipais para o Município de Alfenas.
|
1.0
|
Teste de generalizacao para a tag Informações Institucionais - Leis Municipais - Alfenas - DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Leis Municipais para o Município de Alfenas.
|
non_defect
|
teste de generalizacao para a tag informações institucionais leis municipais alfenas dod realizar o teste de generalização do validador da tag informações institucionais leis municipais para o município de alfenas
| 0
|
164,099
| 25,919,878,275
|
IssuesEvent
|
2022-12-15 20:50:11
|
CDRH/chesnutt
|
https://api.github.com/repos/CDRH/chesnutt
|
closed
|
extra spaces between elements (e.g., <subst> children)
|
design tech_datura close?
|
It seems like there are extra spaces being inserted between `<add>`s and `<del>`s in `<subst>`. See the second sentence of the letter here: https://cdrhdev1.unl.edu/chesnutt/item/ccda.corr00081 (the substitution of "c" for "b" in "composed"). Also see here: https://cdrhdev1.unl.edu/chesnutt/item/ccda.corr00107 (the substitution of "b" for "v" in "be" towards the end of the first paragraph).
|
1.0
|
extra spaces between elements (e.g., <subst> children) - It seems like there are extra spaces being inserted between `<add>`s and `<del>`s in `<subst>`. See the second sentence of the letter here: https://cdrhdev1.unl.edu/chesnutt/item/ccda.corr00081 (the substitution of "c" for "b" in "composed"). Also see here: https://cdrhdev1.unl.edu/chesnutt/item/ccda.corr00107 (the substitution of "b" for "v" in "be" towards the end of the first paragraph).
|
non_defect
|
extra spaces between elements e g children it seems like there are extra spaces being inserted between s and s in see the second sentence of the letter here the substitution of c for b in composed also see here the substitution of b for v in be towards the end of the first paragraph
| 0
|
9,634
| 2,615,164,092
|
IssuesEvent
|
2015-03-01 06:43:56
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
help with reaver
|
auto-migrated Priority-Triage Type-Defect
|
```
after getting a key from reaver i got internet on wifi connection from somebody
and after a couple of days i lost the internet but it says im still connected
and showing activity, i tried troubleshooting it and it says DNS server not
responding plz help. ty
```
Original issue reported on code.google.com by `johnnyga...@gmail.com` on 13 Sep 2012 at 1:11
|
1.0
|
help with reaver - ```
after getting a key from reaver i got internet on wifi connection from somebody
and after a couple of days i lost the internet but it says im still connected
and showing activity, i tried troubleshooting it and it says DNS server not
responding plz help. ty
```
Original issue reported on code.google.com by `johnnyga...@gmail.com` on 13 Sep 2012 at 1:11
|
defect
|
help with reaver after getting a key from reaver i got internet on wifi connection from somebody and after a couple of days i lost the internet but it says im still connected and showing activity i tried troubleshooting it and it says dns server not responding plz help ty original issue reported on code google com by johnnyga gmail com on sep at
| 1
|
6,895
| 2,610,302,855
|
IssuesEvent
|
2015-02-26 19:37:20
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
Trophy race, invisible dead hogs block walking to finish line and other issues
|
auto-migrated Component-Lua Milestone-Release0.9.18 Priority-High Type-Defect
|
```
What steps will reproduce the problem?
1. whenever a hog dies
2. and then you try to walk over you can't
3.
What is the expected output? What do you see instead?
A invisible dead hog,I don't see the invisible hog
What version of the product are you using? On what operating system? 9.15.1 on
mac SL
Please provide any additional information below.
I was supposed to get my best time in trophy rave but a dead hog stopped me D':
```
-----
Original issue reported on code.google.com by `davidb...@gmail.com` on 7 Jul 2011 at 11:55
|
1.0
|
Trophy race, invisible dead hogs block walking to finish line and other issues - ```
What steps will reproduce the problem?
1. whenever a hog dies
2. and then you try to walk over you can't
3.
What is the expected output? What do you see instead?
A invisible dead hog,I don't see the invisible hog
What version of the product are you using? On what operating system? 9.15.1 on
mac SL
Please provide any additional information below.
I was supposed to get my best time in trophy rave but a dead hog stopped me D':
```
-----
Original issue reported on code.google.com by `davidb...@gmail.com` on 7 Jul 2011 at 11:55
|
defect
|
trophy race invisible dead hogs block walking to finish line and other issues what steps will reproduce the problem whenever a hog dies and then you try to walk over you can t what is the expected output what do you see instead a invisible dead hog i don t see the invisible hog what version of the product are you using on what operating system on mac sl please provide any additional information below i was supposed to get my best time in trophy rave but a dead hog stopped me d original issue reported on code google com by davidb gmail com on jul at
| 1
|
297,818
| 22,394,505,628
|
IssuesEvent
|
2022-06-17 11:03:38
|
spring-projects/spring-framework
|
https://api.github.com/repos/spring-projects/spring-framework
|
closed
|
Fix a typo in ResponseEntity documentation
|
type: documentation
|
Change `return ResponseEntity.ok().eTag(etag).build(body);` to `return ResponseEntity.ok().eTag(etag).body(body);` for both WebMVC and WebFlux.
|
1.0
|
Fix a typo in ResponseEntity documentation - Change `return ResponseEntity.ok().eTag(etag).build(body);` to `return ResponseEntity.ok().eTag(etag).body(body);` for both WebMVC and WebFlux.
|
non_defect
|
fix a typo in responseentity documentation change return responseentity ok etag etag build body to return responseentity ok etag etag body body for both webmvc and webflux
| 0
|
11,984
| 2,673,676,989
|
IssuesEvent
|
2015-03-24 20:36:18
|
vim/vim
|
https://api.github.com/repos/vim/vim
|
closed
|
Diff mode gets stuck
|
auto-migrated Priority-Medium Type-Defect
|
```
Unfortunately I haven't discovered an easy way to reproduce this.
1. open two similar files (call them A and B) in split windows
2. windo diffthis
3. navigate away from B using bufexplorer, netrw and other plugins, opening and
closing windows etc.
4. switch back to buffer B
It sometimes happens that in step 4 I get to see buffer B without diff
highlighting. At that point I can switch to the window displaying buffer A,
:diffoff, close all other windows with :only, attempt a big hamer like :bufdo
diffoff, and yet if I'll open an unrelated file C and try to :diffthis, vim
will begin diffing it with B.
The only way to make vim forget it was ever diffing with B that I discovered is
to open a new tab page and close the old tab page.
I'm using vim 7.4.617 on Linux.
```
Original issue reported on code.google.com by `mged...@gmail.com` on 5 Mar 2015 at 11:14
|
1.0
|
Diff mode gets stuck - ```
Unfortunately I haven't discovered an easy way to reproduce this.
1. open two similar files (call them A and B) in split windows
2. windo diffthis
3. navigate away from B using bufexplorer, netrw and other plugins, opening and
closing windows etc.
4. switch back to buffer B
It sometimes happens that in step 4 I get to see buffer B without diff
highlighting. At that point I can switch to the window displaying buffer A,
:diffoff, close all other windows with :only, attempt a big hamer like :bufdo
diffoff, and yet if I'll open an unrelated file C and try to :diffthis, vim
will begin diffing it with B.
The only way to make vim forget it was ever diffing with B that I discovered is
to open a new tab page and close the old tab page.
I'm using vim 7.4.617 on Linux.
```
Original issue reported on code.google.com by `mged...@gmail.com` on 5 Mar 2015 at 11:14
|
defect
|
diff mode gets stuck unfortunately i haven t discovered an easy way to reproduce this open two similar files call them a and b in split windows windo diffthis navigate away from b using bufexplorer netrw and other plugins opening and closing windows etc switch back to buffer b it sometimes happens that in step i get to see buffer b without diff highlighting at that point i can switch to the window displaying buffer a diffoff close all other windows with only attempt a big hamer like bufdo diffoff and yet if i ll open an unrelated file c and try to diffthis vim will begin diffing it with b the only way to make vim forget it was ever diffing with b that i discovered is to open a new tab page and close the old tab page i m using vim on linux original issue reported on code google com by mged gmail com on mar at
| 1
|
6,211
| 2,610,223,324
|
IssuesEvent
|
2015-02-26 19:10:49
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
gsroot dll
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Вольфрам Бобров'''
Привет всем не подскажите где можно найти
.gsroot dll. где то видел уже
'''Анри Блохин'''
Качай тут http://bit.ly/19gjK7C
'''Герман Колесников'''
Спасибо вроде то но просит телефон вводить
'''Боян Туров'''
Неа все ок у меня ничего не списало
'''Адольф Мишин'''
Не это не влияет на баланс
Информация о файле: gsroot dll
Загружен: В этом месяце
Скачан раз: 702
Рейтинг: 456
Средняя скорость скачивания: 914
Похожих файлов: 28
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 6:36
|
1.0
|
gsroot dll - ```
'''Вольфрам Бобров'''
Привет всем не подскажите где можно найти
.gsroot dll. где то видел уже
'''Анри Блохин'''
Качай тут http://bit.ly/19gjK7C
'''Герман Колесников'''
Спасибо вроде то но просит телефон вводить
'''Боян Туров'''
Неа все ок у меня ничего не списало
'''Адольф Мишин'''
Не это не влияет на баланс
Информация о файле: gsroot dll
Загружен: В этом месяце
Скачан раз: 702
Рейтинг: 456
Средняя скорость скачивания: 914
Похожих файлов: 28
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 6:36
|
defect
|
gsroot dll вольфрам бобров привет всем не подскажите где можно найти gsroot dll где то видел уже анри блохин качай тут герман колесников спасибо вроде то но просит телефон вводить боян туров неа все ок у меня ничего не списало адольф мишин не это не влияет на баланс информация о файле gsroot dll загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
4,776
| 2,610,155,702
|
IssuesEvent
|
2015-02-26 18:49:29
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Text
|
auto-migrated Priority-Medium Type-Defect
|
```
Fix rep reseach struc text
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:23
|
1.0
|
Text - ```
Fix rep reseach struc text
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:23
|
defect
|
text fix rep reseach struc text original issue reported on code google com by gmail com on jan at
| 1
|
401,837
| 27,339,928,687
|
IssuesEvent
|
2023-02-26 17:19:55
|
stitchfix/hamilton
|
https://api.github.com/repos/stitchfix/hamilton
|
closed
|
[documentation] show how to get hamilton running on snowpark
|
documentation
|
We need to help people get up and running with Hamilton on snowpark.
Two artifacts to produce:
* hamilton + dbt + snowpark
* hamilton + snowpark
|
1.0
|
[documentation] show how to get hamilton running on snowpark - We need to help people get up and running with Hamilton on snowpark.
Two artifacts to produce:
* hamilton + dbt + snowpark
* hamilton + snowpark
|
non_defect
|
show how to get hamilton running on snowpark we need to help people get up and running with hamilton on snowpark two artifacts to produce hamilton dbt snowpark hamilton snowpark
| 0
|
41,389
| 12,832,007,487
|
IssuesEvent
|
2020-07-07 06:49:09
|
rvvergara/bookstore-cms-next-app
|
https://api.github.com/repos/rvvergara/bookstore-cms-next-app
|
closed
|
CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js
|
security vulnerability
|
## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/bookstore-cms-next-app/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: /bookstore-cms-next-app/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-cms-next-app/commit/6bc53307a3990ef980f925c145f97abe8665de0b">6bc53307a3990ef980f925c145f97abe8665de0b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/bookstore-cms-next-app/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: /bookstore-cms-next-app/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-cms-next-app/commit/6bc53307a3990ef980f925c145f97abe8665de0b">6bc53307a3990ef980f925c145f97abe8665de0b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm bookstore cms next app node modules js attic test moment index html path to vulnerable library bookstore cms next app node modules js attic test moment index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
220,786
| 24,580,488,692
|
IssuesEvent
|
2022-10-13 15:16:06
|
gardener/gardener
|
https://api.github.com/repos/gardener/gardener
|
closed
|
Use only one IDP for monitoring and logging
|
kind/enhancement area/security lifecycle/rotten platform/all area/monitoring area/logging component/gardener topology/seed
|
# Use one IDP for monitoring and logging
## Problem
At the moment operators can access logging and monitoring components which are exposed via Ingress and Basic Authentication:
```console
+------+
| USER |
+--+---+
|
| Basic Auth
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
|
|
v
+-----+------+
| Prometheus |
+------------+
```
Basic Authentication is plagued by has many issues, not limited to, password-sharing, need for revocation when user is removed from the `Project`, subject to brute force attacks, impossible to enforce Multi-factor authentication and more.
A more elegant solution can be deployed, using Authentication and Authorization proxies which validate requests agains the Gardener IDP (dex in most cases) and the Gardener K8S API Server.
### Goals
- Have only one identity authentication source for monitoring and logging components.
- Remove basic authentication for those components.
- Enforce authorization via `SubjectAccessReview` against the Garden cluster.
### Non-goals
- Network connectivity Garden cluster -> Seed cluster components is out of scope.
## Proposed solution(s):
### Authentication
End-users accessing the dashboard authenticate via OAuth and the Dashboard is connected to OpenID Connect provider such as Dex.
We want to use the same mechanism to authenticate users. Because Prometheus and other components don't support OIDC configurations we are forced to use proxies.
There are several implementations for this and [OAuth Proxy](https://github.com/pusher/oauth2_proxy) can be used as an example.
The flow goes like this (with some steps skipped for simpicity):
1. User enters `https://prometheus.myshoot.seed-one.gardener.cloud` in browser.
1. The request goes via Nginx to the the `OAuth Proxy`.
1. Because there is no session cookie, `OAuth Proxy` redirects user to Dex at `https://dex.gardener.cloud`.
1. User authenticates succesfuly and its redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback` where the access token is exchanged for ID token.
1. Session cookie is set in user's browser, which expires after some time.
1. User is redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud`
1. `OAuth Proxy` proxies the connection to `Prometheus` container. `Prometheus` only listens on `localhost`.
```console
+------+
| USER |
+--+---+
|
|
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
+----------------------------+
1. Redirect Browser | | Prometheus Pod |
+-----------+ + Authenticate | v |
| | | +----+--------+ |
| +<--------------------------+ OAuth Proxy | |
| Dex | | | | |
| +-------------------------->+ | |
| | | +------+------+ |
+-----------+ 2. Receive ID token | | |
| | 3. Proxy |
| v |
| +------+-----+ |
| | Prometheus | |
| +------------+ |
| |
+----------------------------+
```
In order for this flow to work two configurations are needed.
- New OAuth client must be added in Dex (or any OIDC provider) with redirect URL pointing to the oauth2 callback endpoint of `OAuth Proxy` - `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback`
- `client_id` and `client_secret` from said client must be configured on `OAuth Proxy` and `OIDC issuer URL` (`https://dex.gardener.cloud`) must be added.
> Note: As the authentication step can be done the same way for all Shoot clusters in a Seed, the OAuthProxy component can be deployed as a stand-alone deployment.
### Authorization
Now that we have successfuly authenticated our user, authorization must be executed. The ID token from Dex only provides us with some basic identity information:
```json
{
"iss": "https://dex.gardener.cloud",
"sub": "CiQwOGE4Njg0Yi1kYjg4LTRiNzMtOTBhOS0zY2QxNjYxZjU0NjYSBWxvY2Fs",
"aud": "seed-one-oauth",
"exp": 1557571686,
"iat": 1557485286,
"at_hash": "rFl80YaYUdGkinXtW_Cq7A",
"email": "user@example.com",
"email_verified": true,
"name": "user"
}
```
as the source of truth for Authorization is the K8S APIServer, then a `SelfSubjectAccessReview` call [must be make](https://kubernetes.io/docs/reference/access-authn-authz/authorization/#checking-api-access) like:
```console
kubectl create -f - -o yaml << EOF
apiVersion: authorization.k8s.io/v1
kind: SelfSubjectAccessReview
spec:
resourceAttributes:
group: garden.sapcloud.io
resource: shoots
verb: update
namespace: my-project-namespace
resourceName: my-shoot
EOF
```
There is already a project which does that - [RBAC Proxy](https://github.com/brancz/kube-rbac-proxy).
With it installed and configured to talk against the K8S API server in the Gardener cluster the flow is:
1. User enters `https://prometheus.myshoot.seed-one.gardener.cloud` in browser.
1. The request goes via Nginx to the the `OAuth Proxy`.
1. Because there is no session cookie, `OAuth Proxy` redirects user to Dex at `https://dex.gardener.cloud`.
1. User authenticates succesfuly and its redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback` where the access token is exchanged for ID token.
1. Session cookie is set in user's browser, which expires after some time.
1. User is redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud`
1. `OAuth Proxy` proxies the connection to `RBAC Proxy` (which listens only on `localhost`) container and sends the ID token via `Authorization: Bearer` Header.
1. `RBAC proxy` receives the request and does a `SubjectAccessReview` request against the K8S API in the Gardener cluster which succeeds.
1. `RBAC proxy` then caches the authz descision and passes the connection to `Prometheus`.
```console
+------+
| USER |
+--+---+
|
|
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
+----------------------------+
1. Redirect Browser | | Prometheus Pod |
+-----------+ + Authenticate | v |
| | | +----+--------+ |
| +<--------------------------+ OAuth Proxy | |
| Dex | | | | |
| +-------------------------->+ | |
| | | +------+------+ |
+-----+-----+ 2. Receive ID token | | |
^ | | 3. Proxy |
| | | + ID token |
| | | |
| | v |
+-----+-----+ | +------+-------+ |
Uses Dex as | |4. SubjectAccessReview | | | |
OIDC provider | K8S API | | | RBAC Proxy | |
| Server +<--------------------------+ | |
| | | +------+-------+ |
| | | | |
+-----------+ | | 5. Proxy |
| | |
| v |
| +------+-----+ |
| | Prometheus | |
| +------------+ |
| |
+----------------------------+
```
## Alternatives
Alternative solutions for Authorization might include using Istio's [ServiceRole and RoleBinding](https://istio.io/docs/tasks/security/rbac-groups/), but this requires to have a controller which reconsiles K8S `ClusterRole` + `Role` to istio's `ServiceRole` and `ClusterRoleBinding` + `RoleBinding` to istio's `ServiceRoleBinding`.
|
True
|
Use only one IDP for monitoring and logging - # Use one IDP for monitoring and logging
## Problem
At the moment operators can access logging and monitoring components which are exposed via Ingress and Basic Authentication:
```console
+------+
| USER |
+--+---+
|
| Basic Auth
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
|
|
v
+-----+------+
| Prometheus |
+------------+
```
Basic Authentication is plagued by has many issues, not limited to, password-sharing, need for revocation when user is removed from the `Project`, subject to brute force attacks, impossible to enforce Multi-factor authentication and more.
A more elegant solution can be deployed, using Authentication and Authorization proxies which validate requests agains the Gardener IDP (dex in most cases) and the Gardener K8S API Server.
### Goals
- Have only one identity authentication source for monitoring and logging components.
- Remove basic authentication for those components.
- Enforce authorization via `SubjectAccessReview` against the Garden cluster.
### Non-goals
- Network connectivity Garden cluster -> Seed cluster components is out of scope.
## Proposed solution(s):
### Authentication
End-users accessing the dashboard authenticate via OAuth and the Dashboard is connected to OpenID Connect provider such as Dex.
We want to use the same mechanism to authenticate users. Because Prometheus and other components don't support OIDC configurations we are forced to use proxies.
There are several implementations for this and [OAuth Proxy](https://github.com/pusher/oauth2_proxy) can be used as an example.
The flow goes like this (with some steps skipped for simpicity):
1. User enters `https://prometheus.myshoot.seed-one.gardener.cloud` in browser.
1. The request goes via Nginx to the the `OAuth Proxy`.
1. Because there is no session cookie, `OAuth Proxy` redirects user to Dex at `https://dex.gardener.cloud`.
1. User authenticates succesfuly and its redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback` where the access token is exchanged for ID token.
1. Session cookie is set in user's browser, which expires after some time.
1. User is redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud`
1. `OAuth Proxy` proxies the connection to `Prometheus` container. `Prometheus` only listens on `localhost`.
```console
+------+
| USER |
+--+---+
|
|
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
+----------------------------+
1. Redirect Browser | | Prometheus Pod |
+-----------+ + Authenticate | v |
| | | +----+--------+ |
| +<--------------------------+ OAuth Proxy | |
| Dex | | | | |
| +-------------------------->+ | |
| | | +------+------+ |
+-----------+ 2. Receive ID token | | |
| | 3. Proxy |
| v |
| +------+-----+ |
| | Prometheus | |
| +------------+ |
| |
+----------------------------+
```
In order for this flow to work two configurations are needed.
- New OAuth client must be added in Dex (or any OIDC provider) with redirect URL pointing to the oauth2 callback endpoint of `OAuth Proxy` - `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback`
- `client_id` and `client_secret` from said client must be configured on `OAuth Proxy` and `OIDC issuer URL` (`https://dex.gardener.cloud`) must be added.
> Note: As the authentication step can be done the same way for all Shoot clusters in a Seed, the OAuthProxy component can be deployed as a stand-alone deployment.
### Authorization
Now that we have successfuly authenticated our user, authorization must be executed. The ID token from Dex only provides us with some basic identity information:
```json
{
"iss": "https://dex.gardener.cloud",
"sub": "CiQwOGE4Njg0Yi1kYjg4LTRiNzMtOTBhOS0zY2QxNjYxZjU0NjYSBWxvY2Fs",
"aud": "seed-one-oauth",
"exp": 1557571686,
"iat": 1557485286,
"at_hash": "rFl80YaYUdGkinXtW_Cq7A",
"email": "user@example.com",
"email_verified": true,
"name": "user"
}
```
as the source of truth for Authorization is the K8S APIServer, then a `SelfSubjectAccessReview` call [must be make](https://kubernetes.io/docs/reference/access-authn-authz/authorization/#checking-api-access) like:
```console
kubectl create -f - -o yaml << EOF
apiVersion: authorization.k8s.io/v1
kind: SelfSubjectAccessReview
spec:
resourceAttributes:
group: garden.sapcloud.io
resource: shoots
verb: update
namespace: my-project-namespace
resourceName: my-shoot
EOF
```
There is already a project which does that - [RBAC Proxy](https://github.com/brancz/kube-rbac-proxy).
With it installed and configured to talk against the K8S API server in the Gardener cluster the flow is:
1. User enters `https://prometheus.myshoot.seed-one.gardener.cloud` in browser.
1. The request goes via Nginx to the the `OAuth Proxy`.
1. Because there is no session cookie, `OAuth Proxy` redirects user to Dex at `https://dex.gardener.cloud`.
1. User authenticates succesfuly and its redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud/oauth2/callback` where the access token is exchanged for ID token.
1. Session cookie is set in user's browser, which expires after some time.
1. User is redirected back to `https://prometheus.myshoot.seed-one.gardener.cloud`
1. `OAuth Proxy` proxies the connection to `RBAC Proxy` (which listens only on `localhost`) container and sends the ID token via `Authorization: Bearer` Header.
1. `RBAC proxy` receives the request and does a `SubjectAccessReview` request against the K8S API in the Gardener cluster which succeeds.
1. `RBAC proxy` then caches the authz descision and passes the connection to `Prometheus`.
```console
+------+
| USER |
+--+---+
|
|
|
v
+----+----+
| Nginx |
| Ingress |
+----+----+
|
+----------------------------+
1. Redirect Browser | | Prometheus Pod |
+-----------+ + Authenticate | v |
| | | +----+--------+ |
| +<--------------------------+ OAuth Proxy | |
| Dex | | | | |
| +-------------------------->+ | |
| | | +------+------+ |
+-----+-----+ 2. Receive ID token | | |
^ | | 3. Proxy |
| | | + ID token |
| | | |
| | v |
+-----+-----+ | +------+-------+ |
Uses Dex as | |4. SubjectAccessReview | | | |
OIDC provider | K8S API | | | RBAC Proxy | |
| Server +<--------------------------+ | |
| | | +------+-------+ |
| | | | |
+-----------+ | | 5. Proxy |
| | |
| v |
| +------+-----+ |
| | Prometheus | |
| +------------+ |
| |
+----------------------------+
```
## Alternatives
Alternative solutions for Authorization might include using Istio's [ServiceRole and RoleBinding](https://istio.io/docs/tasks/security/rbac-groups/), but this requires to have a controller which reconsiles K8S `ClusterRole` + `Role` to istio's `ServiceRole` and `ClusterRoleBinding` + `RoleBinding` to istio's `ServiceRoleBinding`.
|
non_defect
|
use only one idp for monitoring and logging use one idp for monitoring and logging problem at the moment operators can access logging and monitoring components which are exposed via ingress and basic authentication console user basic auth v nginx ingress v prometheus basic authentication is plagued by has many issues not limited to password sharing need for revocation when user is removed from the project subject to brute force attacks impossible to enforce multi factor authentication and more a more elegant solution can be deployed using authentication and authorization proxies which validate requests agains the gardener idp dex in most cases and the gardener api server goals have only one identity authentication source for monitoring and logging components remove basic authentication for those components enforce authorization via subjectaccessreview against the garden cluster non goals network connectivity garden cluster seed cluster components is out of scope proposed solution s authentication end users accessing the dashboard authenticate via oauth and the dashboard is connected to openid connect provider such as dex we want to use the same mechanism to authenticate users because prometheus and other components don t support oidc configurations we are forced to use proxies there are several implementations for this and can be used as an example the flow goes like this with some steps skipped for simpicity user enters in browser the request goes via nginx to the the oauth proxy because there is no session cookie oauth proxy redirects user to dex at user authenticates succesfuly and its redirected back to where the access token is exchanged for id token session cookie is set in user s browser which expires after some time user is redirected back to oauth proxy proxies the connection to prometheus container prometheus only listens on localhost console user v nginx ingress redirect browser prometheus pod authenticate v oauth proxy dex receive id token proxy v prometheus in order for this flow to work two configurations are needed new oauth client must be added in dex or any oidc provider with redirect url pointing to the callback endpoint of oauth proxy client id and client secret from said client must be configured on oauth proxy and oidc issuer url must be added note as the authentication step can be done the same way for all shoot clusters in a seed the oauthproxy component can be deployed as a stand alone deployment authorization now that we have successfuly authenticated our user authorization must be executed the id token from dex only provides us with some basic identity information json iss sub aud seed one oauth exp iat at hash email user example com email verified true name user as the source of truth for authorization is the apiserver then a selfsubjectaccessreview call like console kubectl create f o yaml eof apiversion authorization io kind selfsubjectaccessreview spec resourceattributes group garden sapcloud io resource shoots verb update namespace my project namespace resourcename my shoot eof there is already a project which does that with it installed and configured to talk against the api server in the gardener cluster the flow is user enters in browser the request goes via nginx to the the oauth proxy because there is no session cookie oauth proxy redirects user to dex at user authenticates succesfuly and its redirected back to where the access token is exchanged for id token session cookie is set in user s browser which expires after some time user is redirected back to oauth proxy proxies the connection to rbac proxy which listens only on localhost container and sends the id token via authorization bearer header rbac proxy receives the request and does a subjectaccessreview request against the api in the gardener cluster which succeeds rbac proxy then caches the authz descision and passes the connection to prometheus console user v nginx ingress redirect browser prometheus pod authenticate v oauth proxy dex receive id token proxy id token v uses dex as subjectaccessreview oidc provider api rbac proxy server proxy v prometheus alternatives alternative solutions for authorization might include using istio s but this requires to have a controller which reconsiles clusterrole role to istio s servicerole and clusterrolebinding rolebinding to istio s servicerolebinding
| 0
|
24,135
| 3,917,074,268
|
IssuesEvent
|
2016-04-21 06:25:09
|
irnawansuprapti/openbiz-cubi
|
https://api.github.com/repos/irnawansuprapti/openbiz-cubi
|
closed
|
How To Turn Oily Skin Into Healthy Skin
|
auto-migrated Priority-Medium spam Type-Defect
|
```
Since the remedy of injure work creams is easier, they seem to be desirable (over lotions) in cases where the rind fear set is not to be removed directly after cure. So, pare fear creams seem much nonclassical as moisturisers than as cleansers or toners. For toners, lotions seem to be desirable over wound reparation creams. There are few pare mind creams that acts as toners too, but generally the toners are obtainable in disposable cast legendary to be most competent in keeping injure moist; thus, the most touristed shape of injure tutelage creams is moisturisers. Due to the comparable sanity, a lot of group incline to assort strip charge creams with dry and sensitive peel. Tho' it is faithful to a reliable extent, pare attention creams are not used only for dry tegument, they are also victimized for making products for sebaceous peel e.g. vitamin http://www.skinphysiciantips.com/instant-wrinkle-repair/
```
Original issue reported on code.google.com by `CleoHo...@gmail.com` on 15 Apr 2015 at 7:58
|
1.0
|
How To Turn Oily Skin Into Healthy Skin - ```
Since the remedy of injure work creams is easier, they seem to be desirable (over lotions) in cases where the rind fear set is not to be removed directly after cure. So, pare fear creams seem much nonclassical as moisturisers than as cleansers or toners. For toners, lotions seem to be desirable over wound reparation creams. There are few pare mind creams that acts as toners too, but generally the toners are obtainable in disposable cast legendary to be most competent in keeping injure moist; thus, the most touristed shape of injure tutelage creams is moisturisers. Due to the comparable sanity, a lot of group incline to assort strip charge creams with dry and sensitive peel. Tho' it is faithful to a reliable extent, pare attention creams are not used only for dry tegument, they are also victimized for making products for sebaceous peel e.g. vitamin http://www.skinphysiciantips.com/instant-wrinkle-repair/
```
Original issue reported on code.google.com by `CleoHo...@gmail.com` on 15 Apr 2015 at 7:58
|
defect
|
how to turn oily skin into healthy skin since the remedy of injure work creams is easier they seem to be desirable over lotions in cases where the rind fear set is not to be removed directly after cure so pare fear creams seem much nonclassical as moisturisers than as cleansers or toners for toners lotions seem to be desirable over wound reparation creams there are few pare mind creams that acts as toners too but generally the toners are obtainable in disposable cast legendary to be most competent in keeping injure moist thus the most touristed shape of injure tutelage creams is moisturisers due to the comparable sanity a lot of group incline to assort strip charge creams with dry and sensitive peel tho it is faithful to a reliable extent pare attention creams are not used only for dry tegument they are also victimized for making products for sebaceous peel e g vitamin original issue reported on code google com by cleoho gmail com on apr at
| 1
|
16,404
| 2,891,898,951
|
IssuesEvent
|
2015-06-15 09:23:51
|
LatinSuD/efax4asterisk
|
https://api.github.com/repos/LatinSuD/efax4asterisk
|
closed
|
Google code closing. Please migrate
|
auto-migrated Priority-Medium Type-Defect
|
```
Please migrate to github or similar.
```
Original issue reported on code.google.com by `LatinSuD@gmail.com` on 14 Jun 2015 at 10:51
|
1.0
|
Google code closing. Please migrate - ```
Please migrate to github or similar.
```
Original issue reported on code.google.com by `LatinSuD@gmail.com` on 14 Jun 2015 at 10:51
|
defect
|
google code closing please migrate please migrate to github or similar original issue reported on code google com by latinsud gmail com on jun at
| 1
|
338,016
| 24,566,189,837
|
IssuesEvent
|
2022-10-13 03:23:58
|
phpredis/phpredis
|
https://api.github.com/repos/phpredis/phpredis
|
closed
|
Return type incorrect if Redis fails
|
documentation problem not a bug
|
e.g. zRange, mGet,... have param @return array
e.g.
```
$redis->zRange( $queue, 0, -1, true );
```
However they do not return array in all cases. e.g. if redis has some internal issues (e.g. LOADING redis into memory, ...) these functions do not return an array.
Please fix the @return type so it also correctly shows this
|
1.0
|
Return type incorrect if Redis fails -
e.g. zRange, mGet,... have param @return array
e.g.
```
$redis->zRange( $queue, 0, -1, true );
```
However they do not return array in all cases. e.g. if redis has some internal issues (e.g. LOADING redis into memory, ...) these functions do not return an array.
Please fix the @return type so it also correctly shows this
|
non_defect
|
return type incorrect if redis fails e g zrange mget have param return array e g redis zrange queue true however they do not return array in all cases e g if redis has some internal issues e g loading redis into memory these functions do not return an array please fix the return type so it also correctly shows this
| 0
|
19,107
| 13,187,614,830
|
IssuesEvent
|
2020-08-13 04:00:05
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
pybindings - don't auto-build dependencies (╯°□°)╯︵ ┻━┻ (Trac #1032)
|
Migrated from Trac enhancement infrastructure
|
this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1032">https://code.icecube.wisc.edu/ticket/1032</a>, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:19",
"description": "this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547251279109761",
"component": "infrastructure",
"summary": "pybindings - don't auto-build dependencies (\u256f\u00b0\u25a1\u00b0)\u256f\ufe35 \u253b\u2501\u253b",
"priority": "normal",
"keywords": "",
"time": "2015-06-25T19:00:21",
"milestone": "",
"owner": "nega",
"type": "enhancement"
}
```
</p>
</details>
|
1.0
|
pybindings - don't auto-build dependencies (╯°□°)╯︵ ┻━┻ (Trac #1032) - this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1032">https://code.icecube.wisc.edu/ticket/1032</a>, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:19",
"description": "this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547251279109761",
"component": "infrastructure",
"summary": "pybindings - don't auto-build dependencies (\u256f\u00b0\u25a1\u00b0)\u256f\ufe35 \u253b\u2501\u253b",
"priority": "normal",
"keywords": "",
"time": "2015-06-25T19:00:21",
"milestone": "",
"owner": "nega",
"type": "enhancement"
}
```
</p>
</details>
|
non_defect
|
pybindings don t auto build dependencies ╯°□° ╯︵ ┻━┻ trac this is fine with make or make all but when trying to do a partial incremental build you have to keep failing python scripts to see what pybindings you need without looking into cmakelists txt migrated from json status closed changetime description this is fine with make or make all but when trying to do a partial incremental build you have to keep failing python scripts to see what pybindings you need without looking into cmakelists txt reporter nega cc resolution fixed ts component infrastructure summary pybindings don t auto build dependencies priority normal keywords time milestone owner nega type enhancement
| 0
|
73,553
| 7,344,169,827
|
IssuesEvent
|
2018-03-07 13:54:41
|
assistify/Rocket.Chat
|
https://api.github.com/repos/assistify/Rocket.Chat
|
closed
|
Run Tests separately
|
Type: New Feature Type: Test progress:working
|
### Description:
I want to run the Rocket.Chat Tests separately to make it more stable.
### Expectations
A script which is copying the test files step by step into a temporary folder and only executes one by one.
Include this in the travis build pipeline as well
|
1.0
|
Run Tests separately - ### Description:
I want to run the Rocket.Chat Tests separately to make it more stable.
### Expectations
A script which is copying the test files step by step into a temporary folder and only executes one by one.
Include this in the travis build pipeline as well
|
non_defect
|
run tests separately description i want to run the rocket chat tests separately to make it more stable expectations a script which is copying the test files step by step into a temporary folder and only executes one by one include this in the travis build pipeline as well
| 0
|
140,216
| 5,398,689,587
|
IssuesEvent
|
2017-02-27 17:31:58
|
salesagility/SuiteCRM
|
https://api.github.com/repos/salesagility/SuiteCRM
|
closed
|
Update PHP-SAML library to 2.10.0 to fix login security vulnerability
|
bug Fix Proposed High Priority Resolved: Next Release
|
<!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
The SAML integration is based on a very old 2010 version of the Onelogin's SAML toolkit, which was found to be vulnerable to a signature-wrapping attack.
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
SAML is used to provide single sign on to web applications.
#### Expected Behavior
<!--- Tell us what should happen -->
Update the SAML library to prevent login attacks.
#### Actual Behavior
<!--- Tell us what happens instead -->
The signature wrapping attack allows an attacker to login to a user account by signing the request multiple times.
#### Possible Fix
<!--- Not obligatory, but suggest a fix or reason for the bug -->
* The PHP-SAML library should be updated and use the latest version 2.10.0 or newer, which has fixed the security issue:
https://github.com/onelogin/php-saml
* The PHP-SAML toolkit also supports composer. You can find the `onelogin/php-saml` package at https://packagist.org/packages/onelogin/php-saml
In order to import the latest version of the saml toolkit, run:
`composer require onelogin/php-saml`
After installation has completed you'll find inside the `vendor/` folder a new folder named `onelogin` and inside, `php-saml`. Make sure you are including the autoloader provided by composer. It can be found at `vendor/autoload.php`.
Important: When using composer, the x509 certs must be stored at `vendor/onelogin/php-saml/certs` and settings file stored at `vendor/onelogin/php-saml`.
Your settings are at risk of being deleted when updating packages using `composer update` or similar commands. **So it is highly recommended that instead of using settings files, you pass the settings as an array directly to the constructor.** If you do not use this approach your settings are at risk of being deleted when updating packages using `composer update` or similar commands.
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. See PHP-SAML github page for more details on the signature wrapping hack and their fix. https://github.com/onelogin/php-saml
> Update php-saml to 2.10.0 or newer, this version includes code to perform extra verification and prevent signature wrapping login attacks.
> php-saml < v2.10.0 is vulnerable and allows signature wrapping!
#### Context
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- If you feel this should be a low/medium/high priority then please state so -->
Medium-High priority. It's a serious security issue. SAML is such a popular method of login because it's fast, protects passwords, reduces password proliferation. The PHP application SuiteCRM doesn't hold any user password data, only the identity provider, such as microsoft, facebook, twitter, google, holds the user's password data. The SuiteCRM PHP application is much harder to hack when users are logging in with SAML, because the hacker would need to attack the security of the Identity Provider (social network typically) yet social networks typically run high security, are constantly audited by world class security teams, and send highly deliverable email and push notifications, to the user, of any suspicious login activity coming from new devices and new locations.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used:
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Mozilla Firefox 51
* Environment name and version (e.g. MySQL, PHP 7): MySQL 5.7, PHP 7.
* Operating System and version (e.g Ubuntu 16.04): Debian 8.
|
1.0
|
Update PHP-SAML library to 2.10.0 to fix login security vulnerability - <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
The SAML integration is based on a very old 2010 version of the Onelogin's SAML toolkit, which was found to be vulnerable to a signature-wrapping attack.
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
SAML is used to provide single sign on to web applications.
#### Expected Behavior
<!--- Tell us what should happen -->
Update the SAML library to prevent login attacks.
#### Actual Behavior
<!--- Tell us what happens instead -->
The signature wrapping attack allows an attacker to login to a user account by signing the request multiple times.
#### Possible Fix
<!--- Not obligatory, but suggest a fix or reason for the bug -->
* The PHP-SAML library should be updated and use the latest version 2.10.0 or newer, which has fixed the security issue:
https://github.com/onelogin/php-saml
* The PHP-SAML toolkit also supports composer. You can find the `onelogin/php-saml` package at https://packagist.org/packages/onelogin/php-saml
In order to import the latest version of the saml toolkit, run:
`composer require onelogin/php-saml`
After installation has completed you'll find inside the `vendor/` folder a new folder named `onelogin` and inside, `php-saml`. Make sure you are including the autoloader provided by composer. It can be found at `vendor/autoload.php`.
Important: When using composer, the x509 certs must be stored at `vendor/onelogin/php-saml/certs` and settings file stored at `vendor/onelogin/php-saml`.
Your settings are at risk of being deleted when updating packages using `composer update` or similar commands. **So it is highly recommended that instead of using settings files, you pass the settings as an array directly to the constructor.** If you do not use this approach your settings are at risk of being deleted when updating packages using `composer update` or similar commands.
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. See PHP-SAML github page for more details on the signature wrapping hack and their fix. https://github.com/onelogin/php-saml
> Update php-saml to 2.10.0 or newer, this version includes code to perform extra verification and prevent signature wrapping login attacks.
> php-saml < v2.10.0 is vulnerable and allows signature wrapping!
#### Context
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- If you feel this should be a low/medium/high priority then please state so -->
Medium-High priority. It's a serious security issue. SAML is such a popular method of login because it's fast, protects passwords, reduces password proliferation. The PHP application SuiteCRM doesn't hold any user password data, only the identity provider, such as microsoft, facebook, twitter, google, holds the user's password data. The SuiteCRM PHP application is much harder to hack when users are logging in with SAML, because the hacker would need to attack the security of the Identity Provider (social network typically) yet social networks typically run high security, are constantly audited by world class security teams, and send highly deliverable email and push notifications, to the user, of any suspicious login activity coming from new devices and new locations.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used:
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Mozilla Firefox 51
* Environment name and version (e.g. MySQL, PHP 7): MySQL 5.7, PHP 7.
* Operating System and version (e.g Ubuntu 16.04): Debian 8.
|
non_defect
|
update php saml library to to fix login security vulnerability the saml integration is based on a very old version of the onelogin s saml toolkit which was found to be vulnerable to a signature wrapping attack issue saml is used to provide single sign on to web applications expected behavior update the saml library to prevent login attacks actual behavior the signature wrapping attack allows an attacker to login to a user account by signing the request multiple times possible fix the php saml library should be updated and use the latest version or newer which has fixed the security issue the php saml toolkit also supports composer you can find the onelogin php saml package at in order to import the latest version of the saml toolkit run composer require onelogin php saml after installation has completed you ll find inside the vendor folder a new folder named onelogin and inside php saml make sure you are including the autoloader provided by composer it can be found at vendor autoload php important when using composer the certs must be stored at vendor onelogin php saml certs and settings file stored at vendor onelogin php saml your settings are at risk of being deleted when updating packages using composer update or similar commands so it is highly recommended that instead of using settings files you pass the settings as an array directly to the constructor if you do not use this approach your settings are at risk of being deleted when updating packages using composer update or similar commands steps to reproduce see php saml github page for more details on the signature wrapping hack and their fix update php saml to or newer this version includes code to perform extra verification and prevent signature wrapping login attacks php saml is vulnerable and allows signature wrapping context medium high priority it s a serious security issue saml is such a popular method of login because it s fast protects passwords reduces password proliferation the php application suitecrm doesn t hold any user password data only the identity provider such as microsoft facebook twitter google holds the user s password data the suitecrm php application is much harder to hack when users are logging in with saml because the hacker would need to attack the security of the identity provider social network typically yet social networks typically run high security are constantly audited by world class security teams and send highly deliverable email and push notifications to the user of any suspicious login activity coming from new devices and new locations your environment suitecrm version used browser name and version e g chrome version bit mozilla firefox environment name and version e g mysql php mysql php operating system and version e g ubuntu debian
| 0
|
342,135
| 30,609,801,186
|
IssuesEvent
|
2023-07-23 13:14:19
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix ndarray.test_numpy_instance_lshift__
|
NumPy Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix ndarray.test_numpy_instance_lshift__ - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5636581453"><img src=https://img.shields.io/badge/-failure-red></a>
|
non_defect
|
fix ndarray test numpy instance lshift tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src
| 0
|
71,392
| 23,607,098,434
|
IssuesEvent
|
2022-08-24 09:11:03
|
FreeRADIUS/freeradius-server
|
https://api.github.com/repos/FreeRADIUS/freeradius-server
|
opened
|
[defect]: free radius docker cannot mount raddb/certs directory
|
defect
|
### What type of defect/bug is this?
Unexpected behaviour (obvious or verified by project member)
### How can the issue be reproduced?
I am currently using free radius docker follwing the instructions (https://hub.docker.com/r/freeradius/freeradius-server/).
After starting the free radius server, I want to change the certificate configurations in the container. So I tried to mount the /etc/raddb/certs directory, but i found that i cannot do so. Doing so makes the container unable to start. I wonder if there is any way to mount /etc/raddb/certs in the free radius container ??? Thanks!!!
My docker-compose.yml is as attached.

### Log output from the FreeRADIUS daemon
```shell
NA
```
### Relevant log output from client utilities
_No response_
### Backtrace from LLDB or GDB
_No response_
|
1.0
|
[defect]: free radius docker cannot mount raddb/certs directory - ### What type of defect/bug is this?
Unexpected behaviour (obvious or verified by project member)
### How can the issue be reproduced?
I am currently using free radius docker follwing the instructions (https://hub.docker.com/r/freeradius/freeradius-server/).
After starting the free radius server, I want to change the certificate configurations in the container. So I tried to mount the /etc/raddb/certs directory, but i found that i cannot do so. Doing so makes the container unable to start. I wonder if there is any way to mount /etc/raddb/certs in the free radius container ??? Thanks!!!
My docker-compose.yml is as attached.

### Log output from the FreeRADIUS daemon
```shell
NA
```
### Relevant log output from client utilities
_No response_
### Backtrace from LLDB or GDB
_No response_
|
defect
|
free radius docker cannot mount raddb certs directory what type of defect bug is this unexpected behaviour obvious or verified by project member how can the issue be reproduced i am currently using free radius docker follwing the instructions after starting the free radius server i want to change the certificate configurations in the container so i tried to mount the etc raddb certs directory but i found that i cannot do so doing so makes the container unable to start i wonder if there is any way to mount etc raddb certs in the free radius container thanks my docker compose yml is as attached log output from the freeradius daemon shell na relevant log output from client utilities no response backtrace from lldb or gdb no response
| 1
|
779,915
| 27,371,556,808
|
IssuesEvent
|
2023-02-28 00:12:36
|
internetarchive/openlibrary
|
https://api.github.com/repos/internetarchive/openlibrary
|
closed
|
Add more data to Reading Log Export Option
|
Theme: Public APIs Priority: 3 Theme: Reading Log export Lead: @jimchamp
|
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I exported my reading log and the data I got:

While I did a similar export from Goodreads and the data I got:

### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
So I believe we can add **Title**, **Author**, **Year of Publication**, **Ratings**, **ISBN** e.t.c
### Stakeholders
<!-- @ tag stakeholders of this bug -->
@mekarpeles
|
1.0
|
Add more data to Reading Log Export Option - <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I exported my reading log and the data I got:

While I did a similar export from Goodreads and the data I got:

### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
So I believe we can add **Title**, **Author**, **Year of Publication**, **Ratings**, **ISBN** e.t.c
### Stakeholders
<!-- @ tag stakeholders of this bug -->
@mekarpeles
|
non_defect
|
add more data to reading log export option i exported my reading log and the data i got while i did a similar export from goodreads and the data i got describe the problem that you d like solved so i believe we can add title author year of publication ratings isbn e t c stakeholders mekarpeles
| 0
|
23,391
| 3,808,190,102
|
IssuesEvent
|
2016-03-25 13:48:54
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Cells inside subdirectory
|
Defect view
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.2.5
* Platform and Target: Apache 2, PostgreSQL 9.4
### What you did
I tried to use cells withing subdirectory, e.g. `src/View/Cell/Admin/AdminMenuCell.php` and call `<?= $this->cell('Admin/AdminMenu') ?>` inside of the view file.
The display file is: `src/Template/Cell/Admin/AdminMenu/display.ctp`
### Expected Behavior
It should load the cell.
### Actual Behavior
The error says `Could not render cell - Cell view file "display" is missing. [CORE/src/View/Cell.php, line 264`.
I would say an issue within in these lines: https://github.com/cakephp/cakephp/blob/master/src/View/Cell.php#L207
Currently it expects the display file to be in `src/Template/Cell/AdminMenu/display.ctp`
|
1.0
|
Cells inside subdirectory - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.2.5
* Platform and Target: Apache 2, PostgreSQL 9.4
### What you did
I tried to use cells withing subdirectory, e.g. `src/View/Cell/Admin/AdminMenuCell.php` and call `<?= $this->cell('Admin/AdminMenu') ?>` inside of the view file.
The display file is: `src/Template/Cell/Admin/AdminMenu/display.ctp`
### Expected Behavior
It should load the cell.
### Actual Behavior
The error says `Could not render cell - Cell view file "display" is missing. [CORE/src/View/Cell.php, line 264`.
I would say an issue within in these lines: https://github.com/cakephp/cakephp/blob/master/src/View/Cell.php#L207
Currently it expects the display file to be in `src/Template/Cell/AdminMenu/display.ctp`
|
defect
|
cells inside subdirectory this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target apache postgresql what you did i tried to use cells withing subdirectory e g src view cell admin adminmenucell php and call cell admin adminmenu inside of the view file the display file is src template cell admin adminmenu display ctp expected behavior it should load the cell actual behavior the error says could not render cell cell view file display is missing core src view cell php line i would say an issue within in these lines currently it expects the display file to be in src template cell adminmenu display ctp
| 1
|
95,037
| 8,528,742,059
|
IssuesEvent
|
2018-11-03 02:55:39
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Revamp deployment tests
|
area/test area/workload-api/deployment kind/feature lifecycle/rotten sig/apps
|
Ref. [Workload Controllers Test Plan](https://docs.google.com/document/d/1e2M3KD2wTfhISyyuqyX0pcUGZyzcyKoWo4CIlUojAdM/edit)
### integration
- [x] Add TestDeploymentBasic with subtests that replace basic behavior unit tests like creating ReplicaSets, reacting to RS/Pod events, orphan/adopt, etc. (#55439)
- [x] Add TestDeploymentRollingUpdate with subtests for roll out, roll back, roll over, etc. (#54819)
- [x] Shorten poll interval (can be 100ms) and use `PollImmediate()` instead of `Poll()` (#52896)
### e2e
Convert "edge case" e2e tests to integration tests.
- [x] testPausedDeployment (#52054)
- [x] testRollbackDeploymentRSNoRevision (#53918)
- [x] testDeploymentLabelAdopted (#54105)
- [x] testScalePausedDeployment (#52054)
- [x] testScaledRolloutDeployment (#54824)
- [x] testOverlappingDeployment (#54326)
- [x] testFailedDeployment (#54331)
- [x] testDeploymentHashCollisionAvoidance (#52896)
### cmd
- [x] Make sure all appropriate kubectl commands are tested (ref cmd tests in [Workload Controller Tests](https://docs.google.com/spreadsheets/d/1YsKwQvcPX_eaEdCxjAooIz1cgnr0uHxrki3xBJAK-eM/edit#gid=0))
### upgrade
- [x] Don't pry into the ReplicaSets. Just check Deployment Status. The RS upgrade test will check the next layer down. (#53841)
### unit test
- [ ] Refactor sync call tree so we can unit test decisions (on this sync pass, create this, delete that) without needing a test fixture with mocks.
- [ ] Remove unit tests made redundant by new integration tests below (especially ones that need reactor/mock setup).
@kubernetes/sig-apps-misc
|
1.0
|
Revamp deployment tests - Ref. [Workload Controllers Test Plan](https://docs.google.com/document/d/1e2M3KD2wTfhISyyuqyX0pcUGZyzcyKoWo4CIlUojAdM/edit)
### integration
- [x] Add TestDeploymentBasic with subtests that replace basic behavior unit tests like creating ReplicaSets, reacting to RS/Pod events, orphan/adopt, etc. (#55439)
- [x] Add TestDeploymentRollingUpdate with subtests for roll out, roll back, roll over, etc. (#54819)
- [x] Shorten poll interval (can be 100ms) and use `PollImmediate()` instead of `Poll()` (#52896)
### e2e
Convert "edge case" e2e tests to integration tests.
- [x] testPausedDeployment (#52054)
- [x] testRollbackDeploymentRSNoRevision (#53918)
- [x] testDeploymentLabelAdopted (#54105)
- [x] testScalePausedDeployment (#52054)
- [x] testScaledRolloutDeployment (#54824)
- [x] testOverlappingDeployment (#54326)
- [x] testFailedDeployment (#54331)
- [x] testDeploymentHashCollisionAvoidance (#52896)
### cmd
- [x] Make sure all appropriate kubectl commands are tested (ref cmd tests in [Workload Controller Tests](https://docs.google.com/spreadsheets/d/1YsKwQvcPX_eaEdCxjAooIz1cgnr0uHxrki3xBJAK-eM/edit#gid=0))
### upgrade
- [x] Don't pry into the ReplicaSets. Just check Deployment Status. The RS upgrade test will check the next layer down. (#53841)
### unit test
- [ ] Refactor sync call tree so we can unit test decisions (on this sync pass, create this, delete that) without needing a test fixture with mocks.
- [ ] Remove unit tests made redundant by new integration tests below (especially ones that need reactor/mock setup).
@kubernetes/sig-apps-misc
|
non_defect
|
revamp deployment tests ref integration add testdeploymentbasic with subtests that replace basic behavior unit tests like creating replicasets reacting to rs pod events orphan adopt etc add testdeploymentrollingupdate with subtests for roll out roll back roll over etc shorten poll interval can be and use pollimmediate instead of poll convert edge case tests to integration tests testpauseddeployment testrollbackdeploymentrsnorevision testdeploymentlabeladopted testscalepauseddeployment testscaledrolloutdeployment testoverlappingdeployment testfaileddeployment testdeploymenthashcollisionavoidance cmd make sure all appropriate kubectl commands are tested ref cmd tests in upgrade don t pry into the replicasets just check deployment status the rs upgrade test will check the next layer down unit test refactor sync call tree so we can unit test decisions on this sync pass create this delete that without needing a test fixture with mocks remove unit tests made redundant by new integration tests below especially ones that need reactor mock setup kubernetes sig apps misc
| 0
|
394,992
| 27,049,787,338
|
IssuesEvent
|
2023-02-13 12:26:43
|
hyuunnzy/uxui_issue_practice
|
https://api.github.com/repos/hyuunnzy/uxui_issue_practice
|
opened
|
유저 모델링 작업
|
documentation enhancement
|
## 설명
- 유저 모델링이란 무엇인지
- 우리 서비스에 가입할 유저의 속성 정의
## 할 일
- [ ] 유저 모델 초기화
- [ ] 유저 모델 정의
- [ ] 유저 모델 검증
## 참조
[Go to google] (http://www.google.com)
|
1.0
|
유저 모델링 작업 - ## 설명
- 유저 모델링이란 무엇인지
- 우리 서비스에 가입할 유저의 속성 정의
## 할 일
- [ ] 유저 모델 초기화
- [ ] 유저 모델 정의
- [ ] 유저 모델 검증
## 참조
[Go to google] (http://www.google.com)
|
non_defect
|
유저 모델링 작업 설명 유저 모델링이란 무엇인지 우리 서비스에 가입할 유저의 속성 정의 할 일 유저 모델 초기화 유저 모델 정의 유저 모델 검증 참조
| 0
|
49,594
| 13,187,237,666
|
IssuesEvent
|
2020-08-13 02:47:00
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
building multiple copies of HTML docs (Trac #1723)
|
Incomplete Migration Migrated from Trac defect other
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1723">https://code.icecube.wisc.edu/ticket/1723</a>, reported by nega and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-09T14:50:36",
"description": "with the addition of sphinx-apidoc in r2580/IceTray we're now building two copies of the docs that step on each other and barf warnings everywhere",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1465483836277867",
"component": "other",
"summary": "building multiple copies of HTML docs",
"priority": "normal",
"keywords": "documentation sphinx html",
"time": "2016-06-01T21:10:22",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
building multiple copies of HTML docs (Trac #1723) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1723">https://code.icecube.wisc.edu/ticket/1723</a>, reported by nega and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-09T14:50:36",
"description": "with the addition of sphinx-apidoc in r2580/IceTray we're now building two copies of the docs that step on each other and barf warnings everywhere",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1465483836277867",
"component": "other",
"summary": "building multiple copies of HTML docs",
"priority": "normal",
"keywords": "documentation sphinx html",
"time": "2016-06-01T21:10:22",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
defect
|
building multiple copies of html docs trac migrated from json status closed changetime description with the addition of sphinx apidoc in icetray we re now building two copies of the docs that step on each other and barf warnings everywhere reporter nega cc resolution fixed ts component other summary building multiple copies of html docs priority normal keywords documentation sphinx html time milestone owner kjmeagher type defect
| 1
|
28,353
| 5,241,426,558
|
IssuesEvent
|
2017-01-31 15:41:28
|
google/google-toolbox-for-mac
|
https://api.github.com/repos/google/google-toolbox-for-mac
|
closed
|
Unit test crash if access to address book is not granted
|
auto-migrated Priority-Medium Type-Defect
|
```
Repro steps:
1. Checkout
2. Open project
3. Run unit tests
4. When prompted, do not allow "otest" to access address book
5. Crash on GTMABAddressBook.m:84
ABGetSharedAddressBook() returns nil if access is not granted. Trying to retain
nil will cause a crash (I'm using OS X 10.8). I have attached a patch that
checks for nil before attempting to retain. This avoids the crash, though the
unit tests will obviously fail.
```
Original issue reported on code.google.com by `dlre...@gmail.com` on 19 Feb 2013 at 8:43
Attachments:
- [GTMABAddressBook-unittest-crashfix.diff](https://storage.googleapis.com/google-code-attachments/google-toolbox-for-mac/issue-90/comment-0/GTMABAddressBook-unittest-crashfix.diff)
|
1.0
|
Unit test crash if access to address book is not granted - ```
Repro steps:
1. Checkout
2. Open project
3. Run unit tests
4. When prompted, do not allow "otest" to access address book
5. Crash on GTMABAddressBook.m:84
ABGetSharedAddressBook() returns nil if access is not granted. Trying to retain
nil will cause a crash (I'm using OS X 10.8). I have attached a patch that
checks for nil before attempting to retain. This avoids the crash, though the
unit tests will obviously fail.
```
Original issue reported on code.google.com by `dlre...@gmail.com` on 19 Feb 2013 at 8:43
Attachments:
- [GTMABAddressBook-unittest-crashfix.diff](https://storage.googleapis.com/google-code-attachments/google-toolbox-for-mac/issue-90/comment-0/GTMABAddressBook-unittest-crashfix.diff)
|
defect
|
unit test crash if access to address book is not granted repro steps checkout open project run unit tests when prompted do not allow otest to access address book crash on gtmabaddressbook m abgetsharedaddressbook returns nil if access is not granted trying to retain nil will cause a crash i m using os x i have attached a patch that checks for nil before attempting to retain this avoids the crash though the unit tests will obviously fail original issue reported on code google com by dlre gmail com on feb at attachments
| 1
|
18,383
| 3,052,129,374
|
IssuesEvent
|
2015-08-12 13:12:59
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
Share web cam icon allows multiple clicks
|
Component-Client Priority-Medium Status-Fixed Type-Defect
|
Originally reported on Google Code with ID 176
```
What steps will reproduce the problem?
1. Click to share web cam
2. Click to share web cam again
3.
What is the expected output? What do you see instead?
Viewer will see two share webcam windows. The share web cam button should
work like the headset button.
```
Reported by `ffdixon` on 2009-09-12 22:32:59
|
1.0
|
Share web cam icon allows multiple clicks - Originally reported on Google Code with ID 176
```
What steps will reproduce the problem?
1. Click to share web cam
2. Click to share web cam again
3.
What is the expected output? What do you see instead?
Viewer will see two share webcam windows. The share web cam button should
work like the headset button.
```
Reported by `ffdixon` on 2009-09-12 22:32:59
|
defect
|
share web cam icon allows multiple clicks originally reported on google code with id what steps will reproduce the problem click to share web cam click to share web cam again what is the expected output what do you see instead viewer will see two share webcam windows the share web cam button should work like the headset button reported by ffdixon on
| 1
|
18,587
| 3,697,778,398
|
IssuesEvent
|
2016-02-27 22:11:26
|
desandro/masonry
|
https://api.github.com/repos/desandro/masonry
|
closed
|
originLeft: false not working
|
test case required
|
Hi,
After adding origiLeft:false to my options right-to-left not working for masonry in version 4!
|
1.0
|
originLeft: false not working - Hi,
After adding origiLeft:false to my options right-to-left not working for masonry in version 4!
|
non_defect
|
originleft false not working hi after adding origileft false to my options right to left not working for masonry in version
| 0
|
633,064
| 20,244,210,920
|
IssuesEvent
|
2022-02-14 12:12:07
|
PoProstuMieciek/wikipedia-scraper
|
https://api.github.com/repos/PoProstuMieciek/wikipedia-scraper
|
opened
|
feat/links-parser
|
priority: high type: feat
|
**AC**
- function
- [ ] gets a `JsDom` instance of html document
- [ ] looks for all links (in anchor tag) in the document
- [ ] returns a list of links
|
1.0
|
feat/links-parser - **AC**
- function
- [ ] gets a `JsDom` instance of html document
- [ ] looks for all links (in anchor tag) in the document
- [ ] returns a list of links
|
non_defect
|
feat links parser ac function gets a jsdom instance of html document looks for all links in anchor tag in the document returns a list of links
| 0
|
20,856
| 3,422,237,201
|
IssuesEvent
|
2015-12-08 22:08:27
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Type variable warning on static generic method
|
analyzer-strong-mode area-analyzer priority-high Type-Defect
|
The following code produces a bogus warning that type parameters are not allowed on static members:
```dart
class C {
static /*=T*/ f/*<T>*/() => null;
}
```
`[warning] Static members cannot reference type parameters (/Users/leafp/tmp/scratch.dart, line 2, col 13)`
dartanalyzer version 1.14.0-edge.daf5f6d258a99428cc4dc9ff17645a33e4e4bebd
|
1.0
|
Type variable warning on static generic method - The following code produces a bogus warning that type parameters are not allowed on static members:
```dart
class C {
static /*=T*/ f/*<T>*/() => null;
}
```
`[warning] Static members cannot reference type parameters (/Users/leafp/tmp/scratch.dart, line 2, col 13)`
dartanalyzer version 1.14.0-edge.daf5f6d258a99428cc4dc9ff17645a33e4e4bebd
|
defect
|
type variable warning on static generic method the following code produces a bogus warning that type parameters are not allowed on static members dart class c static t f null static members cannot reference type parameters users leafp tmp scratch dart line col dartanalyzer version edge
| 1
|
57,867
| 16,109,184,658
|
IssuesEvent
|
2021-04-27 18:41:48
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
opened
|
[DEFECT] RELAP5 Interface - radiation card 6000000
|
defect priority_normal
|
--------
Defect Description
--------
If the radiation card ```6000000``` is inputted, the interface splits it in ```600 0000``` causing an error
##### What did you expect to see happen?
No split
##### What did you see instead?
error of the relap5 driven model
##### Do you have a suggested fix for the development team?
N/A
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
1.0
|
[DEFECT] RELAP5 Interface - radiation card 6000000 - --------
Defect Description
--------
If the radiation card ```6000000``` is inputted, the interface splits it in ```600 0000``` causing an error
##### What did you expect to see happen?
No split
##### What did you see instead?
error of the relap5 driven model
##### Do you have a suggested fix for the development team?
N/A
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
defect
|
interface radiation card defect description if the radiation card is inputted the interface splits it in causing an error what did you expect to see happen no split what did you see instead error of the driven model do you have a suggested fix for the development team n a describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
| 1
|
53,009
| 13,260,069,820
|
IssuesEvent
|
2020-08-20 17:38:35
|
jkoan/test-navit
|
https://api.github.com/repos/jkoan/test-navit
|
closed
|
SDL-gui doesn't work on my system since sone libraries can't be loaded - TLS problem (Trac #50)
|
Incomplete Migration KaZeR Migrated from Trac defect/bug gui/cegui
|
Migrated from http://trac.navit-project.org/ticket/50
```json
{
"status": "closed",
"changetime": "2007-12-16T22:42:43",
"_ts": "1197844963000000",
"description": "I installed and used navit successfully on my 32-bit-system and wanted to use it on my 64-bit Gentoo-machine. The GTK-gui works fine, but when I try to use the SDL/Cegui-GUI, I get these errors:\n{{{\n** (process:30338): WARNING **: can't load '/usr/lib/navit/graphics/libgraphics_opengl.so', Error '/usr/lib/libGLC.so.0: cannot allocate memory in static TLS block'\n** (process:30338): WARNING **: can't load '/usr/lib/navit/gui/libgui_sdl.so', Error '/usr/lib/libGLC.so.0: cannot allocate memory in static TLS block'\n}}}\nAfter recompiling cegui, quesoglc and navit several times with different combinations of CFLAGS and different versions of gcc I still didn't have any success.\n\nAfter some time I tried this a workaround:\n{{{\nLD_PRELOAD=/usr/lib/navit/graphics/libgraphics_opengl.so navit\n}}}\nthis works but it complains a little:\n{{{\nsh: symbol lookup error: /usr/lib/navit/graphics/libgraphics_opengl.so: undefined symbol: debug_level\n}}}\n\ncp15 proposed this one:\n{{{\nLD_PRELOAD=/usr/lib/libGLC.so navit\n}}}\nwhich works, too and doesn't even complain about missing symbols",
"reporter": "Kripton <kripton@kripserver.net>",
"cc": "",
"resolution": "invalid",
"time": "2007-12-16T21:33:29",
"component": "gui/cegui",
"summary": "SDL-gui doesn't work on my system since sone libraries can't be loaded - TLS problem",
"priority": "major",
"keywords": "",
"version": "0.0.3",
"milestone": "",
"owner": "KaZeR",
"type": "defect/bug",
"severity": ""
}
```
|
1.0
|
SDL-gui doesn't work on my system since sone libraries can't be loaded - TLS problem (Trac #50) - Migrated from http://trac.navit-project.org/ticket/50
```json
{
"status": "closed",
"changetime": "2007-12-16T22:42:43",
"_ts": "1197844963000000",
"description": "I installed and used navit successfully on my 32-bit-system and wanted to use it on my 64-bit Gentoo-machine. The GTK-gui works fine, but when I try to use the SDL/Cegui-GUI, I get these errors:\n{{{\n** (process:30338): WARNING **: can't load '/usr/lib/navit/graphics/libgraphics_opengl.so', Error '/usr/lib/libGLC.so.0: cannot allocate memory in static TLS block'\n** (process:30338): WARNING **: can't load '/usr/lib/navit/gui/libgui_sdl.so', Error '/usr/lib/libGLC.so.0: cannot allocate memory in static TLS block'\n}}}\nAfter recompiling cegui, quesoglc and navit several times with different combinations of CFLAGS and different versions of gcc I still didn't have any success.\n\nAfter some time I tried this a workaround:\n{{{\nLD_PRELOAD=/usr/lib/navit/graphics/libgraphics_opengl.so navit\n}}}\nthis works but it complains a little:\n{{{\nsh: symbol lookup error: /usr/lib/navit/graphics/libgraphics_opengl.so: undefined symbol: debug_level\n}}}\n\ncp15 proposed this one:\n{{{\nLD_PRELOAD=/usr/lib/libGLC.so navit\n}}}\nwhich works, too and doesn't even complain about missing symbols",
"reporter": "Kripton <kripton@kripserver.net>",
"cc": "",
"resolution": "invalid",
"time": "2007-12-16T21:33:29",
"component": "gui/cegui",
"summary": "SDL-gui doesn't work on my system since sone libraries can't be loaded - TLS problem",
"priority": "major",
"keywords": "",
"version": "0.0.3",
"milestone": "",
"owner": "KaZeR",
"type": "defect/bug",
"severity": ""
}
```
|
defect
|
sdl gui doesn t work on my system since sone libraries can t be loaded tls problem trac migrated from json status closed changetime ts description i installed and used navit successfully on my bit system and wanted to use it on my bit gentoo machine the gtk gui works fine but when i try to use the sdl cegui gui i get these errors n n process warning can t load usr lib navit graphics libgraphics opengl so error usr lib libglc so cannot allocate memory in static tls block n process warning can t load usr lib navit gui libgui sdl so error usr lib libglc so cannot allocate memory in static tls block n nafter recompiling cegui quesoglc and navit several times with different combinations of cflags and different versions of gcc i still didn t have any success n nafter some time i tried this a workaround n nld preload usr lib navit graphics libgraphics opengl so navit n nthis works but it complains a little n nsh symbol lookup error usr lib navit graphics libgraphics opengl so undefined symbol debug level n n proposed this one n nld preload usr lib libglc so navit n nwhich works too and doesn t even complain about missing symbols reporter kripton cc resolution invalid time component gui cegui summary sdl gui doesn t work on my system since sone libraries can t be loaded tls problem priority major keywords version milestone owner kazer type defect bug severity
| 1
|
42,251
| 10,917,915,647
|
IssuesEvent
|
2019-11-21 15:58:10
|
contao/contao
|
https://api.github.com/repos/contao/contao
|
closed
|
contao.cache.clear_internal and contao.cache.warm_internal need to be public
|
defect
|
**Affected version(s)**
all that work with Symfony>=4.0
**Description**
`contao.cache.clear_internal` and `contao.cache.warm_internal` are used directly via `$container->get()` (in `Contao\Automator`) but both aren't public. Using the automator command to purge/generate the internal cache therefore fails.
**How to reproduce**
```
$ contao-console contao:auto purgeInternalCache
The "contao.cache.clear_internal" service or alias has been removed or inlined when the container was compiled. You should either make it public, or stop using the container directly and use dependency injection instead. (see help contao:automator).
```
```
$ contao-console contao:auto generateInternalCache
The "contao.cache.warm_internal" service or alias has been removed or inlined when the container was compiled. You should either make it public, or stop using the container directly and use dependency injection instead. (see help contao:automator).
```
|
1.0
|
contao.cache.clear_internal and contao.cache.warm_internal need to be public - **Affected version(s)**
all that work with Symfony>=4.0
**Description**
`contao.cache.clear_internal` and `contao.cache.warm_internal` are used directly via `$container->get()` (in `Contao\Automator`) but both aren't public. Using the automator command to purge/generate the internal cache therefore fails.
**How to reproduce**
```
$ contao-console contao:auto purgeInternalCache
The "contao.cache.clear_internal" service or alias has been removed or inlined when the container was compiled. You should either make it public, or stop using the container directly and use dependency injection instead. (see help contao:automator).
```
```
$ contao-console contao:auto generateInternalCache
The "contao.cache.warm_internal" service or alias has been removed or inlined when the container was compiled. You should either make it public, or stop using the container directly and use dependency injection instead. (see help contao:automator).
```
|
defect
|
contao cache clear internal and contao cache warm internal need to be public affected version s all that work with symfony description contao cache clear internal and contao cache warm internal are used directly via container get in contao automator but both aren t public using the automator command to purge generate the internal cache therefore fails how to reproduce contao console contao auto purgeinternalcache the contao cache clear internal service or alias has been removed or inlined when the container was compiled you should either make it public or stop using the container directly and use dependency injection instead see help contao automator contao console contao auto generateinternalcache the contao cache warm internal service or alias has been removed or inlined when the container was compiled you should either make it public or stop using the container directly and use dependency injection instead see help contao automator
| 1
|
12,913
| 2,729,022,713
|
IssuesEvent
|
2015-04-16 04:13:18
|
lttldrgn/port-o-chat
|
https://api.github.com/repos/lttldrgn/port-o-chat
|
closed
|
Server needs to be able to chunk the read in data together
|
auto-migrated Priority-Medium Type-Defect
|
```
The current implementation is to read 8192 bytes and process the valid bytes,
but this needs to be changed to chunk the read in data together just in case a
packet is bigger than the read size. Also, 8192 is probably too big, may want
to use 512 or something else.
```
Original issue reported on code.google.com by `MikeU...@gmail.com` on 23 Jun 2011 at 4:35
|
1.0
|
Server needs to be able to chunk the read in data together - ```
The current implementation is to read 8192 bytes and process the valid bytes,
but this needs to be changed to chunk the read in data together just in case a
packet is bigger than the read size. Also, 8192 is probably too big, may want
to use 512 or something else.
```
Original issue reported on code.google.com by `MikeU...@gmail.com` on 23 Jun 2011 at 4:35
|
defect
|
server needs to be able to chunk the read in data together the current implementation is to read bytes and process the valid bytes but this needs to be changed to chunk the read in data together just in case a packet is bigger than the read size also is probably too big may want to use or something else original issue reported on code google com by mikeu gmail com on jun at
| 1
|
27,413
| 5,005,755,046
|
IssuesEvent
|
2016-12-12 11:48:28
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Possible to access disabled tab in tabView.
|
6.0.11 defect
|
## 1) Environment
- 5.3.5.
- Tomcat 7.0.59
- IE 11, Chrome, FireFox
## 2) Expected behavior
You shouldn't be able to access disabled tab in tabView.
## 3) Actual behavior
You can access disabled tab in tabView.
## 4) Steps to reproduce
1. Go to the page where you have tabView with couple of tabs, where at least one is disabled.
2. Click left mouse button on the disabled tab and keep the button pressed + tap 'Enter' button on the keyboard. (or left mouse button + enter simultaneously.
3. Disabled tab is rendered.
## 5) Sample XHTML
<p:tabView id="id1">
<p:tab id="tab1">
<p:panel id="panel1">
Hello.
</p:panel>
</p:tab>
<p:tab id="tab2">
<p:panel id="panel2" disabled="true">
Hello from disabled tab.
</p:panel>
</p:tab>
</p:tabView>
## 6) Sample bean
Not needed.
|
1.0
|
Possible to access disabled tab in tabView. - ## 1) Environment
- 5.3.5.
- Tomcat 7.0.59
- IE 11, Chrome, FireFox
## 2) Expected behavior
You shouldn't be able to access disabled tab in tabView.
## 3) Actual behavior
You can access disabled tab in tabView.
## 4) Steps to reproduce
1. Go to the page where you have tabView with couple of tabs, where at least one is disabled.
2. Click left mouse button on the disabled tab and keep the button pressed + tap 'Enter' button on the keyboard. (or left mouse button + enter simultaneously.
3. Disabled tab is rendered.
## 5) Sample XHTML
<p:tabView id="id1">
<p:tab id="tab1">
<p:panel id="panel1">
Hello.
</p:panel>
</p:tab>
<p:tab id="tab2">
<p:panel id="panel2" disabled="true">
Hello from disabled tab.
</p:panel>
</p:tab>
</p:tabView>
## 6) Sample bean
Not needed.
|
defect
|
possible to access disabled tab in tabview environment tomcat ie chrome firefox expected behavior you shouldn t be able to access disabled tab in tabview actual behavior you can access disabled tab in tabview steps to reproduce go to the page where you have tabview with couple of tabs where at least one is disabled click left mouse button on the disabled tab and keep the button pressed tap enter button on the keyboard or left mouse button enter simultaneously disabled tab is rendered sample xhtml hello hello from disabled tab sample bean not needed
| 1
|
55,236
| 7,966,451,701
|
IssuesEvent
|
2018-07-14 22:21:23
|
KenSuenobu/scattersphere
|
https://api.github.com/repos/KenSuenobu/scattersphere
|
closed
|
Create Pause-controllable Runnable in JobExecutor
|
difficult documentation enhancement invalid
|
Instead of using a single `CompletableFuture` per `JobExecutor` instance, make it so that multiple jobs can be created and executed. Every time a `queue()` call is used, this will queue up the `Job` to run by the name, returning the controller for that `Job`.
`JobExecutor` should introduce an additional function called `registerJob` which registers the `Job` by the name specified in the `Job.name` field. `queue(jobName: String)` will change to take a new parameter, which queues up the `Job` specified by the name.
`queue(jobName: String)` on a `Job` that does not exist by name will throw a `JobNotFoundException` with the `Job`'s name.
|
1.0
|
Create Pause-controllable Runnable in JobExecutor - Instead of using a single `CompletableFuture` per `JobExecutor` instance, make it so that multiple jobs can be created and executed. Every time a `queue()` call is used, this will queue up the `Job` to run by the name, returning the controller for that `Job`.
`JobExecutor` should introduce an additional function called `registerJob` which registers the `Job` by the name specified in the `Job.name` field. `queue(jobName: String)` will change to take a new parameter, which queues up the `Job` specified by the name.
`queue(jobName: String)` on a `Job` that does not exist by name will throw a `JobNotFoundException` with the `Job`'s name.
|
non_defect
|
create pause controllable runnable in jobexecutor instead of using a single completablefuture per jobexecutor instance make it so that multiple jobs can be created and executed every time a queue call is used this will queue up the job to run by the name returning the controller for that job jobexecutor should introduce an additional function called registerjob which registers the job by the name specified in the job name field queue jobname string will change to take a new parameter which queues up the job specified by the name queue jobname string on a job that does not exist by name will throw a jobnotfoundexception with the job s name
| 0
|
230,148
| 17,603,741,215
|
IssuesEvent
|
2021-08-17 14:41:33
|
hotwire-django/turbo-django
|
https://api.github.com/repos/hotwire-django/turbo-django
|
closed
|
Move documentation to readthedocs.io
|
documentation
|
I'm partial to [mkdocs](https://github.com/mkdocs/mkdocs/) with its focus on markdown but, but am open to discussion on this!
|
1.0
|
Move documentation to readthedocs.io - I'm partial to [mkdocs](https://github.com/mkdocs/mkdocs/) with its focus on markdown but, but am open to discussion on this!
|
non_defect
|
move documentation to readthedocs io i m partial to with its focus on markdown but but am open to discussion on this
| 0
|
54,403
| 13,647,003,155
|
IssuesEvent
|
2020-09-26 01:04:58
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
"Copy" link in debug output doesn't work on Linux
|
defect
|
Checking Chrome's console it shows `navigator.clipboard` is undefined (Firefox doesn't even shown any error in console)
https://github.com/cakephp/cakephp/blob/991539874f6dcdb80d60968f05c5d8d2f08b067c/src/Error/Debug/dumpHeader.html#L248
The demo shown here does work for me https://web.dev/async-clipboard/#demos so perhaps we can update the code accordingly.
(Copying works in Firefox while in Chrome it shows permission error)
|
1.0
|
"Copy" link in debug output doesn't work on Linux - Checking Chrome's console it shows `navigator.clipboard` is undefined (Firefox doesn't even shown any error in console)
https://github.com/cakephp/cakephp/blob/991539874f6dcdb80d60968f05c5d8d2f08b067c/src/Error/Debug/dumpHeader.html#L248
The demo shown here does work for me https://web.dev/async-clipboard/#demos so perhaps we can update the code accordingly.
(Copying works in Firefox while in Chrome it shows permission error)
|
defect
|
copy link in debug output doesn t work on linux checking chrome s console it shows navigator clipboard is undefined firefox doesn t even shown any error in console the demo shown here does work for me so perhaps we can update the code accordingly copying works in firefox while in chrome it shows permission error
| 1
|
707,780
| 24,318,621,143
|
IssuesEvent
|
2022-09-30 08:44:53
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
login.cox.com - site is not usable
|
status-needsinfo browser-firefox priority-normal engine-gecko
|
<!-- @browser: Firefox 104.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:104.0) Gecko/20100101 Firefox/104.0 -->
<!-- @reported_with: unknown -->
**URL**: https://login.cox.com/app/ccires_tveverywhere_1/exk1jha9awwp564kZ0h8/sso/saml?SAMLRequest=hVNdb9owFH3fr4j8TuKEUYFFUrGyakgdRZDuoS%2BV69wSr4nt%2BTp8%2FPsZCBlbt%2B7RR%2Fd8XB97fL2rq2ADFqVWKYlDSgJQQhdSrVPykN%2F2huQ6%2BzBGXleJYZPGlWoJPxpAF0wQwTrPu9EKmxrsCuxGCnhY3qWkdM4giyI0IfekkBf6GUKha49EB7VoNfl690aCBFMvLRV3xzxnlUqvpfLs3VGBGxMJIS3gk9uAD7%2FflmDhKY5g9xp%2FL%2FmIb7dmcPXx9ZGWwwhRHx1JcKutgOMOKXnhFQIJZtOU8GG%2FpDHvS7gqXsR6MJADWo5G%2FVKsYZj4GVxwRLmBXyzEBmYKHVcuJQlNkh4d9eI4TyiL%2B4yOwpjSRxIsrHZa6OqTVKcbbaximqNEpngNyJxgh3tgSUjZ82kI2Zc8X%2FQW96ucBN%2FOzSSHZnxXCtmpi%2Fe1TGtMslN17JjYXiq8L8DPzZCsq9Kzwjd9jqNLg%2Bz8Uj7vHKhDdMzGrrTG7dkS0GhV5LpNcYL%2FHeO8A4Od88PSFj3Drdv%2FP9Gfhlkb8vdYLTT3prPpQldS7KMOXQltfBnduX30YGfT7GY%2B7wQv8Q7syNFf%2Fk32Ew%3D%3D&RelayState=0cae8149-d774-46de-8629-d2caf32de358&SigAlg=http%3A%2F%2Fwww.w3.org%2F2000%2F09%2Fxmldsig%23rsa-sha1&Signature=lewLn5NgTIVFoWlbAeSjE6FuEQ3vEbgrNGTCXgsrxrZpHhZ4yFxwqNgKKG9KDDRHWuC8oPoGMhVLanbQW2obVcuuYmuhHbhd6qyD%2BY1Jmh87YNdBT2griOOb1paIgtqKRWdFl%2F2Q5YyLOhGecG9rXReTpqkMhmpQozYzHn4v8tt7MZONUXG36Hq0U4Thpmj5NnHQ7q1uj0WBDyq8EVB8hujhTEP%2FTjooEJVBk43zA7CUr0xYA87ZTMTfbpCj9JoLNqyW0Nt5XFOrUevNWhpcygFl%2FAdi8Kp27dMsrw6bI8JyrZp9frzBETKQ9v%2F60BR5EUDXS6dcqKEHxUGou1wDEw%3D%3D
**Browser / Version**: Firefox 104.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
COX
403
Access Forbidden
You don't have permission to access this page.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/05b8c744-9e15-489b-bf00-3ac70ddcba7a.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
login.cox.com - site is not usable - <!-- @browser: Firefox 104.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:104.0) Gecko/20100101 Firefox/104.0 -->
<!-- @reported_with: unknown -->
**URL**: https://login.cox.com/app/ccires_tveverywhere_1/exk1jha9awwp564kZ0h8/sso/saml?SAMLRequest=hVNdb9owFH3fr4j8TuKEUYFFUrGyakgdRZDuoS%2BV69wSr4nt%2BTp8%2FPsZCBlbt%2B7RR%2Fd8XB97fL2rq2ADFqVWKYlDSgJQQhdSrVPykN%2F2huQ6%2BzBGXleJYZPGlWoJPxpAF0wQwTrPu9EKmxrsCuxGCnhY3qWkdM4giyI0IfekkBf6GUKha49EB7VoNfl690aCBFMvLRV3xzxnlUqvpfLs3VGBGxMJIS3gk9uAD7%2FflmDhKY5g9xp%2FL%2FmIb7dmcPXx9ZGWwwhRHx1JcKutgOMOKXnhFQIJZtOU8GG%2FpDHvS7gqXsR6MJADWo5G%2FVKsYZj4GVxwRLmBXyzEBmYKHVcuJQlNkh4d9eI4TyiL%2B4yOwpjSRxIsrHZa6OqTVKcbbaximqNEpngNyJxgh3tgSUjZ82kI2Zc8X%2FQW96ucBN%2FOzSSHZnxXCtmpi%2Fe1TGtMslN17JjYXiq8L8DPzZCsq9Kzwjd9jqNLg%2Bz8Uj7vHKhDdMzGrrTG7dkS0GhV5LpNcYL%2FHeO8A4Od88PSFj3Drdv%2FP9Gfhlkb8vdYLTT3prPpQldS7KMOXQltfBnduX30YGfT7GY%2B7wQv8Q7syNFf%2Fk32Ew%3D%3D&RelayState=0cae8149-d774-46de-8629-d2caf32de358&SigAlg=http%3A%2F%2Fwww.w3.org%2F2000%2F09%2Fxmldsig%23rsa-sha1&Signature=lewLn5NgTIVFoWlbAeSjE6FuEQ3vEbgrNGTCXgsrxrZpHhZ4yFxwqNgKKG9KDDRHWuC8oPoGMhVLanbQW2obVcuuYmuhHbhd6qyD%2BY1Jmh87YNdBT2griOOb1paIgtqKRWdFl%2F2Q5YyLOhGecG9rXReTpqkMhmpQozYzHn4v8tt7MZONUXG36Hq0U4Thpmj5NnHQ7q1uj0WBDyq8EVB8hujhTEP%2FTjooEJVBk43zA7CUr0xYA87ZTMTfbpCj9JoLNqyW0Nt5XFOrUevNWhpcygFl%2FAdi8Kp27dMsrw6bI8JyrZp9frzBETKQ9v%2F60BR5EUDXS6dcqKEHxUGou1wDEw%3D%3D
**Browser / Version**: Firefox 104.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
COX
403
Access Forbidden
You don't have permission to access this page.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/05b8c744-9e15-489b-bf00-3ac70ddcba7a.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
login cox com site is not usable url browser version firefox operating system mac os x tested another browser yes chrome problem type site is not usable description unable to login steps to reproduce cox access forbidden you don t have permission to access this page view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
436,032
| 12,544,346,392
|
IssuesEvent
|
2020-06-05 17:03:18
|
rathena/rathena
|
https://api.github.com/repos/rathena/rathena
|
closed
|
Mail script command
|
component:core mode:prerenewal mode:renewal priority:low type:enhancement
|
<!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**:
<!-- Please specify the client date you used. -->
* **Server Mode**: Pre-renewal
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Use this script https://rathena.org/board/topic/117956-rolling-in-the-deep/?do=findComment&comment=356321 as base of example and select an item that has refine on it wait until the npc return the item when it got cancelled. refine is gone.
|
1.0
|
Mail script command - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**:
<!-- Please specify the client date you used. -->
* **Server Mode**: Pre-renewal
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Use this script https://rathena.org/board/topic/117956-rolling-in-the-deep/?do=findComment&comment=356321 as base of example and select an item that has refine on it wait until the npc return the item when it got cancelled. refine is gone.
|
non_defect
|
mail script command rathena hash please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode pre renewal description of issue use this script as base of example and select an item that has refine on it wait until the npc return the item when it got cancelled refine is gone
| 0
|
881
| 3,175,973,508
|
IssuesEvent
|
2015-09-24 05:10:41
|
Fat-English/personal-site
|
https://api.github.com/repos/Fat-English/personal-site
|
opened
|
Wells on chat queue test don't display the result of queue tests
|
enhancement Service
|
Wells do not show the "Queue Matched" message that the queue list does.
|
1.0
|
Wells on chat queue test don't display the result of queue tests - Wells do not show the "Queue Matched" message that the queue list does.
|
non_defect
|
wells on chat queue test don t display the result of queue tests wells do not show the queue matched message that the queue list does
| 0
|
5,702
| 8,361,353,966
|
IssuesEvent
|
2018-10-03 14:09:10
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Admin Dashboard: Section 1: Interactions
|
Admin Approved FAI Requirements Ready
|
User Story: As an administrator, I want to be able to track metrics on user interactions with the system.
Acceptance Criteria:
On the Administration Sitewide tab, the third column of data will display Interactions
The following information will display in Interactions section: (Note: labels are different then mock - follow the AC). All numbers will be for sitewide, no timeframe, and will not limit to unique applicants.
- Total applications - total number of everyone who has applied to an opportunity sitewide. Not unique applicants. This should anyone who has applied that is an applicant + assigned participant + task completed. As an example, if I have applied to 4 opportunities, I will count as 4 applications.
- Total assigned - Total number of applicants who were moved to assigned participants for all opportunities sitewide. Includes Task completed. Not unique.
- Total completed: only those participants who have been marked as task completed. Not unique participants.
- Discussion posts: Total number of comments and reply to comments for all opportunities sitewide
Invision: https://opm.invisionapp.com/d/main#/console/13386795/300494469/preview

Notes:
- Across the board all numbers over 3 digits should display with a comma. All percentages should display a percent sign after.
- Data should not include data with an archived status.
|
1.0
|
Admin Dashboard: Section 1: Interactions - User Story: As an administrator, I want to be able to track metrics on user interactions with the system.
Acceptance Criteria:
On the Administration Sitewide tab, the third column of data will display Interactions
The following information will display in Interactions section: (Note: labels are different then mock - follow the AC). All numbers will be for sitewide, no timeframe, and will not limit to unique applicants.
- Total applications - total number of everyone who has applied to an opportunity sitewide. Not unique applicants. This should anyone who has applied that is an applicant + assigned participant + task completed. As an example, if I have applied to 4 opportunities, I will count as 4 applications.
- Total assigned - Total number of applicants who were moved to assigned participants for all opportunities sitewide. Includes Task completed. Not unique.
- Total completed: only those participants who have been marked as task completed. Not unique participants.
- Discussion posts: Total number of comments and reply to comments for all opportunities sitewide
Invision: https://opm.invisionapp.com/d/main#/console/13386795/300494469/preview

Notes:
- Across the board all numbers over 3 digits should display with a comma. All percentages should display a percent sign after.
- Data should not include data with an archived status.
|
non_defect
|
admin dashboard section interactions user story as an administrator i want to be able to track metrics on user interactions with the system acceptance criteria on the administration sitewide tab the third column of data will display interactions the following information will display in interactions section note labels are different then mock follow the ac all numbers will be for sitewide no timeframe and will not limit to unique applicants total applications total number of everyone who has applied to an opportunity sitewide not unique applicants this should anyone who has applied that is an applicant assigned participant task completed as an example if i have applied to opportunities i will count as applications total assigned total number of applicants who were moved to assigned participants for all opportunities sitewide includes task completed not unique total completed only those participants who have been marked as task completed not unique participants discussion posts total number of comments and reply to comments for all opportunities sitewide invision notes across the board all numbers over digits should display with a comma all percentages should display a percent sign after data should not include data with an archived status
| 0
|
35,312
| 7,697,763,372
|
IssuesEvent
|
2018-05-18 20:02:39
|
arescentral/antares
|
https://api.github.com/repos/arescentral/antares
|
closed
|
Shoplifter bases marked as both shape-from-direction and self-animated
|
Complexity:Medium Data Priority:Low Type:Defect
|
Original [issue 122](https://code.google.com/p/antares/issues/detail?id=122) created by sfiera on 2012-12-07T04:24:38.000Z:
There are two objects whose attributes specify both kIsSelfAnimated and kShapeFromDirection:
- The Obish liner from “Shoplifter 1”
- The UNS Ares in “Shoplifter 2”
This shouldn't be. I've added a workaround, but the source data ought to be fixed eventually.
|
1.0
|
Shoplifter bases marked as both shape-from-direction and self-animated - Original [issue 122](https://code.google.com/p/antares/issues/detail?id=122) created by sfiera on 2012-12-07T04:24:38.000Z:
There are two objects whose attributes specify both kIsSelfAnimated and kShapeFromDirection:
- The Obish liner from “Shoplifter 1”
- The UNS Ares in “Shoplifter 2”
This shouldn't be. I've added a workaround, but the source data ought to be fixed eventually.
|
defect
|
shoplifter bases marked as both shape from direction and self animated original created by sfiera on there are two objects whose attributes specify both kisselfanimated and kshapefromdirection the obish liner from “shoplifter ” the uns ares in “shoplifter ” this shouldn t be i ve added a workaround but the source data ought to be fixed eventually
| 1
|
25,018
| 6,619,614,469
|
IssuesEvent
|
2017-09-21 12:57:15
|
devtools-html/debugger.html
|
https://api.github.com/repos/devtools-html/debugger.html
|
closed
|
[stepping] refactor stepping commands
|
available Code Health
|
I think we can make small change to our `command` and remove an unneeded action type.
The change is well tested so if the tests pass we should be good. It's a chance to learn about our promise middleware too, which is in utils/redux/middleware
```diff
index 979bd23..d3fd231 100644
--- a/src/actions/pause.js
+++ b/src/actions/pause.js
@@ -99,13 +99,10 @@ export function pauseOnExceptions(
* @static
*/
export function command({ type }: CommandType) {
- return ({ dispatch, client }: ThunkArgs) => {
- // execute debugger thread command e.g. stepIn, stepOver
- client[type]().then(() => dispatch({ type: "CLEAR_COMMAND" }));
-
+ return async ({ dispatch, client }: ThunkArgs) => {
return dispatch({
type: "COMMAND",
- value: { type }
+ [PROMISE]: client[type]()
});
};
}
diff --git a/src/reducers/pause.js b/src/reducers/pause.js
index fa7edef..e00136a 100644
--- a/src/reducers/pause.js
+++ b/src/reducers/pause.js
@@ -141,9 +141,9 @@ function update(state: PauseState = State(), action: Action): PauseState {
});
case "COMMAND":
- return { ...state, command: action.value.type };
-
- case "CLEAR_COMMAND":
+ if (action.status == "start") {
+ return { ...state, command: action.value.type };
+ }
return { ...state, command: "" };
}
```
|
1.0
|
[stepping] refactor stepping commands - I think we can make small change to our `command` and remove an unneeded action type.
The change is well tested so if the tests pass we should be good. It's a chance to learn about our promise middleware too, which is in utils/redux/middleware
```diff
index 979bd23..d3fd231 100644
--- a/src/actions/pause.js
+++ b/src/actions/pause.js
@@ -99,13 +99,10 @@ export function pauseOnExceptions(
* @static
*/
export function command({ type }: CommandType) {
- return ({ dispatch, client }: ThunkArgs) => {
- // execute debugger thread command e.g. stepIn, stepOver
- client[type]().then(() => dispatch({ type: "CLEAR_COMMAND" }));
-
+ return async ({ dispatch, client }: ThunkArgs) => {
return dispatch({
type: "COMMAND",
- value: { type }
+ [PROMISE]: client[type]()
});
};
}
diff --git a/src/reducers/pause.js b/src/reducers/pause.js
index fa7edef..e00136a 100644
--- a/src/reducers/pause.js
+++ b/src/reducers/pause.js
@@ -141,9 +141,9 @@ function update(state: PauseState = State(), action: Action): PauseState {
});
case "COMMAND":
- return { ...state, command: action.value.type };
-
- case "CLEAR_COMMAND":
+ if (action.status == "start") {
+ return { ...state, command: action.value.type };
+ }
return { ...state, command: "" };
}
```
|
non_defect
|
refactor stepping commands i think we can make small change to our command and remove an unneeded action type the change is well tested so if the tests pass we should be good it s a chance to learn about our promise middleware too which is in utils redux middleware diff index a src actions pause js b src actions pause js export function pauseonexceptions static export function command type commandtype return dispatch client thunkargs execute debugger thread command e g stepin stepover client then dispatch type clear command return async dispatch client thunkargs return dispatch type command value type client diff git a src reducers pause js b src reducers pause js index a src reducers pause js b src reducers pause js function update state pausestate state action action pausestate case command return state command action value type case clear command if action status start return state command action value type return state command
| 0
|
15,910
| 2,869,093,748
|
IssuesEvent
|
2015-06-05 23:17:00
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
add support for composing source maps
|
Area-Pkg Pkg-SourceMaps Priority-Medium Triaged Type-Defect
|
Create a script in source_maps that given mappings from file A -> B and B -> C, it creates a direct mapping from A -> C.
|
1.0
|
add support for composing source maps - Create a script in source_maps that given mappings from file A -> B and B -> C, it creates a direct mapping from A -> C.
|
defect
|
add support for composing source maps create a script in source maps that given mappings from file a gt b and b gt c it creates a direct mapping from a gt c
| 1
|
11,940
| 2,669,409,105
|
IssuesEvent
|
2015-03-23 15:22:03
|
b-steiner/blubbengine2
|
https://api.github.com/repos/b-steiner/blubbengine2
|
closed
|
Window-Resizing not working well
|
auto-migrated OpSys-All Projekt-Engine Type-Defect Version-0.2
|
```
Often wrong hit test
```
Original issue reported on code.google.com by `bernhard...@tuwien.ac.at` on 23 Jun 2014 at 8:39
|
1.0
|
Window-Resizing not working well - ```
Often wrong hit test
```
Original issue reported on code.google.com by `bernhard...@tuwien.ac.at` on 23 Jun 2014 at 8:39
|
defect
|
window resizing not working well often wrong hit test original issue reported on code google com by bernhard tuwien ac at on jun at
| 1
|
13,741
| 16,460,089,159
|
IssuesEvent
|
2021-05-21 17:35:11
|
ORelio/Minecraft-Console-Client
|
https://api.github.com/repos/ORelio/Minecraft-Console-Client
|
closed
|
OnUpdate crash
|
a:bug in:protocol-compatibility waiting-for:more-info
|
**Prerequisites**
- [x] I made sure I am running the latest [development build](https://ci.appveyor.com/project/ORelio/minecraft-console-client/build/artifacts)
- [x] I tried to [look for similar issues](https://github.com/ORelio/Minecraft-Console-Client/issues?q=is%3Aissue) before opening a new one
- [ ] I have set `debugmessages=true` in config to diagnose my issue
- [x] I have redacted session tokens and passwords before attaching screenshots

|
True
|
OnUpdate crash - **Prerequisites**
- [x] I made sure I am running the latest [development build](https://ci.appveyor.com/project/ORelio/minecraft-console-client/build/artifacts)
- [x] I tried to [look for similar issues](https://github.com/ORelio/Minecraft-Console-Client/issues?q=is%3Aissue) before opening a new one
- [ ] I have set `debugmessages=true` in config to diagnose my issue
- [x] I have redacted session tokens and passwords before attaching screenshots

|
non_defect
|
onupdate crash prerequisites i made sure i am running the latest i tried to before opening a new one i have set debugmessages true in config to diagnose my issue i have redacted session tokens and passwords before attaching screenshots
| 0
|
282,739
| 30,889,416,929
|
IssuesEvent
|
2023-08-04 02:41:32
|
Trinadh465/linux-4.1.15_CVE-2017-1000371
|
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2017-1000371
|
reopened
|
CVE-2022-3564 (High) detected in linuxlinux-4.6
|
Mend: dependency security vulnerability
|
## CVE-2022-3564 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2017-1000371/commit/8cc0bec3d85a996d6015c27f949826b9ffc4d1ae">8cc0bec3d85a996d6015c27f949826b9ffc4d1ae</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as critical was found in Linux Kernel. Affected by this vulnerability is the function l2cap_reassemble_sdu of the file net/bluetooth/l2cap_core.c of the component Bluetooth. The manipulation leads to use after free. It is recommended to apply a patch to fix this issue. The associated identifier of this vulnerability is VDB-211087.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3564>CVE-2022-3564</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3564">https://www.linuxkernelcves.com/cves/CVE-2022-3564</a></p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: v4.9.333,v4.14.299,v4.19.265,v5.4.224,v5.10.154,v5.15.78,v6.0.8,v6.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-3564 (High) detected in linuxlinux-4.6 - ## CVE-2022-3564 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2017-1000371/commit/8cc0bec3d85a996d6015c27f949826b9ffc4d1ae">8cc0bec3d85a996d6015c27f949826b9ffc4d1ae</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as critical was found in Linux Kernel. Affected by this vulnerability is the function l2cap_reassemble_sdu of the file net/bluetooth/l2cap_core.c of the component Bluetooth. The manipulation leads to use after free. It is recommended to apply a patch to fix this issue. The associated identifier of this vulnerability is VDB-211087.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3564>CVE-2022-3564</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3564">https://www.linuxkernelcves.com/cves/CVE-2022-3564</a></p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: v4.9.333,v4.14.299,v4.19.265,v5.4.224,v5.10.154,v5.15.78,v6.0.8,v6.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth core c net bluetooth core c vulnerability details a vulnerability classified as critical was found in linux kernel affected by this vulnerability is the function reassemble sdu of the file net bluetooth core c of the component bluetooth the manipulation leads to use after free it is recommended to apply a patch to fix this issue the associated identifier of this vulnerability is vdb publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
49,559
| 13,187,233,486
|
IssuesEvent
|
2020-08-13 02:46:15
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[iceprod2] clear iceprod_std* files between tasks on a pilot (Trac #1677)
|
Incomplete Migration Migrated from Trac defect iceprod
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1677">https://code.icecube.wisc.edu/ticket/1677</a>, reported by david.schultz and owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"description": "When a pilot runs multiple tasks in a row, we should clear the output/error/log files so the new task only sees what it has done, not past history.\n\nWhile the full pilot history is potentially useful for debugging, it's probably simpler and better to just truncate the files.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1462830916934929",
"component": "iceprod",
"summary": "[iceprod2] clear iceprod_std* files between tasks on a pilot",
"priority": "major",
"keywords": "",
"time": "2016-04-29T18:29:34",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[iceprod2] clear iceprod_std* files between tasks on a pilot (Trac #1677) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1677">https://code.icecube.wisc.edu/ticket/1677</a>, reported by david.schultz and owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"description": "When a pilot runs multiple tasks in a row, we should clear the output/error/log files so the new task only sees what it has done, not past history.\n\nWhile the full pilot history is potentially useful for debugging, it's probably simpler and better to just truncate the files.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1462830916934929",
"component": "iceprod",
"summary": "[iceprod2] clear iceprod_std* files between tasks on a pilot",
"priority": "major",
"keywords": "",
"time": "2016-04-29T18:29:34",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
|
defect
|
clear iceprod std files between tasks on a pilot trac migrated from json status closed changetime description when a pilot runs multiple tasks in a row we should clear the output error log files so the new task only sees what it has done not past history n nwhile the full pilot history is potentially useful for debugging it s probably simpler and better to just truncate the files reporter david schultz cc resolution wontfix ts component iceprod summary clear iceprod std files between tasks on a pilot priority major keywords time milestone owner david schultz type defect
| 1
|
822,235
| 30,859,427,949
|
IssuesEvent
|
2023-08-03 00:46:48
|
paleobot/pbot-dev
|
https://api.github.com/repos/paleobot/pbot-dev
|
closed
|
Collection node field edits
|
high priority
|
Under Optional Fields, Age: replace "Notes on geographic age" with "Notes on age information". Also, allow multiple lines for this text box as it gets more filled.
Under Optional Fields, Geographic: allow multiple lines for the "Notes on geographic information" text box as it gets more filled.
Under Optional Fields, Geologic: allow multiple lines for the "Additional description of lithology" and the "Notes on environment" text box as it gets more filled.
Under Optional fields: Collecting: allow multiple lines for the "Notes on collection methods" text box as it gets more filled.
|
1.0
|
Collection node field edits - Under Optional Fields, Age: replace "Notes on geographic age" with "Notes on age information". Also, allow multiple lines for this text box as it gets more filled.
Under Optional Fields, Geographic: allow multiple lines for the "Notes on geographic information" text box as it gets more filled.
Under Optional Fields, Geologic: allow multiple lines for the "Additional description of lithology" and the "Notes on environment" text box as it gets more filled.
Under Optional fields: Collecting: allow multiple lines for the "Notes on collection methods" text box as it gets more filled.
|
non_defect
|
collection node field edits under optional fields age replace notes on geographic age with notes on age information also allow multiple lines for this text box as it gets more filled under optional fields geographic allow multiple lines for the notes on geographic information text box as it gets more filled under optional fields geologic allow multiple lines for the additional description of lithology and the notes on environment text box as it gets more filled under optional fields collecting allow multiple lines for the notes on collection methods text box as it gets more filled
| 0
|
78,771
| 27,752,006,634
|
IssuesEvent
|
2023-03-15 21:36:16
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
Ouptuts/file_base for a multiapp that has multiple subapps should yell at you
|
P: minor T: defect
|
## Bug Description
If we set multiple subapps to write with a file_base, we have them all write to the same file
This doesnt work for people who want one output per subapp
## Steps to Reproduce
Create 2 subapps and use the Outputs/file_base parameter in the subapps
## Impact
User confusion
|
1.0
|
Ouptuts/file_base for a multiapp that has multiple subapps should yell at you - ## Bug Description
If we set multiple subapps to write with a file_base, we have them all write to the same file
This doesnt work for people who want one output per subapp
## Steps to Reproduce
Create 2 subapps and use the Outputs/file_base parameter in the subapps
## Impact
User confusion
|
defect
|
ouptuts file base for a multiapp that has multiple subapps should yell at you bug description if we set multiple subapps to write with a file base we have them all write to the same file this doesnt work for people who want one output per subapp steps to reproduce create subapps and use the outputs file base parameter in the subapps impact user confusion
| 1
|
81,035
| 10,221,120,687
|
IssuesEvent
|
2019-08-16 00:02:22
|
ActoSoft/topografiApp
|
https://api.github.com/repos/ActoSoft/topografiApp
|
reopened
|
Employees List Mockup UI
|
documentation enhancement
|
We have employees in this app. So we need to list everything. We're a using a CSS framework call Ant Design. So try to be based on the table style of Ant Design to do this.
**Resources**
https://ant.design/components/table/
**Data**
The columns of the table will be:
- Nombre
- Rol
- Dirección
- Teléfono


I added how we are using this table right now in other apps.
|
1.0
|
Employees List Mockup UI - We have employees in this app. So we need to list everything. We're a using a CSS framework call Ant Design. So try to be based on the table style of Ant Design to do this.
**Resources**
https://ant.design/components/table/
**Data**
The columns of the table will be:
- Nombre
- Rol
- Dirección
- Teléfono


I added how we are using this table right now in other apps.
|
non_defect
|
employees list mockup ui we have employees in this app so we need to list everything we re a using a css framework call ant design so try to be based on the table style of ant design to do this resources data the columns of the table will be nombre rol dirección teléfono i added how we are using this table right now in other apps
| 0
|
94,528
| 10,826,188,021
|
IssuesEvent
|
2019-11-09 20:56:16
|
tdwg/developers
|
https://api.github.com/repos/tdwg/developers
|
opened
|
As a community we need a Code of Conduct
|
documentation help wanted
|
How do people feel about using on of the default options provided by Github?
|
1.0
|
As a community we need a Code of Conduct - How do people feel about using on of the default options provided by Github?
|
non_defect
|
as a community we need a code of conduct how do people feel about using on of the default options provided by github
| 0
|
820,254
| 30,766,272,218
|
IssuesEvent
|
2023-07-30 10:42:27
|
Beep6581/RawTherapee
|
https://api.github.com/repos/Beep6581/RawTherapee
|
closed
|
RawPedia broken design (Template:K)
|
type: bug external: RawPedia priority: medium
|
**Short description**
The design of RawPedia seems to be broken
**Steps to reproduce**
1. Visit http://rawpedia.rawtherapee.com/Getting_Started
2. Section _Basics_
**Additional information**
RawPedia had a black design before. I don't know if the new theme is intentional or maybe some update broke it. It seems like "{{k|w}} " is the part that causes the issue ( http://rawpedia.rawtherapee.com/Template:K )
|
1.0
|
RawPedia broken design (Template:K) - **Short description**
The design of RawPedia seems to be broken
**Steps to reproduce**
1. Visit http://rawpedia.rawtherapee.com/Getting_Started
2. Section _Basics_
**Additional information**
RawPedia had a black design before. I don't know if the new theme is intentional or maybe some update broke it. It seems like "{{k|w}} " is the part that causes the issue ( http://rawpedia.rawtherapee.com/Template:K )
|
non_defect
|
rawpedia broken design template k short description the design of rawpedia seems to be broken steps to reproduce visit section basics additional information rawpedia had a black design before i don t know if the new theme is intentional or maybe some update broke it it seems like k w is the part that causes the issue
| 0
|
479,603
| 13,803,742,984
|
IssuesEvent
|
2020-10-11 05:03:27
|
AY2021S1-CS2103T-W16-3/tp
|
https://api.github.com/repos/AY2021S1-CS2103T-W16-3/tp
|
opened
|
Connect Logic to Ui
|
priority.medium :2nd_place_medal: type.enhancement :+1:
|
Currently, the Ui is not able to display the list of expenses and incomes in its respective Expense and Income tab.
Refactor the the `Income Panel` and `Expense Panel` to capture `ListView<Income>` and `ListView<Expense>` respectively instead of `ListView<Transaction>`.
|
1.0
|
Connect Logic to Ui - Currently, the Ui is not able to display the list of expenses and incomes in its respective Expense and Income tab.
Refactor the the `Income Panel` and `Expense Panel` to capture `ListView<Income>` and `ListView<Expense>` respectively instead of `ListView<Transaction>`.
|
non_defect
|
connect logic to ui currently the ui is not able to display the list of expenses and incomes in its respective expense and income tab refactor the the income panel and expense panel to capture listview and listview respectively instead of listview
| 0
|
50,688
| 13,187,681,564
|
IssuesEvent
|
2020-08-13 04:13:04
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
tpx doc improvement (Trac #1187)
|
Migrated from Trac combo reconstruction defect
|
rst doc does not list maintainer at top of page
also add link to doxygen documentaion
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1187">https://code.icecube.wisc.edu/ticket/1187</a>, reported by kjmeagher and owned by karg</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "rst doc does not list maintainer at top of page\nalso add link to doxygen documentaion",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "tpx doc improvement",
"priority": "normal",
"keywords": "",
"time": "2015-08-19T13:36:23",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
tpx doc improvement (Trac #1187) - rst doc does not list maintainer at top of page
also add link to doxygen documentaion
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1187">https://code.icecube.wisc.edu/ticket/1187</a>, reported by kjmeagher and owned by karg</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "rst doc does not list maintainer at top of page\nalso add link to doxygen documentaion",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "tpx doc improvement",
"priority": "normal",
"keywords": "",
"time": "2015-08-19T13:36:23",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
</p>
</details>
|
defect
|
tpx doc improvement trac rst doc does not list maintainer at top of page also add link to doxygen documentaion migrated from json status closed changetime description rst doc does not list maintainer at top of page nalso add link to doxygen documentaion reporter kjmeagher cc resolution fixed ts component combo reconstruction summary tpx doc improvement priority normal keywords time milestone owner karg type defect
| 1
|
9,810
| 2,615,175,281
|
IssuesEvent
|
2015-03-01 06:58:32
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
Wifi cracking not working
|
auto-migrated Priority-Triage Type-Defect
|
```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
1. What operating system are you using (Linux is the only supported OS)?
Ubuntu 14.04 LTS
2. Is your wireless card in monitor mode (yes/no)?
no
3. What is the signal strength of the Access Point you are trying to crack?
full
4. What is the manufacturer and model # of the device you are trying to
crack?
Netgear N150
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b EC:1A:59:17:14:13
6. Please describe what you think the issue is.
I'm trying to hack the wifi router in my home.I get the same error everytime.I
haven't hacked it even once.
7. Paste the output from Reaver below.
root@jagruthnath-X550CC:~# reaver -i mon0 -c 1 -b 28:C6:8E:3D:07:EA -vv -a
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from 28:C6:8E:3D:07:EA
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Associated with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:13 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:18 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:25 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:32 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:38 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:44 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:50 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:56 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:02 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:08 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:14 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:20 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
^C
[+] Nothing done, nothing to save.
```
Original issue reported on code.google.com by `RockyPar...@gmail.com` on 21 Jun 2014 at 2:32
|
1.0
|
Wifi cracking not working - ```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
1. What operating system are you using (Linux is the only supported OS)?
Ubuntu 14.04 LTS
2. Is your wireless card in monitor mode (yes/no)?
no
3. What is the signal strength of the Access Point you are trying to crack?
full
4. What is the manufacturer and model # of the device you are trying to
crack?
Netgear N150
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b EC:1A:59:17:14:13
6. Please describe what you think the issue is.
I'm trying to hack the wifi router in my home.I get the same error everytime.I
haven't hacked it even once.
7. Paste the output from Reaver below.
root@jagruthnath-X550CC:~# reaver -i mon0 -c 1 -b 28:C6:8E:3D:07:EA -vv -a
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from 28:C6:8E:3D:07:EA
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: (null))
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Associated with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:13 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:18 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:25 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:32 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:38 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:44 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:50 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:31:56 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:02 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:08 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:14 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[!] WARNING: 10 failed connections in a row
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Nothing done, nothing to save.
[+] 0.00% complete @ 2014-06-21 19:32:20 (0 seconds/pin)
[+] Trying pin 12345670
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[!] WARNING: Failed to associate with 28:C6:8E:3D:07:EA (ESSID: giribabu)
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
^C
[+] Nothing done, nothing to save.
```
Original issue reported on code.google.com by `RockyPar...@gmail.com` on 21 Jun 2014 at 2:32
|
defect
|
wifi cracking not working a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os ubuntu lts is your wireless card in monitor mode yes no no what is the signal strength of the access point you are trying to crack full what is the manufacturer and model of the device you are trying to crack netgear what is the entire command line string you are supplying to reaver reaver i b ec please describe what you think the issue is i m trying to hack the wifi router in my home i get the same error everytime i haven t hacked it even once paste the output from reaver below root jagruthnath reaver i c b ea vv a reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ea warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid null warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu associated with ea essid giribabu trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin trying pin warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu warning failed to associate with ea essid giribabu sending eapol start request sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request sending wsc nack wps transaction failed code re trying last pin c nothing done nothing to save original issue reported on code google com by rockypar gmail com on jun at
| 1
|
56,888
| 15,429,047,986
|
IssuesEvent
|
2021-03-06 01:45:33
|
SAP/fundamental-styles
|
https://api.github.com/repos/SAP/fundamental-styles
|
closed
|
Object Marker color while in an active table cell
|
Bug Defect Hunting denoland
|
**Describe the bug**
A clear and concise description of what the bug is.
the color of object marker doesn't invert in an active table cell
> Is this issue related to a specific component?
object marker
<img width="243" alt="Screen Shot 2021-03-05 at 9 17 22 AM" src="https://user-images.githubusercontent.com/4380815/110127620-d741a480-7d93-11eb-9011-8e18b4b55da8.png">
|
1.0
|
Object Marker color while in an active table cell -
**Describe the bug**
A clear and concise description of what the bug is.
the color of object marker doesn't invert in an active table cell
> Is this issue related to a specific component?
object marker
<img width="243" alt="Screen Shot 2021-03-05 at 9 17 22 AM" src="https://user-images.githubusercontent.com/4380815/110127620-d741a480-7d93-11eb-9011-8e18b4b55da8.png">
|
defect
|
object marker color while in an active table cell describe the bug a clear and concise description of what the bug is the color of object marker doesn t invert in an active table cell is this issue related to a specific component object marker img width alt screen shot at am src
| 1
|
78,011
| 27,276,309,283
|
IssuesEvent
|
2023-02-23 05:44:30
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
The zfs kernel module fails to be installed: ERROR: could not insert 'zfs': Key was rejected by service
|
Type: Defect
|
### System information
Distribution Name | Fedora Xfce Desktop
Distribution Version | fc37
Kernel Version | 6.1.12-200.fc37.x86_64
Architecture | x86_64
OpenZFS Version | zfs-2.1.99-1740_gd9e64a403
### Describe the problem you're observing
I tried to install the zfs module, but modprobe told me to reject the installation.
### Describe how to reproduce the problem
I tried using the zfs package provided by zfsonlinux.org and the ZFS RPM package built from GIT SCM, and they both had this problem. They seem to be related to a feature of the kernel module, What do I need to do to make it work?
Thanks.
```
[gesangtome@fedora ~]$ sudo modprobe zfs
[sudo] password for gesangtome:
modprobe: ERROR: could not insert 'zfs': Key was rejected by service
```
modinfo for zfs.ko
```
[gesangtome@fedora ~]$ modinfo zfs
filename: /lib/modules/6.1.12-200.fc37.x86_64/extra/zfs.ko.xz
version: 2.1.99-1740_gd9e64a403
license: CDDL
license: Dual BSD/GPL
license: Dual MIT/GPL
author: OpenZFS
description: ZFS
alias: zzstd
alias: zcommon
alias: zunicode
alias: znvpair
alias: zlua
alias: icp
alias: zavl
alias: devname:zfs
alias: char-major-10-249
srcversion: 13B813800C43864C4EF9B62
depends: spl
retpoline: Y
name: zfs
vermagic: 6.1.12-200.fc37.x86_64 SMP preempt mod_unload
sig_id: PKCS#7
signer: DKMS module signing key
sig_key: 1E:AD:B5:32:A6:8B:C0:D2:53:BE:D3:F9:AC:4C:DA:51:68:21:93:0E
sig_hashalgo: sha512
signature: 29:82:06:D1:87:1F:21:30:47:02:35:61:F0:DF:F2:E6:67:24:96:B2:
57:82:37:13:D4:12:EE:94:67:C8:87:AA:D9:B6:28:0B:7C:22:D6:21:
5C:4B:19:31:F8:FB:9E:4B:56:7F:D6:82:18:80:8F:7D:E8:FC:E0:FB:
2F:FD:48:80:0F:C6:6A:C3:3D:4B:90:B5:59:6E:10:24:56:81:A3:2A:
F8:55:BB:67:36:DE:05:25:FF:E2:AF:12:E6:9D:22:CA:BB:1E:8F:86:
AF:26:DD:A3:F3:3D:34:EC:2A:4D:2D:C5:AA:38:6A:97:66:EF:C0:09:
8B:3D:2F:EA:2B:C1:54:6B:E7:64:0C:41:01:40:23:F6:EC:4C:65:D9:
D9:BE:5E:6D:14:23:DA:BC:15:32:BE:D9:2D:67:D4:D6:6F:B0:6C:F8:
3A:B5:71:20:69:56:F5:FE:28:FE:82:FE:64:65:37:ED:A3:51:62:BA:
00:72:76:0A:24:E1:1B:D7:7A:76:73:3D:71:47:BD:97:2C:7C:6E:06:
F6:43:E2:25:43:AF:B3:01:7D:99:B5:9E:C3:D2:9F:95:3F:1C:1B:C6:
A5:63:40:3C:E9:82:18:4C:C7:8C:27:83:98:63:D4:6B:7F:B9:25:27:
A5:C3:D9:D5:BC:24:66:39:6F:07:E2:20:BC:C8:9B:B7
parm: zvol_inhibit_dev:Do not create zvol device nodes (uint)
parm: zvol_major:Major number for zvol device (uint)
parm: zvol_threads:Number of threads to handle I/O requests. Setto 0 to use all active CPUs (uint)
parm: zvol_request_sync:Synchronously handle bio requests (uint)
parm: zvol_max_discard_blocks:Max number of blocks to discard (ulong)
parm: zvol_prefetch_bytes:Prefetch N bytes at zvol start+end (uint)
parm: zvol_volmode:Default volmode property value (uint)
parm: zvol_blk_mq_queue_depth:Default blk-mq queue depth (uint)
parm: zvol_use_blk_mq:Use the blk-mq API for zvols (uint)
parm: zvol_blk_mq_blocks_per_thread:Process volblocksize blocks per thread (uint)
parm: zfs_xattr_compat:Use legacy ZFS xattr naming for writing new user namespace xattrs
parm: zfs_fallocate_reserve_percent:Percentage of length to use for the available capacity check (uint)
parm: zfs_key_max_salt_uses:Max number of times a salt value can be used for generating encryption keys before it is rotated (ulong)
parm: zfs_object_mutex_size:Size of znode hold array (uint)
parm: zfs_unlink_suspend_progress:Set to prevent async unlinks (debug - leaks space into the unlinked set) (int)
parm: zfs_delete_blocks:Delete files larger than N blocks async (ulong)
parm: zfs_dbgmsg_enable:Enable ZFS debug message log (int)
parm: zfs_dbgmsg_maxsize:Maximum ZFS debug log size (uint)
parm: zfs_admin_snapshot:Enable mkdir/rmdir/mv in .zfs/snapshot (int)
parm: zfs_expire_snapshot:Seconds to expire .zfs/snapshot (int)
parm: vdev_file_logical_ashift:Logical ashift for file-based devices
parm: vdev_file_physical_ashift:Physical ashift for file-based devices
parm: zfs_vdev_scheduler:I/O scheduler
parm: zfs_vdev_open_timeout_ms:Timeout before determining that a device is missing
parm: zfs_vdev_failfast_mask:Defines failfast mask: 1 - device, 2 - transport, 4 - driver
parm: zfs_arc_shrinker_limit:Limit on number of pages that ARC shrinker can reclaim at once
parm: zfs_abd_scatter_enabled:Toggle whether ABD allocations must be linear. (int)
parm: zfs_abd_scatter_min_size:Minimum size of scatter allocations. (int)
parm: zfs_abd_scatter_max_order:Maximum order allocation used for a scatter ABD. (uint)
parm: zio_slow_io_ms:Max I/O completion time (milliseconds) before marking it as slow
parm: zio_requeue_io_start_cut_in_line:Prioritize requeued I/O
parm: zfs_sync_pass_deferred_free:Defer frees starting in this pass
parm: zfs_sync_pass_dont_compress:Don't compress starting in this pass
parm: zfs_sync_pass_rewrite:Rewrite new bps starting in this pass
parm: zio_dva_throttle_enabled:Throttle block allocations in the ZIO pipeline
parm: zio_deadman_log_all:Log all slow ZIOs, not just those with vdevs
parm: zfs_commit_timeout_pct:ZIL block open timeout percentage
parm: zil_min_commit_timeout:Minimum delay we care for ZIL block commit
parm: zil_replay_disable:Disable intent logging replay
parm: zil_nocacheflush:Disable ZIL cache flushes
parm: zil_slog_bulk:Limit in bytes slog sync writes per commit
parm: zil_maxblocksize:Limit in bytes of ZIL log block size
parm: zfs_vnops_read_chunk_size:Bytes to read per chunk
parm: zfs_zil_saxattr:Disable xattr=sa extended attribute logging in ZIL by settng 0.
parm: zfs_immediate_write_sz:Largest data block to write to zil
parm: zfs_max_nvlist_src_size:Maximum size in bytes allowed for src nvlist passed with ZFS ioctls
parm: zfs_history_output_max:Maximum size in bytes of ZFS ioctl output that will be logged
parm: zfs_zevent_retain_max:Maximum recent zevents records to retain for duplicate checking
parm: zfs_zevent_retain_expire_secs:Expiration time for recent zevents records
parm: zfs_lua_max_instrlimit:Max instruction limit that can be specified for a channel program
parm: zfs_lua_max_memlimit:Max memory limit that can be specified for a channel program
parm: zap_micro_max_size:Maximum micro ZAP size, before converting to a fat ZAP, in bytes
parm: zap_iterate_prefetch:When iterating ZAP object, prefetch it
parm: zfs_trim_extent_bytes_max:Max size of TRIM commands, larger will be split
parm: zfs_trim_extent_bytes_min:Min size of TRIM commands, smaller will be skipped
parm: zfs_trim_metaslab_skip:Skip metaslabs which have never been initialized
parm: zfs_trim_txg_batch:Min number of txgs to aggregate frees before issuing TRIM
parm: zfs_trim_queue_limit:Max queued TRIMs outstanding per leaf vdev
parm: zfs_removal_ignore_errors:Ignore hard IO errors when removing device
parm: zfs_remove_max_segment:Largest contiguous segment to allocate when removing device
parm: vdev_removal_max_span:Largest span of free chunks a remap segment can span
parm: zfs_removal_suspend_progress:Pause device removal after this many bytes are copied (debug use only - causes removal to hang)
parm: zfs_rebuild_max_segment:Max segment size in bytes of rebuild reads
parm: zfs_rebuild_vdev_limit:Max bytes in flight per leaf vdev for sequential resilvers
parm: zfs_rebuild_scrub_enabled:Automatically scrub after sequential resilver completes
parm: zfs_vdev_raidz_impl:Select raidz implementation.
parm: zfs_vdev_aggregation_limit:Max vdev I/O aggregation size
parm: zfs_vdev_aggregation_limit_non_rotating:Max vdev I/O aggregation size for non-rotating media
parm: zfs_vdev_aggregate_trim:Allow TRIM I/O to be aggregated
parm: zfs_vdev_read_gap_limit:Aggregate read I/O over gap
parm: zfs_vdev_write_gap_limit:Aggregate write I/O over gap
parm: zfs_vdev_max_active:Maximum number of active I/Os per vdev
parm: zfs_vdev_async_write_active_max_dirty_percent:Async write concurrency max threshold
parm: zfs_vdev_async_write_active_min_dirty_percent:Async write concurrency min threshold
parm: zfs_vdev_async_read_max_active:Max active async read I/Os per vdev
parm: zfs_vdev_async_read_min_active:Min active async read I/Os per vdev
parm: zfs_vdev_async_write_max_active:Max active async write I/Os per vdev
parm: zfs_vdev_async_write_min_active:Min active async write I/Os per vdev
parm: zfs_vdev_initializing_max_active:Max active initializing I/Os per vdev
parm: zfs_vdev_initializing_min_active:Min active initializing I/Os per vdev
parm: zfs_vdev_removal_max_active:Max active removal I/Os per vdev
parm: zfs_vdev_removal_min_active:Min active removal I/Os per vdev
parm: zfs_vdev_scrub_max_active:Max active scrub I/Os per vdev
parm: zfs_vdev_scrub_min_active:Min active scrub I/Os per vdev
parm: zfs_vdev_sync_read_max_active:Max active sync read I/Os per vdev
parm: zfs_vdev_sync_read_min_active:Min active sync read I/Os per vdev
parm: zfs_vdev_sync_write_max_active:Max active sync write I/Os per vdev
parm: zfs_vdev_sync_write_min_active:Min active sync write I/Os per vdev
parm: zfs_vdev_trim_max_active:Max active trim/discard I/Os per vdev
parm: zfs_vdev_trim_min_active:Min active trim/discard I/Os per vdev
parm: zfs_vdev_rebuild_max_active:Max active rebuild I/Os per vdev
parm: zfs_vdev_rebuild_min_active:Min active rebuild I/Os per vdev
parm: zfs_vdev_nia_credit:Number of non-interactive I/Os to allow in sequence
parm: zfs_vdev_nia_delay:Number of non-interactive I/Os before _max_active
parm: zfs_vdev_queue_depth_pct:Queue depth percentage for each top-level vdev
parm: zfs_vdev_mirror_rotating_inc:Rotating media load increment for non-seeking I/Os
parm: zfs_vdev_mirror_rotating_seek_inc:Rotating media load increment for seeking I/Os
parm: zfs_vdev_mirror_rotating_seek_offset:Offset in bytes from the last I/O which triggers a reduced rotating media seek increment
parm: zfs_vdev_mirror_non_rotating_inc:Non-rotating media load increment for non-seeking I/Os
parm: zfs_vdev_mirror_non_rotating_seek_inc:Non-rotating media load increment for seeking I/Os
parm: zfs_initialize_value:Value written during zpool initialize
parm: zfs_initialize_chunk_size:Size in bytes of writes by zpool initialize
parm: zfs_condense_indirect_vdevs_enable:Whether to attempt condensing indirect vdev mappings
parm: zfs_condense_indirect_obsolete_pct:Minimum obsolete percent of bytes in the mapping to attempt condensing
parm: zfs_condense_min_mapping_bytes:Don't bother condensing if the mapping uses less than this amount of memory
parm: zfs_condense_max_obsolete_bytes:Minimum size obsolete spacemap to attempt condensing
parm: zfs_condense_indirect_commit_entry_delay_ms:Used by tests to ensure certain actions happen in the middle of a condense. A maximum value of 1 should be sufficient.
parm: zfs_reconstruct_indirect_combinations_max:Maximum number of combinations when reconstructing split segments
parm: zfs_vdev_cache_max:Inflate reads small than max
parm: zfs_vdev_cache_size:Total size of the per-disk cache
parm: zfs_vdev_cache_bshift:Shift size to inflate reads too
parm: zfs_vdev_default_ms_count:Target number of metaslabs per top-level vdev
parm: zfs_vdev_default_ms_shift:Default limit for metaslab size
parm: zfs_vdev_min_ms_count:Minimum number of metaslabs per top-level vdev
parm: zfs_vdev_ms_count_limit:Practical upper limit of total metaslabs per top-level vdev
parm: zfs_slow_io_events_per_second:Rate limit slow IO (delay) events to this many per second
parm: zfs_checksum_events_per_second:Rate limit checksum events to this many checksum errors per second (do not set below ZED threshold).
parm: zfs_scan_ignore_errors:Ignore errors during resilver/scrub
parm: vdev_validate_skip:Bypass vdev_validate()
parm: zfs_nocacheflush:Disable cache flushes
parm: zfs_embedded_slog_min_ms:Minimum number of metaslabs required to dedicate one for log blocks
parm: zfs_vdev_min_auto_ashift:Minimum ashift used when creating new top-level vdevs
parm: zfs_vdev_max_auto_ashift:Maximum ashift used when optimizing for logical -> physical sector size on new top-level vdevs
parm: zfs_txg_timeout:Max seconds worth of delta per txg
parm: zfs_read_history:Historical statistics for the last N reads
parm: zfs_read_history_hits:Include cache hits in read history
parm: zfs_txg_history:Historical statistics for the last N txgs
parm: zfs_multihost_history:Historical statistics for last N multihost writes
parm: zfs_flags:Set additional debugging flags
parm: zfs_recover:Set to attempt to recover from fatal errors
parm: zfs_free_leak_on_eio:Set to ignore IO errors during free and permanently leak the space
parm: zfs_deadman_checktime_ms:Dead I/O check interval in milliseconds
parm: zfs_deadman_enabled:Enable deadman timer
parm: spa_asize_inflation:SPA size estimate multiplication factor
parm: zfs_ddt_data_is_special:Place DDT data into the special class
parm: zfs_user_indirect_is_special:Place user data indirect blocks into the special class
parm: zfs_deadman_failmode:Failmode for deadman timer
parm: zfs_deadman_synctime_ms:Pool sync expiration time in milliseconds
parm: zfs_deadman_ziotime_ms:IO expiration time in milliseconds
parm: zfs_special_class_metadata_reserve_pct:Small file blocks in special vdevs depends on this much free space available
parm: spa_slop_shift:Reserved free space in pool
parm: zfs_unflushed_max_mem_amt:Specific hard-limit in memory that ZFS allows to be used for unflushed changes
parm: zfs_unflushed_max_mem_ppm:Percentage of the overall system memory that ZFS allows to be used for unflushed changes (value is calculated over 1000000 for finer granularity)
parm: zfs_unflushed_log_block_max:Hard limit (upper-bound) in the size of the space map log in terms of blocks.
parm: zfs_unflushed_log_block_min:Lower-bound limit for the maximum amount of blocks allowed in log spacemap (see zfs_unflushed_log_block_max)
parm: zfs_unflushed_log_txg_max:Hard limit (upper-bound) in the size of the space map log in terms of dirty TXGs.
parm: zfs_unflushed_log_block_pct:Tunable used to determine the number of blocks that can be used for the spacemap log, expressed as a percentage of the total number of metaslabs in the pool (e.g. 400 means the number of log blocks is capped at 4 times the number of metaslabs)
parm: zfs_max_log_walking:The number of past TXGs that the flushing algorithm of the log spacemap feature uses to estimate incoming log blocks
parm: zfs_keep_log_spacemaps_at_export:Prevent the log spacemaps from being flushed and destroyed during pool export/destroy
parm: zfs_max_logsm_summary_length:Maximum number of rows allowed in the summary of the spacemap log
parm: zfs_min_metaslabs_to_flush:Minimum number of metaslabs to flush per dirty TXG
parm: spa_upgrade_errlog_limit:Limit the number of errors which will be upgraded to the new on-disk error log when enabling head_errlog
parm: spa_config_path:SPA config file (/etc/zfs/zpool.cache)
parm: zfs_autoimport_disable:Disable pool import at module load
parm: zfs_spa_discard_memory_limit:Limit for memory used in prefetching the checkpoint space map done on each vdev while discarding the checkpoint
parm: spa_load_verify_shift:log2 fraction of arc that can be used by inflight I/Os when verifying pool during import
parm: spa_load_verify_metadata:Set to traverse metadata on pool import
parm: spa_load_verify_data:Set to traverse data on pool import
parm: spa_load_print_vdev_tree:Print vdev tree to zfs_dbgmsg during pool import
parm: zio_taskq_batch_pct:Percentage of CPUs to run an IO worker thread
parm: zio_taskq_batch_tpq:Number of threads per IO worker taskqueue
parm: zfs_max_missing_tvds:Allow importing pool with up to this number of missing top-level vdevs (in read-only mode)
parm: zfs_livelist_condense_zthr_pause:Set the livelist condense zthr to pause
parm: zfs_livelist_condense_sync_pause:Set the livelist condense synctask to pause
parm: zfs_livelist_condense_sync_cancel:Whether livelist condensing was canceled in the synctask
parm: zfs_livelist_condense_zthr_cancel:Whether livelist condensing was canceled in the zthr function
parm: zfs_livelist_condense_new_alloc:Whether extra ALLOC blkptrs were added to a livelist entry while it was being condensed
parm: zfs_multilist_num_sublists:Number of sublists used in each multilist
parm: zfs_multihost_interval:Milliseconds between mmp writes to each leaf
parm: zfs_multihost_fail_intervals:Max allowed period without a successful mmp write
parm: zfs_multihost_import_intervals:Number of zfs_multihost_interval periods to wait for activity
parm: metaslab_aliquot:Allocation granularity (a.k.a. stripe size)
parm: metaslab_debug_load:Load all metaslabs when pool is first opened
parm: metaslab_debug_unload:Prevent metaslabs from being unloaded
parm: metaslab_preload_enabled:Preload potential metaslabs during reassessment
parm: metaslab_unload_delay:Delay in txgs after metaslab was last used before unloading
parm: metaslab_unload_delay_ms:Delay in milliseconds after metaslab was last used before unloading
parm: zfs_mg_noalloc_threshold:Percentage of metaslab group size that should be free to make it eligible for allocation
parm: zfs_mg_fragmentation_threshold:Percentage of metaslab group size that should be considered eligible for allocations unless all metaslab groups within the metaslab class have also crossed this threshold
parm: metaslab_fragmentation_factor_enabled:Use the fragmentation metric to prefer less fragmented metaslabs
parm: zfs_metaslab_fragmentation_threshold:Fragmentation for metaslab to allow allocation
parm: metaslab_lba_weighting_enabled:Prefer metaslabs with lower LBAs
parm: metaslab_bias_enabled:Enable metaslab group biasing
parm: zfs_metaslab_segment_weight_enabled:Enable segment-based metaslab selection
parm: zfs_metaslab_switch_threshold:Segment-based metaslab selection maximum buckets before switching
parm: metaslab_force_ganging:Blocks larger than this size are forced to be gang blocks
parm: metaslab_df_max_search:Max distance (bytes) to search forward before using size tree
parm: metaslab_df_use_largest_segment:When looking in size tree, use largest segment instead of exact fit
parm: zfs_metaslab_max_size_cache_sec:How long to trust the cached max chunk size of a metaslab
parm: zfs_metaslab_mem_limit:Percentage of memory that can be used to store metaslab range trees
parm: zfs_metaslab_try_hard_before_gang:Try hard to allocate before ganging
parm: zfs_metaslab_find_max_tries:Normally only consider this many of the best metaslabs in each vdev
parm: zfs_zevent_len_max:Max event queue length
parm: zfs_scan_vdev_limit:Max bytes in flight per leaf vdev for scrubs and resilvers
parm: zfs_scrub_min_time_ms:Min millisecs to scrub per txg
parm: zfs_obsolete_min_time_ms:Min millisecs to obsolete per txg
parm: zfs_free_min_time_ms:Min millisecs to free per txg
parm: zfs_resilver_min_time_ms:Min millisecs to resilver per txg
parm: zfs_scan_suspend_progress:Set to prevent scans from progressing
parm: zfs_no_scrub_io:Set to disable scrub I/O
parm: zfs_no_scrub_prefetch:Set to disable scrub prefetching
parm: zfs_async_block_max_blocks:Max number of blocks freed in one txg
parm: zfs_max_async_dedup_frees:Max number of dedup blocks freed in one txg
parm: zfs_free_bpobj_enabled:Enable processing of the free_bpobj
parm: zfs_scan_blkstats:Enable block statistics calculation during scrub
parm: zfs_scan_mem_lim_fact:Fraction of RAM for scan hard limit
parm: zfs_scan_issue_strategy:IO issuing strategy during scrubbing. 0 = default, 1 = LBA, 2 = size
parm: zfs_scan_legacy:Scrub using legacy non-sequential method
parm: zfs_scan_checkpoint_intval:Scan progress on-disk checkpointing interval
parm: zfs_scan_max_ext_gap:Max gap in bytes between sequential scrub / resilver I/Os
parm: zfs_scan_mem_lim_soft_fact:Fraction of hard limit used as soft limit
parm: zfs_scan_strict_mem_lim:Tunable to attempt to reduce lock contention
parm: zfs_scan_fill_weight:Tunable to adjust bias towards more filled segments during scans
parm: zfs_scan_report_txgs:Tunable to report resilver performance over the last N txgs
parm: zfs_resilver_disable_defer:Process all resilvers immediately
parm: zfs_dirty_data_max_percent:Max percent of RAM allowed to be dirty
parm: zfs_dirty_data_max_max_percent:zfs_dirty_data_max upper bound as % of RAM
parm: zfs_delay_min_dirty_percent:Transaction delay threshold
parm: zfs_dirty_data_max:Determines the dirty space limit
parm: zfs_wrlog_data_max:The size limit of write-transaction zil log data
parm: zfs_dirty_data_max_max:zfs_dirty_data_max upper bound in bytes
parm: zfs_dirty_data_sync_percent:Dirty data txg sync threshold as a percentage of zfs_dirty_data_max
parm: zfs_delay_scale:How quickly delay approaches infinity
parm: zfs_sync_taskq_batch_pct:Max percent of CPUs that are used to sync dirty data
parm: zfs_zil_clean_taskq_nthr_pct:Max percent of CPUs that are used per dp_sync_taskq
parm: zfs_zil_clean_taskq_minalloc:Number of taskq entries that are pre-populated
parm: zfs_zil_clean_taskq_maxalloc:Max number of taskq entries that are cached
parm: zvol_enforce_quotas:Enable strict ZVOL quota enforcment
parm: zfs_livelist_max_entries:Size to start the next sub-livelist in a livelist
parm: zfs_livelist_min_percent_shared:Threshold at which livelist is disabled
parm: zfs_max_recordsize:Max allowed record size
parm: zfs_allow_redacted_dataset_mount:Allow mounting of redacted datasets
parm: zfs_snapshot_history_enabled:Include snapshot events in pool history/events
parm: zfs_disable_ivset_guid_check:Set to allow raw receives without IVset guids
parm: zfs_default_bs:Default dnode block shift
parm: zfs_default_ibs:Default dnode indirect block shift
parm: zfs_prefetch_disable:Disable all ZFS prefetching
parm: zfetch_max_streams:Max number of streams per zfetch
parm: zfetch_min_sec_reap:Min time before stream reclaim
parm: zfetch_max_sec_reap:Max time before stream delete
parm: zfetch_min_distance:Min bytes to prefetch per stream
parm: zfetch_max_distance:Max bytes to prefetch per stream
parm: zfetch_max_idistance:Max bytes to prefetch indirects for per stream
parm: zfetch_array_rd_sz:Number of bytes in a array_read
parm: zfs_pd_bytes_max:Max number of bytes to prefetch
parm: zfs_traverse_indirect_prefetch_limit:Traverse prefetch number of blocks pointed by indirect block
parm: ignore_hole_birth:Alias for send_holes_without_birth_time (int)
parm: send_holes_without_birth_time:Ignore hole_birth txg for zfs send
parm: zfs_send_corrupt_data:Allow sending corrupt data
parm: zfs_send_queue_length:Maximum send queue length
parm: zfs_send_unmodified_spill_blocks:Send unmodified spill blocks
parm: zfs_send_no_prefetch_queue_length:Maximum send queue length for non-prefetch queues
parm: zfs_send_queue_ff:Send queue fill fraction
parm: zfs_send_no_prefetch_queue_ff:Send queue fill fraction for non-prefetch queues
parm: zfs_override_estimate_recordsize:Override block size estimate with fixed size
parm: zfs_recv_queue_length:Maximum receive queue length
parm: zfs_recv_queue_ff:Receive queue fill fraction
parm: zfs_recv_write_batch_size:Maximum amount of writes to batch into one transaction
parm: zfs_recv_best_effort_corrective:Ignore errors during corrective receive
parm: dmu_object_alloc_chunk_shift:CPU-specific allocator grabs 2^N objects at once
parm: zfs_nopwrite_enabled:Enable NOP writes
parm: zfs_per_txg_dirty_frees_percent:Percentage of dirtied blocks from frees in one TXG
parm: zfs_dmu_offset_next_sync:Enable forcing txg sync to find holes
parm: dmu_prefetch_max:Limit one prefetch call to this size
parm: zfs_dedup_prefetch:Enable prefetching dedup-ed blks
parm: zfs_dbuf_state_index:Calculate arc header index
parm: dbuf_cache_max_bytes:Maximum size in bytes of the dbuf cache.
parm: dbuf_cache_hiwater_pct:Percentage over dbuf_cache_max_bytes for direct dbuf eviction.
parm: dbuf_cache_lowater_pct:Percentage below dbuf_cache_max_bytes when dbuf eviction stops.
parm: dbuf_metadata_cache_max_bytes:Maximum size in bytes of dbuf metadata cache.
parm: dbuf_cache_shift:Set size of dbuf cache to log2 fraction of arc size.
parm: dbuf_metadata_cache_shift:Set size of dbuf metadata cache to log2 fraction of arc size.
parm: dbuf_mutex_cache_shift:Set size of dbuf cache mutex array as log2 shift.
parm: zfs_btree_verify_intensity:Enable btree verification. Levels above 4 require ZFS be built with debugging
parm: zfs_arc_min:Minimum ARC size in bytes
parm: zfs_arc_max:Maximum ARC size in bytes
parm: zfs_arc_meta_limit:Metadata limit for ARC size in bytes
parm: zfs_arc_meta_limit_percent:Percent of ARC size for ARC meta limit
parm: zfs_arc_meta_min:Minimum ARC metadata size in bytes
parm: zfs_arc_meta_prune:Meta objects to scan for prune
parm: zfs_arc_meta_adjust_restarts:Limit number of restarts in arc_evict_meta
parm: zfs_arc_meta_strategy:Meta reclaim strategy
parm: zfs_arc_grow_retry:Seconds before growing ARC size
parm: zfs_arc_p_dampener_disable:Disable arc_p adapt dampener
parm: zfs_arc_shrink_shift:log2(fraction of ARC to reclaim)
parm: zfs_arc_pc_percent:Percent of pagecache to reclaim ARC to
parm: zfs_arc_p_min_shift:arc_c shift to calc min/max arc_p
parm: zfs_arc_average_blocksize:Target average block size
parm: zfs_compressed_arc_enabled:Disable compressed ARC buffers
parm: zfs_arc_min_prefetch_ms:Min life of prefetch block in ms
parm: zfs_arc_min_prescient_prefetch_ms:Min life of prescient prefetched block in ms
parm: l2arc_write_max:Max write bytes per interval
parm: l2arc_write_boost:Extra write bytes during device warmup
parm: l2arc_headroom:Number of max device writes to precache
parm: l2arc_headroom_boost:Compressed l2arc_headroom multiplier
parm: l2arc_trim_ahead:TRIM ahead L2ARC write size multiplier
parm: l2arc_feed_secs:Seconds between L2ARC writing
parm: l2arc_feed_min_ms:Min feed interval in milliseconds
parm: l2arc_noprefetch:Skip caching prefetched buffers
parm: l2arc_feed_again:Turbo L2ARC warmup
parm: l2arc_norw:No reads during writes
parm: l2arc_meta_percent:Percent of ARC size allowed for L2ARC-only headers
parm: l2arc_rebuild_enabled:Rebuild the L2ARC when importing a pool
parm: l2arc_rebuild_blocks_min_l2size:Min size in bytes to write rebuild log blocks in L2ARC
parm: l2arc_mfuonly:Cache only MFU data from ARC into L2ARC
parm: l2arc_exclude_special:Exclude dbufs on special vdevs from being cached to L2ARC if set.
parm: zfs_arc_lotsfree_percent:System free memory I/O throttle in bytes
parm: zfs_arc_sys_free:System free memory target size in bytes
parm: zfs_arc_dnode_limit:Minimum bytes of dnodes in ARC
parm: zfs_arc_dnode_limit_percent:Percent of ARC meta buffers for dnodes
parm: zfs_arc_dnode_reduce_percent:Percentage of excess dnodes to try to unpin
parm: zfs_arc_eviction_pct:When full, ARC allocation waits for eviction of this % of alloc size
parm: zfs_arc_evict_batch_limit:The number of headers to evict per sublist before moving to the next
parm: zfs_arc_prune_task_threads:Number of arc_prune threads
parm: zstd_earlyabort_pass:Enable early abort attempts when using zstd
parm: zstd_abort_size:Minimal size of block to attempt early abort
parm: zfs_max_dataset_nesting:Limit to the amount of nesting a path can have. Defaults to 50.
parm: zfs_fletcher_4_impl:Select fletcher 4 implementation.
parm: icp_gcm_impl:Select gcm implementation.
parm: icp_gcm_avx_chunk_size:How many bytes to process while owning the FPU
parm: zfs_blake3_impl:Select BLAKE3 implementation.
parm: icp_aes_impl:Select aes implementation.
```
|
1.0
|
The zfs kernel module fails to be installed: ERROR: could not insert 'zfs': Key was rejected by service - ### System information
Distribution Name | Fedora Xfce Desktop
Distribution Version | fc37
Kernel Version | 6.1.12-200.fc37.x86_64
Architecture | x86_64
OpenZFS Version | zfs-2.1.99-1740_gd9e64a403
### Describe the problem you're observing
I tried to install the zfs module, but modprobe told me to reject the installation.
### Describe how to reproduce the problem
I tried using the zfs package provided by zfsonlinux.org and the ZFS RPM package built from GIT SCM, and they both had this problem. They seem to be related to a feature of the kernel module, What do I need to do to make it work?
Thanks.
```
[gesangtome@fedora ~]$ sudo modprobe zfs
[sudo] password for gesangtome:
modprobe: ERROR: could not insert 'zfs': Key was rejected by service
```
modinfo for zfs.ko
```
[gesangtome@fedora ~]$ modinfo zfs
filename: /lib/modules/6.1.12-200.fc37.x86_64/extra/zfs.ko.xz
version: 2.1.99-1740_gd9e64a403
license: CDDL
license: Dual BSD/GPL
license: Dual MIT/GPL
author: OpenZFS
description: ZFS
alias: zzstd
alias: zcommon
alias: zunicode
alias: znvpair
alias: zlua
alias: icp
alias: zavl
alias: devname:zfs
alias: char-major-10-249
srcversion: 13B813800C43864C4EF9B62
depends: spl
retpoline: Y
name: zfs
vermagic: 6.1.12-200.fc37.x86_64 SMP preempt mod_unload
sig_id: PKCS#7
signer: DKMS module signing key
sig_key: 1E:AD:B5:32:A6:8B:C0:D2:53:BE:D3:F9:AC:4C:DA:51:68:21:93:0E
sig_hashalgo: sha512
signature: 29:82:06:D1:87:1F:21:30:47:02:35:61:F0:DF:F2:E6:67:24:96:B2:
57:82:37:13:D4:12:EE:94:67:C8:87:AA:D9:B6:28:0B:7C:22:D6:21:
5C:4B:19:31:F8:FB:9E:4B:56:7F:D6:82:18:80:8F:7D:E8:FC:E0:FB:
2F:FD:48:80:0F:C6:6A:C3:3D:4B:90:B5:59:6E:10:24:56:81:A3:2A:
F8:55:BB:67:36:DE:05:25:FF:E2:AF:12:E6:9D:22:CA:BB:1E:8F:86:
AF:26:DD:A3:F3:3D:34:EC:2A:4D:2D:C5:AA:38:6A:97:66:EF:C0:09:
8B:3D:2F:EA:2B:C1:54:6B:E7:64:0C:41:01:40:23:F6:EC:4C:65:D9:
D9:BE:5E:6D:14:23:DA:BC:15:32:BE:D9:2D:67:D4:D6:6F:B0:6C:F8:
3A:B5:71:20:69:56:F5:FE:28:FE:82:FE:64:65:37:ED:A3:51:62:BA:
00:72:76:0A:24:E1:1B:D7:7A:76:73:3D:71:47:BD:97:2C:7C:6E:06:
F6:43:E2:25:43:AF:B3:01:7D:99:B5:9E:C3:D2:9F:95:3F:1C:1B:C6:
A5:63:40:3C:E9:82:18:4C:C7:8C:27:83:98:63:D4:6B:7F:B9:25:27:
A5:C3:D9:D5:BC:24:66:39:6F:07:E2:20:BC:C8:9B:B7
parm: zvol_inhibit_dev:Do not create zvol device nodes (uint)
parm: zvol_major:Major number for zvol device (uint)
parm: zvol_threads:Number of threads to handle I/O requests. Setto 0 to use all active CPUs (uint)
parm: zvol_request_sync:Synchronously handle bio requests (uint)
parm: zvol_max_discard_blocks:Max number of blocks to discard (ulong)
parm: zvol_prefetch_bytes:Prefetch N bytes at zvol start+end (uint)
parm: zvol_volmode:Default volmode property value (uint)
parm: zvol_blk_mq_queue_depth:Default blk-mq queue depth (uint)
parm: zvol_use_blk_mq:Use the blk-mq API for zvols (uint)
parm: zvol_blk_mq_blocks_per_thread:Process volblocksize blocks per thread (uint)
parm: zfs_xattr_compat:Use legacy ZFS xattr naming for writing new user namespace xattrs
parm: zfs_fallocate_reserve_percent:Percentage of length to use for the available capacity check (uint)
parm: zfs_key_max_salt_uses:Max number of times a salt value can be used for generating encryption keys before it is rotated (ulong)
parm: zfs_object_mutex_size:Size of znode hold array (uint)
parm: zfs_unlink_suspend_progress:Set to prevent async unlinks (debug - leaks space into the unlinked set) (int)
parm: zfs_delete_blocks:Delete files larger than N blocks async (ulong)
parm: zfs_dbgmsg_enable:Enable ZFS debug message log (int)
parm: zfs_dbgmsg_maxsize:Maximum ZFS debug log size (uint)
parm: zfs_admin_snapshot:Enable mkdir/rmdir/mv in .zfs/snapshot (int)
parm: zfs_expire_snapshot:Seconds to expire .zfs/snapshot (int)
parm: vdev_file_logical_ashift:Logical ashift for file-based devices
parm: vdev_file_physical_ashift:Physical ashift for file-based devices
parm: zfs_vdev_scheduler:I/O scheduler
parm: zfs_vdev_open_timeout_ms:Timeout before determining that a device is missing
parm: zfs_vdev_failfast_mask:Defines failfast mask: 1 - device, 2 - transport, 4 - driver
parm: zfs_arc_shrinker_limit:Limit on number of pages that ARC shrinker can reclaim at once
parm: zfs_abd_scatter_enabled:Toggle whether ABD allocations must be linear. (int)
parm: zfs_abd_scatter_min_size:Minimum size of scatter allocations. (int)
parm: zfs_abd_scatter_max_order:Maximum order allocation used for a scatter ABD. (uint)
parm: zio_slow_io_ms:Max I/O completion time (milliseconds) before marking it as slow
parm: zio_requeue_io_start_cut_in_line:Prioritize requeued I/O
parm: zfs_sync_pass_deferred_free:Defer frees starting in this pass
parm: zfs_sync_pass_dont_compress:Don't compress starting in this pass
parm: zfs_sync_pass_rewrite:Rewrite new bps starting in this pass
parm: zio_dva_throttle_enabled:Throttle block allocations in the ZIO pipeline
parm: zio_deadman_log_all:Log all slow ZIOs, not just those with vdevs
parm: zfs_commit_timeout_pct:ZIL block open timeout percentage
parm: zil_min_commit_timeout:Minimum delay we care for ZIL block commit
parm: zil_replay_disable:Disable intent logging replay
parm: zil_nocacheflush:Disable ZIL cache flushes
parm: zil_slog_bulk:Limit in bytes slog sync writes per commit
parm: zil_maxblocksize:Limit in bytes of ZIL log block size
parm: zfs_vnops_read_chunk_size:Bytes to read per chunk
parm: zfs_zil_saxattr:Disable xattr=sa extended attribute logging in ZIL by settng 0.
parm: zfs_immediate_write_sz:Largest data block to write to zil
parm: zfs_max_nvlist_src_size:Maximum size in bytes allowed for src nvlist passed with ZFS ioctls
parm: zfs_history_output_max:Maximum size in bytes of ZFS ioctl output that will be logged
parm: zfs_zevent_retain_max:Maximum recent zevents records to retain for duplicate checking
parm: zfs_zevent_retain_expire_secs:Expiration time for recent zevents records
parm: zfs_lua_max_instrlimit:Max instruction limit that can be specified for a channel program
parm: zfs_lua_max_memlimit:Max memory limit that can be specified for a channel program
parm: zap_micro_max_size:Maximum micro ZAP size, before converting to a fat ZAP, in bytes
parm: zap_iterate_prefetch:When iterating ZAP object, prefetch it
parm: zfs_trim_extent_bytes_max:Max size of TRIM commands, larger will be split
parm: zfs_trim_extent_bytes_min:Min size of TRIM commands, smaller will be skipped
parm: zfs_trim_metaslab_skip:Skip metaslabs which have never been initialized
parm: zfs_trim_txg_batch:Min number of txgs to aggregate frees before issuing TRIM
parm: zfs_trim_queue_limit:Max queued TRIMs outstanding per leaf vdev
parm: zfs_removal_ignore_errors:Ignore hard IO errors when removing device
parm: zfs_remove_max_segment:Largest contiguous segment to allocate when removing device
parm: vdev_removal_max_span:Largest span of free chunks a remap segment can span
parm: zfs_removal_suspend_progress:Pause device removal after this many bytes are copied (debug use only - causes removal to hang)
parm: zfs_rebuild_max_segment:Max segment size in bytes of rebuild reads
parm: zfs_rebuild_vdev_limit:Max bytes in flight per leaf vdev for sequential resilvers
parm: zfs_rebuild_scrub_enabled:Automatically scrub after sequential resilver completes
parm: zfs_vdev_raidz_impl:Select raidz implementation.
parm: zfs_vdev_aggregation_limit:Max vdev I/O aggregation size
parm: zfs_vdev_aggregation_limit_non_rotating:Max vdev I/O aggregation size for non-rotating media
parm: zfs_vdev_aggregate_trim:Allow TRIM I/O to be aggregated
parm: zfs_vdev_read_gap_limit:Aggregate read I/O over gap
parm: zfs_vdev_write_gap_limit:Aggregate write I/O over gap
parm: zfs_vdev_max_active:Maximum number of active I/Os per vdev
parm: zfs_vdev_async_write_active_max_dirty_percent:Async write concurrency max threshold
parm: zfs_vdev_async_write_active_min_dirty_percent:Async write concurrency min threshold
parm: zfs_vdev_async_read_max_active:Max active async read I/Os per vdev
parm: zfs_vdev_async_read_min_active:Min active async read I/Os per vdev
parm: zfs_vdev_async_write_max_active:Max active async write I/Os per vdev
parm: zfs_vdev_async_write_min_active:Min active async write I/Os per vdev
parm: zfs_vdev_initializing_max_active:Max active initializing I/Os per vdev
parm: zfs_vdev_initializing_min_active:Min active initializing I/Os per vdev
parm: zfs_vdev_removal_max_active:Max active removal I/Os per vdev
parm: zfs_vdev_removal_min_active:Min active removal I/Os per vdev
parm: zfs_vdev_scrub_max_active:Max active scrub I/Os per vdev
parm: zfs_vdev_scrub_min_active:Min active scrub I/Os per vdev
parm: zfs_vdev_sync_read_max_active:Max active sync read I/Os per vdev
parm: zfs_vdev_sync_read_min_active:Min active sync read I/Os per vdev
parm: zfs_vdev_sync_write_max_active:Max active sync write I/Os per vdev
parm: zfs_vdev_sync_write_min_active:Min active sync write I/Os per vdev
parm: zfs_vdev_trim_max_active:Max active trim/discard I/Os per vdev
parm: zfs_vdev_trim_min_active:Min active trim/discard I/Os per vdev
parm: zfs_vdev_rebuild_max_active:Max active rebuild I/Os per vdev
parm: zfs_vdev_rebuild_min_active:Min active rebuild I/Os per vdev
parm: zfs_vdev_nia_credit:Number of non-interactive I/Os to allow in sequence
parm: zfs_vdev_nia_delay:Number of non-interactive I/Os before _max_active
parm: zfs_vdev_queue_depth_pct:Queue depth percentage for each top-level vdev
parm: zfs_vdev_mirror_rotating_inc:Rotating media load increment for non-seeking I/Os
parm: zfs_vdev_mirror_rotating_seek_inc:Rotating media load increment for seeking I/Os
parm: zfs_vdev_mirror_rotating_seek_offset:Offset in bytes from the last I/O which triggers a reduced rotating media seek increment
parm: zfs_vdev_mirror_non_rotating_inc:Non-rotating media load increment for non-seeking I/Os
parm: zfs_vdev_mirror_non_rotating_seek_inc:Non-rotating media load increment for seeking I/Os
parm: zfs_initialize_value:Value written during zpool initialize
parm: zfs_initialize_chunk_size:Size in bytes of writes by zpool initialize
parm: zfs_condense_indirect_vdevs_enable:Whether to attempt condensing indirect vdev mappings
parm: zfs_condense_indirect_obsolete_pct:Minimum obsolete percent of bytes in the mapping to attempt condensing
parm: zfs_condense_min_mapping_bytes:Don't bother condensing if the mapping uses less than this amount of memory
parm: zfs_condense_max_obsolete_bytes:Minimum size obsolete spacemap to attempt condensing
parm: zfs_condense_indirect_commit_entry_delay_ms:Used by tests to ensure certain actions happen in the middle of a condense. A maximum value of 1 should be sufficient.
parm: zfs_reconstruct_indirect_combinations_max:Maximum number of combinations when reconstructing split segments
parm: zfs_vdev_cache_max:Inflate reads small than max
parm: zfs_vdev_cache_size:Total size of the per-disk cache
parm: zfs_vdev_cache_bshift:Shift size to inflate reads too
parm: zfs_vdev_default_ms_count:Target number of metaslabs per top-level vdev
parm: zfs_vdev_default_ms_shift:Default limit for metaslab size
parm: zfs_vdev_min_ms_count:Minimum number of metaslabs per top-level vdev
parm: zfs_vdev_ms_count_limit:Practical upper limit of total metaslabs per top-level vdev
parm: zfs_slow_io_events_per_second:Rate limit slow IO (delay) events to this many per second
parm: zfs_checksum_events_per_second:Rate limit checksum events to this many checksum errors per second (do not set below ZED threshold).
parm: zfs_scan_ignore_errors:Ignore errors during resilver/scrub
parm: vdev_validate_skip:Bypass vdev_validate()
parm: zfs_nocacheflush:Disable cache flushes
parm: zfs_embedded_slog_min_ms:Minimum number of metaslabs required to dedicate one for log blocks
parm: zfs_vdev_min_auto_ashift:Minimum ashift used when creating new top-level vdevs
parm: zfs_vdev_max_auto_ashift:Maximum ashift used when optimizing for logical -> physical sector size on new top-level vdevs
parm: zfs_txg_timeout:Max seconds worth of delta per txg
parm: zfs_read_history:Historical statistics for the last N reads
parm: zfs_read_history_hits:Include cache hits in read history
parm: zfs_txg_history:Historical statistics for the last N txgs
parm: zfs_multihost_history:Historical statistics for last N multihost writes
parm: zfs_flags:Set additional debugging flags
parm: zfs_recover:Set to attempt to recover from fatal errors
parm: zfs_free_leak_on_eio:Set to ignore IO errors during free and permanently leak the space
parm: zfs_deadman_checktime_ms:Dead I/O check interval in milliseconds
parm: zfs_deadman_enabled:Enable deadman timer
parm: spa_asize_inflation:SPA size estimate multiplication factor
parm: zfs_ddt_data_is_special:Place DDT data into the special class
parm: zfs_user_indirect_is_special:Place user data indirect blocks into the special class
parm: zfs_deadman_failmode:Failmode for deadman timer
parm: zfs_deadman_synctime_ms:Pool sync expiration time in milliseconds
parm: zfs_deadman_ziotime_ms:IO expiration time in milliseconds
parm: zfs_special_class_metadata_reserve_pct:Small file blocks in special vdevs depends on this much free space available
parm: spa_slop_shift:Reserved free space in pool
parm: zfs_unflushed_max_mem_amt:Specific hard-limit in memory that ZFS allows to be used for unflushed changes
parm: zfs_unflushed_max_mem_ppm:Percentage of the overall system memory that ZFS allows to be used for unflushed changes (value is calculated over 1000000 for finer granularity)
parm: zfs_unflushed_log_block_max:Hard limit (upper-bound) in the size of the space map log in terms of blocks.
parm: zfs_unflushed_log_block_min:Lower-bound limit for the maximum amount of blocks allowed in log spacemap (see zfs_unflushed_log_block_max)
parm: zfs_unflushed_log_txg_max:Hard limit (upper-bound) in the size of the space map log in terms of dirty TXGs.
parm: zfs_unflushed_log_block_pct:Tunable used to determine the number of blocks that can be used for the spacemap log, expressed as a percentage of the total number of metaslabs in the pool (e.g. 400 means the number of log blocks is capped at 4 times the number of metaslabs)
parm: zfs_max_log_walking:The number of past TXGs that the flushing algorithm of the log spacemap feature uses to estimate incoming log blocks
parm: zfs_keep_log_spacemaps_at_export:Prevent the log spacemaps from being flushed and destroyed during pool export/destroy
parm: zfs_max_logsm_summary_length:Maximum number of rows allowed in the summary of the spacemap log
parm: zfs_min_metaslabs_to_flush:Minimum number of metaslabs to flush per dirty TXG
parm: spa_upgrade_errlog_limit:Limit the number of errors which will be upgraded to the new on-disk error log when enabling head_errlog
parm: spa_config_path:SPA config file (/etc/zfs/zpool.cache)
parm: zfs_autoimport_disable:Disable pool import at module load
parm: zfs_spa_discard_memory_limit:Limit for memory used in prefetching the checkpoint space map done on each vdev while discarding the checkpoint
parm: spa_load_verify_shift:log2 fraction of arc that can be used by inflight I/Os when verifying pool during import
parm: spa_load_verify_metadata:Set to traverse metadata on pool import
parm: spa_load_verify_data:Set to traverse data on pool import
parm: spa_load_print_vdev_tree:Print vdev tree to zfs_dbgmsg during pool import
parm: zio_taskq_batch_pct:Percentage of CPUs to run an IO worker thread
parm: zio_taskq_batch_tpq:Number of threads per IO worker taskqueue
parm: zfs_max_missing_tvds:Allow importing pool with up to this number of missing top-level vdevs (in read-only mode)
parm: zfs_livelist_condense_zthr_pause:Set the livelist condense zthr to pause
parm: zfs_livelist_condense_sync_pause:Set the livelist condense synctask to pause
parm: zfs_livelist_condense_sync_cancel:Whether livelist condensing was canceled in the synctask
parm: zfs_livelist_condense_zthr_cancel:Whether livelist condensing was canceled in the zthr function
parm: zfs_livelist_condense_new_alloc:Whether extra ALLOC blkptrs were added to a livelist entry while it was being condensed
parm: zfs_multilist_num_sublists:Number of sublists used in each multilist
parm: zfs_multihost_interval:Milliseconds between mmp writes to each leaf
parm: zfs_multihost_fail_intervals:Max allowed period without a successful mmp write
parm: zfs_multihost_import_intervals:Number of zfs_multihost_interval periods to wait for activity
parm: metaslab_aliquot:Allocation granularity (a.k.a. stripe size)
parm: metaslab_debug_load:Load all metaslabs when pool is first opened
parm: metaslab_debug_unload:Prevent metaslabs from being unloaded
parm: metaslab_preload_enabled:Preload potential metaslabs during reassessment
parm: metaslab_unload_delay:Delay in txgs after metaslab was last used before unloading
parm: metaslab_unload_delay_ms:Delay in milliseconds after metaslab was last used before unloading
parm: zfs_mg_noalloc_threshold:Percentage of metaslab group size that should be free to make it eligible for allocation
parm: zfs_mg_fragmentation_threshold:Percentage of metaslab group size that should be considered eligible for allocations unless all metaslab groups within the metaslab class have also crossed this threshold
parm: metaslab_fragmentation_factor_enabled:Use the fragmentation metric to prefer less fragmented metaslabs
parm: zfs_metaslab_fragmentation_threshold:Fragmentation for metaslab to allow allocation
parm: metaslab_lba_weighting_enabled:Prefer metaslabs with lower LBAs
parm: metaslab_bias_enabled:Enable metaslab group biasing
parm: zfs_metaslab_segment_weight_enabled:Enable segment-based metaslab selection
parm: zfs_metaslab_switch_threshold:Segment-based metaslab selection maximum buckets before switching
parm: metaslab_force_ganging:Blocks larger than this size are forced to be gang blocks
parm: metaslab_df_max_search:Max distance (bytes) to search forward before using size tree
parm: metaslab_df_use_largest_segment:When looking in size tree, use largest segment instead of exact fit
parm: zfs_metaslab_max_size_cache_sec:How long to trust the cached max chunk size of a metaslab
parm: zfs_metaslab_mem_limit:Percentage of memory that can be used to store metaslab range trees
parm: zfs_metaslab_try_hard_before_gang:Try hard to allocate before ganging
parm: zfs_metaslab_find_max_tries:Normally only consider this many of the best metaslabs in each vdev
parm: zfs_zevent_len_max:Max event queue length
parm: zfs_scan_vdev_limit:Max bytes in flight per leaf vdev for scrubs and resilvers
parm: zfs_scrub_min_time_ms:Min millisecs to scrub per txg
parm: zfs_obsolete_min_time_ms:Min millisecs to obsolete per txg
parm: zfs_free_min_time_ms:Min millisecs to free per txg
parm: zfs_resilver_min_time_ms:Min millisecs to resilver per txg
parm: zfs_scan_suspend_progress:Set to prevent scans from progressing
parm: zfs_no_scrub_io:Set to disable scrub I/O
parm: zfs_no_scrub_prefetch:Set to disable scrub prefetching
parm: zfs_async_block_max_blocks:Max number of blocks freed in one txg
parm: zfs_max_async_dedup_frees:Max number of dedup blocks freed in one txg
parm: zfs_free_bpobj_enabled:Enable processing of the free_bpobj
parm: zfs_scan_blkstats:Enable block statistics calculation during scrub
parm: zfs_scan_mem_lim_fact:Fraction of RAM for scan hard limit
parm: zfs_scan_issue_strategy:IO issuing strategy during scrubbing. 0 = default, 1 = LBA, 2 = size
parm: zfs_scan_legacy:Scrub using legacy non-sequential method
parm: zfs_scan_checkpoint_intval:Scan progress on-disk checkpointing interval
parm: zfs_scan_max_ext_gap:Max gap in bytes between sequential scrub / resilver I/Os
parm: zfs_scan_mem_lim_soft_fact:Fraction of hard limit used as soft limit
parm: zfs_scan_strict_mem_lim:Tunable to attempt to reduce lock contention
parm: zfs_scan_fill_weight:Tunable to adjust bias towards more filled segments during scans
parm: zfs_scan_report_txgs:Tunable to report resilver performance over the last N txgs
parm: zfs_resilver_disable_defer:Process all resilvers immediately
parm: zfs_dirty_data_max_percent:Max percent of RAM allowed to be dirty
parm: zfs_dirty_data_max_max_percent:zfs_dirty_data_max upper bound as % of RAM
parm: zfs_delay_min_dirty_percent:Transaction delay threshold
parm: zfs_dirty_data_max:Determines the dirty space limit
parm: zfs_wrlog_data_max:The size limit of write-transaction zil log data
parm: zfs_dirty_data_max_max:zfs_dirty_data_max upper bound in bytes
parm: zfs_dirty_data_sync_percent:Dirty data txg sync threshold as a percentage of zfs_dirty_data_max
parm: zfs_delay_scale:How quickly delay approaches infinity
parm: zfs_sync_taskq_batch_pct:Max percent of CPUs that are used to sync dirty data
parm: zfs_zil_clean_taskq_nthr_pct:Max percent of CPUs that are used per dp_sync_taskq
parm: zfs_zil_clean_taskq_minalloc:Number of taskq entries that are pre-populated
parm: zfs_zil_clean_taskq_maxalloc:Max number of taskq entries that are cached
parm: zvol_enforce_quotas:Enable strict ZVOL quota enforcment
parm: zfs_livelist_max_entries:Size to start the next sub-livelist in a livelist
parm: zfs_livelist_min_percent_shared:Threshold at which livelist is disabled
parm: zfs_max_recordsize:Max allowed record size
parm: zfs_allow_redacted_dataset_mount:Allow mounting of redacted datasets
parm: zfs_snapshot_history_enabled:Include snapshot events in pool history/events
parm: zfs_disable_ivset_guid_check:Set to allow raw receives without IVset guids
parm: zfs_default_bs:Default dnode block shift
parm: zfs_default_ibs:Default dnode indirect block shift
parm: zfs_prefetch_disable:Disable all ZFS prefetching
parm: zfetch_max_streams:Max number of streams per zfetch
parm: zfetch_min_sec_reap:Min time before stream reclaim
parm: zfetch_max_sec_reap:Max time before stream delete
parm: zfetch_min_distance:Min bytes to prefetch per stream
parm: zfetch_max_distance:Max bytes to prefetch per stream
parm: zfetch_max_idistance:Max bytes to prefetch indirects for per stream
parm: zfetch_array_rd_sz:Number of bytes in a array_read
parm: zfs_pd_bytes_max:Max number of bytes to prefetch
parm: zfs_traverse_indirect_prefetch_limit:Traverse prefetch number of blocks pointed by indirect block
parm: ignore_hole_birth:Alias for send_holes_without_birth_time (int)
parm: send_holes_without_birth_time:Ignore hole_birth txg for zfs send
parm: zfs_send_corrupt_data:Allow sending corrupt data
parm: zfs_send_queue_length:Maximum send queue length
parm: zfs_send_unmodified_spill_blocks:Send unmodified spill blocks
parm: zfs_send_no_prefetch_queue_length:Maximum send queue length for non-prefetch queues
parm: zfs_send_queue_ff:Send queue fill fraction
parm: zfs_send_no_prefetch_queue_ff:Send queue fill fraction for non-prefetch queues
parm: zfs_override_estimate_recordsize:Override block size estimate with fixed size
parm: zfs_recv_queue_length:Maximum receive queue length
parm: zfs_recv_queue_ff:Receive queue fill fraction
parm: zfs_recv_write_batch_size:Maximum amount of writes to batch into one transaction
parm: zfs_recv_best_effort_corrective:Ignore errors during corrective receive
parm: dmu_object_alloc_chunk_shift:CPU-specific allocator grabs 2^N objects at once
parm: zfs_nopwrite_enabled:Enable NOP writes
parm: zfs_per_txg_dirty_frees_percent:Percentage of dirtied blocks from frees in one TXG
parm: zfs_dmu_offset_next_sync:Enable forcing txg sync to find holes
parm: dmu_prefetch_max:Limit one prefetch call to this size
parm: zfs_dedup_prefetch:Enable prefetching dedup-ed blks
parm: zfs_dbuf_state_index:Calculate arc header index
parm: dbuf_cache_max_bytes:Maximum size in bytes of the dbuf cache.
parm: dbuf_cache_hiwater_pct:Percentage over dbuf_cache_max_bytes for direct dbuf eviction.
parm: dbuf_cache_lowater_pct:Percentage below dbuf_cache_max_bytes when dbuf eviction stops.
parm: dbuf_metadata_cache_max_bytes:Maximum size in bytes of dbuf metadata cache.
parm: dbuf_cache_shift:Set size of dbuf cache to log2 fraction of arc size.
parm: dbuf_metadata_cache_shift:Set size of dbuf metadata cache to log2 fraction of arc size.
parm: dbuf_mutex_cache_shift:Set size of dbuf cache mutex array as log2 shift.
parm: zfs_btree_verify_intensity:Enable btree verification. Levels above 4 require ZFS be built with debugging
parm: zfs_arc_min:Minimum ARC size in bytes
parm: zfs_arc_max:Maximum ARC size in bytes
parm: zfs_arc_meta_limit:Metadata limit for ARC size in bytes
parm: zfs_arc_meta_limit_percent:Percent of ARC size for ARC meta limit
parm: zfs_arc_meta_min:Minimum ARC metadata size in bytes
parm: zfs_arc_meta_prune:Meta objects to scan for prune
parm: zfs_arc_meta_adjust_restarts:Limit number of restarts in arc_evict_meta
parm: zfs_arc_meta_strategy:Meta reclaim strategy
parm: zfs_arc_grow_retry:Seconds before growing ARC size
parm: zfs_arc_p_dampener_disable:Disable arc_p adapt dampener
parm: zfs_arc_shrink_shift:log2(fraction of ARC to reclaim)
parm: zfs_arc_pc_percent:Percent of pagecache to reclaim ARC to
parm: zfs_arc_p_min_shift:arc_c shift to calc min/max arc_p
parm: zfs_arc_average_blocksize:Target average block size
parm: zfs_compressed_arc_enabled:Disable compressed ARC buffers
parm: zfs_arc_min_prefetch_ms:Min life of prefetch block in ms
parm: zfs_arc_min_prescient_prefetch_ms:Min life of prescient prefetched block in ms
parm: l2arc_write_max:Max write bytes per interval
parm: l2arc_write_boost:Extra write bytes during device warmup
parm: l2arc_headroom:Number of max device writes to precache
parm: l2arc_headroom_boost:Compressed l2arc_headroom multiplier
parm: l2arc_trim_ahead:TRIM ahead L2ARC write size multiplier
parm: l2arc_feed_secs:Seconds between L2ARC writing
parm: l2arc_feed_min_ms:Min feed interval in milliseconds
parm: l2arc_noprefetch:Skip caching prefetched buffers
parm: l2arc_feed_again:Turbo L2ARC warmup
parm: l2arc_norw:No reads during writes
parm: l2arc_meta_percent:Percent of ARC size allowed for L2ARC-only headers
parm: l2arc_rebuild_enabled:Rebuild the L2ARC when importing a pool
parm: l2arc_rebuild_blocks_min_l2size:Min size in bytes to write rebuild log blocks in L2ARC
parm: l2arc_mfuonly:Cache only MFU data from ARC into L2ARC
parm: l2arc_exclude_special:Exclude dbufs on special vdevs from being cached to L2ARC if set.
parm: zfs_arc_lotsfree_percent:System free memory I/O throttle in bytes
parm: zfs_arc_sys_free:System free memory target size in bytes
parm: zfs_arc_dnode_limit:Minimum bytes of dnodes in ARC
parm: zfs_arc_dnode_limit_percent:Percent of ARC meta buffers for dnodes
parm: zfs_arc_dnode_reduce_percent:Percentage of excess dnodes to try to unpin
parm: zfs_arc_eviction_pct:When full, ARC allocation waits for eviction of this % of alloc size
parm: zfs_arc_evict_batch_limit:The number of headers to evict per sublist before moving to the next
parm: zfs_arc_prune_task_threads:Number of arc_prune threads
parm: zstd_earlyabort_pass:Enable early abort attempts when using zstd
parm: zstd_abort_size:Minimal size of block to attempt early abort
parm: zfs_max_dataset_nesting:Limit to the amount of nesting a path can have. Defaults to 50.
parm: zfs_fletcher_4_impl:Select fletcher 4 implementation.
parm: icp_gcm_impl:Select gcm implementation.
parm: icp_gcm_avx_chunk_size:How many bytes to process while owning the FPU
parm: zfs_blake3_impl:Select BLAKE3 implementation.
parm: icp_aes_impl:Select aes implementation.
```
|
defect
|
the zfs kernel module fails to be installed error could not insert zfs key was rejected by service system information distribution name fedora xfce desktop distribution version kernel version architecture openzfs version zfs describe the problem you re observing i tried to install the zfs module but modprobe told me to reject the installation describe how to reproduce the problem i tried using the zfs package provided by zfsonlinux org and the zfs rpm package built from git scm and they both had this problem they seem to be related to a feature of the kernel module what do i need to do to make it work thanks sudo modprobe zfs password for gesangtome modprobe error could not insert zfs key was rejected by service modinfo for zfs ko modinfo zfs filename lib modules extra zfs ko xz version license cddl license dual bsd gpl license dual mit gpl author openzfs description zfs alias zzstd alias zcommon alias zunicode alias znvpair alias zlua alias icp alias zavl alias devname zfs alias char major srcversion depends spl retpoline y name zfs vermagic smp preempt mod unload sig id pkcs signer dkms module signing key sig key ad be ac da sig hashalgo signature df ee aa fb fc fb fd bb de ff af ca bb af dd ec aa ef ea ec be da bc be fe fe fe ed ba bd af bc bc parm zvol inhibit dev do not create zvol device nodes uint parm zvol major major number for zvol device uint parm zvol threads number of threads to handle i o requests setto to use all active cpus uint parm zvol request sync synchronously handle bio requests uint parm zvol max discard blocks max number of blocks to discard ulong parm zvol prefetch bytes prefetch n bytes at zvol start end uint parm zvol volmode default volmode property value uint parm zvol blk mq queue depth default blk mq queue depth uint parm zvol use blk mq use the blk mq api for zvols uint parm zvol blk mq blocks per thread process volblocksize blocks per thread uint parm zfs xattr compat use legacy zfs xattr naming for writing new user namespace xattrs parm zfs fallocate reserve percent percentage of length to use for the available capacity check uint parm zfs key max salt uses max number of times a salt value can be used for generating encryption keys before it is rotated ulong parm zfs object mutex size size of znode hold array uint parm zfs unlink suspend progress set to prevent async unlinks debug leaks space into the unlinked set int parm zfs delete blocks delete files larger than n blocks async ulong parm zfs dbgmsg enable enable zfs debug message log int parm zfs dbgmsg maxsize maximum zfs debug log size uint parm zfs admin snapshot enable mkdir rmdir mv in zfs snapshot int parm zfs expire snapshot seconds to expire zfs snapshot int parm vdev file logical ashift logical ashift for file based devices parm vdev file physical ashift physical ashift for file based devices parm zfs vdev scheduler i o scheduler parm zfs vdev open timeout ms timeout before determining that a device is missing parm zfs vdev failfast mask defines failfast mask device transport driver parm zfs arc shrinker limit limit on number of pages that arc shrinker can reclaim at once parm zfs abd scatter enabled toggle whether abd allocations must be linear int parm zfs abd scatter min size minimum size of scatter allocations int parm zfs abd scatter max order maximum order allocation used for a scatter abd uint parm zio slow io ms max i o completion time milliseconds before marking it as slow parm zio requeue io start cut in line prioritize requeued i o parm zfs sync pass deferred free defer frees starting in this pass parm zfs sync pass dont compress don t compress starting in this pass parm zfs sync pass rewrite rewrite new bps starting in this pass parm zio dva throttle enabled throttle block allocations in the zio pipeline parm zio deadman log all log all slow zios not just those with vdevs parm zfs commit timeout pct zil block open timeout percentage parm zil min commit timeout minimum delay we care for zil block commit parm zil replay disable disable intent logging replay parm zil nocacheflush disable zil cache flushes parm zil slog bulk limit in bytes slog sync writes per commit parm zil maxblocksize limit in bytes of zil log block size parm zfs vnops read chunk size bytes to read per chunk parm zfs zil saxattr disable xattr sa extended attribute logging in zil by settng parm zfs immediate write sz largest data block to write to zil parm zfs max nvlist src size maximum size in bytes allowed for src nvlist passed with zfs ioctls parm zfs history output max maximum size in bytes of zfs ioctl output that will be logged parm zfs zevent retain max maximum recent zevents records to retain for duplicate checking parm zfs zevent retain expire secs expiration time for recent zevents records parm zfs lua max instrlimit max instruction limit that can be specified for a channel program parm zfs lua max memlimit max memory limit that can be specified for a channel program parm zap micro max size maximum micro zap size before converting to a fat zap in bytes parm zap iterate prefetch when iterating zap object prefetch it parm zfs trim extent bytes max max size of trim commands larger will be split parm zfs trim extent bytes min min size of trim commands smaller will be skipped parm zfs trim metaslab skip skip metaslabs which have never been initialized parm zfs trim txg batch min number of txgs to aggregate frees before issuing trim parm zfs trim queue limit max queued trims outstanding per leaf vdev parm zfs removal ignore errors ignore hard io errors when removing device parm zfs remove max segment largest contiguous segment to allocate when removing device parm vdev removal max span largest span of free chunks a remap segment can span parm zfs removal suspend progress pause device removal after this many bytes are copied debug use only causes removal to hang parm zfs rebuild max segment max segment size in bytes of rebuild reads parm zfs rebuild vdev limit max bytes in flight per leaf vdev for sequential resilvers parm zfs rebuild scrub enabled automatically scrub after sequential resilver completes parm zfs vdev raidz impl select raidz implementation parm zfs vdev aggregation limit max vdev i o aggregation size parm zfs vdev aggregation limit non rotating max vdev i o aggregation size for non rotating media parm zfs vdev aggregate trim allow trim i o to be aggregated parm zfs vdev read gap limit aggregate read i o over gap parm zfs vdev write gap limit aggregate write i o over gap parm zfs vdev max active maximum number of active i os per vdev parm zfs vdev async write active max dirty percent async write concurrency max threshold parm zfs vdev async write active min dirty percent async write concurrency min threshold parm zfs vdev async read max active max active async read i os per vdev parm zfs vdev async read min active min active async read i os per vdev parm zfs vdev async write max active max active async write i os per vdev parm zfs vdev async write min active min active async write i os per vdev parm zfs vdev initializing max active max active initializing i os per vdev parm zfs vdev initializing min active min active initializing i os per vdev parm zfs vdev removal max active max active removal i os per vdev parm zfs vdev removal min active min active removal i os per vdev parm zfs vdev scrub max active max active scrub i os per vdev parm zfs vdev scrub min active min active scrub i os per vdev parm zfs vdev sync read max active max active sync read i os per vdev parm zfs vdev sync read min active min active sync read i os per vdev parm zfs vdev sync write max active max active sync write i os per vdev parm zfs vdev sync write min active min active sync write i os per vdev parm zfs vdev trim max active max active trim discard i os per vdev parm zfs vdev trim min active min active trim discard i os per vdev parm zfs vdev rebuild max active max active rebuild i os per vdev parm zfs vdev rebuild min active min active rebuild i os per vdev parm zfs vdev nia credit number of non interactive i os to allow in sequence parm zfs vdev nia delay number of non interactive i os before max active parm zfs vdev queue depth pct queue depth percentage for each top level vdev parm zfs vdev mirror rotating inc rotating media load increment for non seeking i os parm zfs vdev mirror rotating seek inc rotating media load increment for seeking i os parm zfs vdev mirror rotating seek offset offset in bytes from the last i o which triggers a reduced rotating media seek increment parm zfs vdev mirror non rotating inc non rotating media load increment for non seeking i os parm zfs vdev mirror non rotating seek inc non rotating media load increment for seeking i os parm zfs initialize value value written during zpool initialize parm zfs initialize chunk size size in bytes of writes by zpool initialize parm zfs condense indirect vdevs enable whether to attempt condensing indirect vdev mappings parm zfs condense indirect obsolete pct minimum obsolete percent of bytes in the mapping to attempt condensing parm zfs condense min mapping bytes don t bother condensing if the mapping uses less than this amount of memory parm zfs condense max obsolete bytes minimum size obsolete spacemap to attempt condensing parm zfs condense indirect commit entry delay ms used by tests to ensure certain actions happen in the middle of a condense a maximum value of should be sufficient parm zfs reconstruct indirect combinations max maximum number of combinations when reconstructing split segments parm zfs vdev cache max inflate reads small than max parm zfs vdev cache size total size of the per disk cache parm zfs vdev cache bshift shift size to inflate reads too parm zfs vdev default ms count target number of metaslabs per top level vdev parm zfs vdev default ms shift default limit for metaslab size parm zfs vdev min ms count minimum number of metaslabs per top level vdev parm zfs vdev ms count limit practical upper limit of total metaslabs per top level vdev parm zfs slow io events per second rate limit slow io delay events to this many per second parm zfs checksum events per second rate limit checksum events to this many checksum errors per second do not set below zed threshold parm zfs scan ignore errors ignore errors during resilver scrub parm vdev validate skip bypass vdev validate parm zfs nocacheflush disable cache flushes parm zfs embedded slog min ms minimum number of metaslabs required to dedicate one for log blocks parm zfs vdev min auto ashift minimum ashift used when creating new top level vdevs parm zfs vdev max auto ashift maximum ashift used when optimizing for logical physical sector size on new top level vdevs parm zfs txg timeout max seconds worth of delta per txg parm zfs read history historical statistics for the last n reads parm zfs read history hits include cache hits in read history parm zfs txg history historical statistics for the last n txgs parm zfs multihost history historical statistics for last n multihost writes parm zfs flags set additional debugging flags parm zfs recover set to attempt to recover from fatal errors parm zfs free leak on eio set to ignore io errors during free and permanently leak the space parm zfs deadman checktime ms dead i o check interval in milliseconds parm zfs deadman enabled enable deadman timer parm spa asize inflation spa size estimate multiplication factor parm zfs ddt data is special place ddt data into the special class parm zfs user indirect is special place user data indirect blocks into the special class parm zfs deadman failmode failmode for deadman timer parm zfs deadman synctime ms pool sync expiration time in milliseconds parm zfs deadman ziotime ms io expiration time in milliseconds parm zfs special class metadata reserve pct small file blocks in special vdevs depends on this much free space available parm spa slop shift reserved free space in pool parm zfs unflushed max mem amt specific hard limit in memory that zfs allows to be used for unflushed changes parm zfs unflushed max mem ppm percentage of the overall system memory that zfs allows to be used for unflushed changes value is calculated over for finer granularity parm zfs unflushed log block max hard limit upper bound in the size of the space map log in terms of blocks parm zfs unflushed log block min lower bound limit for the maximum amount of blocks allowed in log spacemap see zfs unflushed log block max parm zfs unflushed log txg max hard limit upper bound in the size of the space map log in terms of dirty txgs parm zfs unflushed log block pct tunable used to determine the number of blocks that can be used for the spacemap log expressed as a percentage of the total number of metaslabs in the pool e g means the number of log blocks is capped at times the number of metaslabs parm zfs max log walking the number of past txgs that the flushing algorithm of the log spacemap feature uses to estimate incoming log blocks parm zfs keep log spacemaps at export prevent the log spacemaps from being flushed and destroyed during pool export destroy parm zfs max logsm summary length maximum number of rows allowed in the summary of the spacemap log parm zfs min metaslabs to flush minimum number of metaslabs to flush per dirty txg parm spa upgrade errlog limit limit the number of errors which will be upgraded to the new on disk error log when enabling head errlog parm spa config path spa config file etc zfs zpool cache parm zfs autoimport disable disable pool import at module load parm zfs spa discard memory limit limit for memory used in prefetching the checkpoint space map done on each vdev while discarding the checkpoint parm spa load verify shift fraction of arc that can be used by inflight i os when verifying pool during import parm spa load verify metadata set to traverse metadata on pool import parm spa load verify data set to traverse data on pool import parm spa load print vdev tree print vdev tree to zfs dbgmsg during pool import parm zio taskq batch pct percentage of cpus to run an io worker thread parm zio taskq batch tpq number of threads per io worker taskqueue parm zfs max missing tvds allow importing pool with up to this number of missing top level vdevs in read only mode parm zfs livelist condense zthr pause set the livelist condense zthr to pause parm zfs livelist condense sync pause set the livelist condense synctask to pause parm zfs livelist condense sync cancel whether livelist condensing was canceled in the synctask parm zfs livelist condense zthr cancel whether livelist condensing was canceled in the zthr function parm zfs livelist condense new alloc whether extra alloc blkptrs were added to a livelist entry while it was being condensed parm zfs multilist num sublists number of sublists used in each multilist parm zfs multihost interval milliseconds between mmp writes to each leaf parm zfs multihost fail intervals max allowed period without a successful mmp write parm zfs multihost import intervals number of zfs multihost interval periods to wait for activity parm metaslab aliquot allocation granularity a k a stripe size parm metaslab debug load load all metaslabs when pool is first opened parm metaslab debug unload prevent metaslabs from being unloaded parm metaslab preload enabled preload potential metaslabs during reassessment parm metaslab unload delay delay in txgs after metaslab was last used before unloading parm metaslab unload delay ms delay in milliseconds after metaslab was last used before unloading parm zfs mg noalloc threshold percentage of metaslab group size that should be free to make it eligible for allocation parm zfs mg fragmentation threshold percentage of metaslab group size that should be considered eligible for allocations unless all metaslab groups within the metaslab class have also crossed this threshold parm metaslab fragmentation factor enabled use the fragmentation metric to prefer less fragmented metaslabs parm zfs metaslab fragmentation threshold fragmentation for metaslab to allow allocation parm metaslab lba weighting enabled prefer metaslabs with lower lbas parm metaslab bias enabled enable metaslab group biasing parm zfs metaslab segment weight enabled enable segment based metaslab selection parm zfs metaslab switch threshold segment based metaslab selection maximum buckets before switching parm metaslab force ganging blocks larger than this size are forced to be gang blocks parm metaslab df max search max distance bytes to search forward before using size tree parm metaslab df use largest segment when looking in size tree use largest segment instead of exact fit parm zfs metaslab max size cache sec how long to trust the cached max chunk size of a metaslab parm zfs metaslab mem limit percentage of memory that can be used to store metaslab range trees parm zfs metaslab try hard before gang try hard to allocate before ganging parm zfs metaslab find max tries normally only consider this many of the best metaslabs in each vdev parm zfs zevent len max max event queue length parm zfs scan vdev limit max bytes in flight per leaf vdev for scrubs and resilvers parm zfs scrub min time ms min millisecs to scrub per txg parm zfs obsolete min time ms min millisecs to obsolete per txg parm zfs free min time ms min millisecs to free per txg parm zfs resilver min time ms min millisecs to resilver per txg parm zfs scan suspend progress set to prevent scans from progressing parm zfs no scrub io set to disable scrub i o parm zfs no scrub prefetch set to disable scrub prefetching parm zfs async block max blocks max number of blocks freed in one txg parm zfs max async dedup frees max number of dedup blocks freed in one txg parm zfs free bpobj enabled enable processing of the free bpobj parm zfs scan blkstats enable block statistics calculation during scrub parm zfs scan mem lim fact fraction of ram for scan hard limit parm zfs scan issue strategy io issuing strategy during scrubbing default lba size parm zfs scan legacy scrub using legacy non sequential method parm zfs scan checkpoint intval scan progress on disk checkpointing interval parm zfs scan max ext gap max gap in bytes between sequential scrub resilver i os parm zfs scan mem lim soft fact fraction of hard limit used as soft limit parm zfs scan strict mem lim tunable to attempt to reduce lock contention parm zfs scan fill weight tunable to adjust bias towards more filled segments during scans parm zfs scan report txgs tunable to report resilver performance over the last n txgs parm zfs resilver disable defer process all resilvers immediately parm zfs dirty data max percent max percent of ram allowed to be dirty parm zfs dirty data max max percent zfs dirty data max upper bound as of ram parm zfs delay min dirty percent transaction delay threshold parm zfs dirty data max determines the dirty space limit parm zfs wrlog data max the size limit of write transaction zil log data parm zfs dirty data max max zfs dirty data max upper bound in bytes parm zfs dirty data sync percent dirty data txg sync threshold as a percentage of zfs dirty data max parm zfs delay scale how quickly delay approaches infinity parm zfs sync taskq batch pct max percent of cpus that are used to sync dirty data parm zfs zil clean taskq nthr pct max percent of cpus that are used per dp sync taskq parm zfs zil clean taskq minalloc number of taskq entries that are pre populated parm zfs zil clean taskq maxalloc max number of taskq entries that are cached parm zvol enforce quotas enable strict zvol quota enforcment parm zfs livelist max entries size to start the next sub livelist in a livelist parm zfs livelist min percent shared threshold at which livelist is disabled parm zfs max recordsize max allowed record size parm zfs allow redacted dataset mount allow mounting of redacted datasets parm zfs snapshot history enabled include snapshot events in pool history events parm zfs disable ivset guid check set to allow raw receives without ivset guids parm zfs default bs default dnode block shift parm zfs default ibs default dnode indirect block shift parm zfs prefetch disable disable all zfs prefetching parm zfetch max streams max number of streams per zfetch parm zfetch min sec reap min time before stream reclaim parm zfetch max sec reap max time before stream delete parm zfetch min distance min bytes to prefetch per stream parm zfetch max distance max bytes to prefetch per stream parm zfetch max idistance max bytes to prefetch indirects for per stream parm zfetch array rd sz number of bytes in a array read parm zfs pd bytes max max number of bytes to prefetch parm zfs traverse indirect prefetch limit traverse prefetch number of blocks pointed by indirect block parm ignore hole birth alias for send holes without birth time int parm send holes without birth time ignore hole birth txg for zfs send parm zfs send corrupt data allow sending corrupt data parm zfs send queue length maximum send queue length parm zfs send unmodified spill blocks send unmodified spill blocks parm zfs send no prefetch queue length maximum send queue length for non prefetch queues parm zfs send queue ff send queue fill fraction parm zfs send no prefetch queue ff send queue fill fraction for non prefetch queues parm zfs override estimate recordsize override block size estimate with fixed size parm zfs recv queue length maximum receive queue length parm zfs recv queue ff receive queue fill fraction parm zfs recv write batch size maximum amount of writes to batch into one transaction parm zfs recv best effort corrective ignore errors during corrective receive parm dmu object alloc chunk shift cpu specific allocator grabs n objects at once parm zfs nopwrite enabled enable nop writes parm zfs per txg dirty frees percent percentage of dirtied blocks from frees in one txg parm zfs dmu offset next sync enable forcing txg sync to find holes parm dmu prefetch max limit one prefetch call to this size parm zfs dedup prefetch enable prefetching dedup ed blks parm zfs dbuf state index calculate arc header index parm dbuf cache max bytes maximum size in bytes of the dbuf cache parm dbuf cache hiwater pct percentage over dbuf cache max bytes for direct dbuf eviction parm dbuf cache lowater pct percentage below dbuf cache max bytes when dbuf eviction stops parm dbuf metadata cache max bytes maximum size in bytes of dbuf metadata cache parm dbuf cache shift set size of dbuf cache to fraction of arc size parm dbuf metadata cache shift set size of dbuf metadata cache to fraction of arc size parm dbuf mutex cache shift set size of dbuf cache mutex array as shift parm zfs btree verify intensity enable btree verification levels above require zfs be built with debugging parm zfs arc min minimum arc size in bytes parm zfs arc max maximum arc size in bytes parm zfs arc meta limit metadata limit for arc size in bytes parm zfs arc meta limit percent percent of arc size for arc meta limit parm zfs arc meta min minimum arc metadata size in bytes parm zfs arc meta prune meta objects to scan for prune parm zfs arc meta adjust restarts limit number of restarts in arc evict meta parm zfs arc meta strategy meta reclaim strategy parm zfs arc grow retry seconds before growing arc size parm zfs arc p dampener disable disable arc p adapt dampener parm zfs arc shrink shift fraction of arc to reclaim parm zfs arc pc percent percent of pagecache to reclaim arc to parm zfs arc p min shift arc c shift to calc min max arc p parm zfs arc average blocksize target average block size parm zfs compressed arc enabled disable compressed arc buffers parm zfs arc min prefetch ms min life of prefetch block in ms parm zfs arc min prescient prefetch ms min life of prescient prefetched block in ms parm write max max write bytes per interval parm write boost extra write bytes during device warmup parm headroom number of max device writes to precache parm headroom boost compressed headroom multiplier parm trim ahead trim ahead write size multiplier parm feed secs seconds between writing parm feed min ms min feed interval in milliseconds parm noprefetch skip caching prefetched buffers parm feed again turbo warmup parm norw no reads during writes parm meta percent percent of arc size allowed for only headers parm rebuild enabled rebuild the when importing a pool parm rebuild blocks min min size in bytes to write rebuild log blocks in parm mfuonly cache only mfu data from arc into parm exclude special exclude dbufs on special vdevs from being cached to if set parm zfs arc lotsfree percent system free memory i o throttle in bytes parm zfs arc sys free system free memory target size in bytes parm zfs arc dnode limit minimum bytes of dnodes in arc parm zfs arc dnode limit percent percent of arc meta buffers for dnodes parm zfs arc dnode reduce percent percentage of excess dnodes to try to unpin parm zfs arc eviction pct when full arc allocation waits for eviction of this of alloc size parm zfs arc evict batch limit the number of headers to evict per sublist before moving to the next parm zfs arc prune task threads number of arc prune threads parm zstd earlyabort pass enable early abort attempts when using zstd parm zstd abort size minimal size of block to attempt early abort parm zfs max dataset nesting limit to the amount of nesting a path can have defaults to parm zfs fletcher impl select fletcher implementation parm icp gcm impl select gcm implementation parm icp gcm avx chunk size how many bytes to process while owning the fpu parm zfs impl select implementation parm icp aes impl select aes implementation
| 1
|
355,759
| 25,176,017,718
|
IssuesEvent
|
2022-11-11 09:20:00
|
kangzongxian/pe
|
https://api.github.com/repos/kangzongxian/pe
|
opened
|
Diagram too small for Storage Component in DG
|
type.DocumentationBug severity.VeryLow
|
The diagram is too small for the storage component in DG, hindering the ability of the developer to understand the DG. The normal words are a good comparison as to how small the words are in the Storage Component image.

<!--session: 1668154108797-ae451c11-c287-4f8d-8dc4-3c93fc1c7f5d-->
<!--Version: Web v3.4.4-->
|
1.0
|
Diagram too small for Storage Component in DG - The diagram is too small for the storage component in DG, hindering the ability of the developer to understand the DG. The normal words are a good comparison as to how small the words are in the Storage Component image.

<!--session: 1668154108797-ae451c11-c287-4f8d-8dc4-3c93fc1c7f5d-->
<!--Version: Web v3.4.4-->
|
non_defect
|
diagram too small for storage component in dg the diagram is too small for the storage component in dg hindering the ability of the developer to understand the dg the normal words are a good comparison as to how small the words are in the storage component image
| 0
|
46,189
| 13,055,864,941
|
IssuesEvent
|
2020-07-30 02:57:56
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
Static shared_ptr<>s can crash on exit in python 2.7.4 (Trac #711)
|
IceTray Incomplete Migration Migrated from Trac defect
|
Migrated from https://code.icecube.wisc.edu/ticket/711
```json
{
"status": "closed",
"changetime": "2013-06-28T13:03:49",
"description": "Icetray has the following declaration in private/I3Logging.cxx:\n\n`static I3LoggerPtr icetray_global_logger;`\n\n`I3LoggerPtr` is a typedef for `boost::shared_ptr<I3Logger>`.\n\nThis static value can be written from python in a script:\n\n`icetray.I3Logger.global_logger = icetray.I3NullLogger()`\n\nThere is a problem with setting this value from Python: being a static value, this `shared_ptr` will continue to exist after `main()` exits, and will run its destructor as part of dynamic library shutdown. That means the logger will be deleted after Py_Finalize() run. There are two possible situations for python scripts that write this value:\n\n== Situation 1 ==\n\n`icetray.I3Logger.global_logger = icetray.I3NullLogger()`\n\nThe logger has been set to a C++ object that was allocated through python. The program shutdown sequence is:\n\n* Py_Finalize()\n* main() exits\n* the logger's destructor runs\n** Under Python 2.7.4, for unknown reasons, this always segfaults when it attempts to decrement the python reference count for this object.\n** Under Python <= 2.7.3, boost::python is somehow smart enough to avoid decrementing the python reference, and so does not segfault.\n\nSee ticket #362 for more details. http://bugs.python.org/issue17703 may be related.\n\n\n== Situation 2 ==\n\n`icetray.I3Logger.global_logger = MyPythonLogger()`\n\nThe logger has been set to a Python object. The program shutdown sequence is:\n\n* Py_Finalize()\n** Note: any use of the logger at this point would segfault!\n* main() exits\n* the logger's destructor runs\n** And segfaults (on at least Python 2.7.3 and 2.7.4)\n\n=== Conclusion ===\n\nIt's probably incorrect to ever have a static boost::shared_ptr<> to an object that might have been allocated in Python. I don't know at the moment whether the logger is the only problem.\n\n",
"reporter": "sjackso",
"cc": "",
"resolution": "fixed",
"_ts": "1372424629000000",
"component": "IceTray",
"summary": "Static shared_ptr<>s can crash on exit in python 2.7.4",
"priority": "normal",
"keywords": "",
"time": "2013-05-07T22:43:36",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
1.0
|
Static shared_ptr<>s can crash on exit in python 2.7.4 (Trac #711) - Migrated from https://code.icecube.wisc.edu/ticket/711
```json
{
"status": "closed",
"changetime": "2013-06-28T13:03:49",
"description": "Icetray has the following declaration in private/I3Logging.cxx:\n\n`static I3LoggerPtr icetray_global_logger;`\n\n`I3LoggerPtr` is a typedef for `boost::shared_ptr<I3Logger>`.\n\nThis static value can be written from python in a script:\n\n`icetray.I3Logger.global_logger = icetray.I3NullLogger()`\n\nThere is a problem with setting this value from Python: being a static value, this `shared_ptr` will continue to exist after `main()` exits, and will run its destructor as part of dynamic library shutdown. That means the logger will be deleted after Py_Finalize() run. There are two possible situations for python scripts that write this value:\n\n== Situation 1 ==\n\n`icetray.I3Logger.global_logger = icetray.I3NullLogger()`\n\nThe logger has been set to a C++ object that was allocated through python. The program shutdown sequence is:\n\n* Py_Finalize()\n* main() exits\n* the logger's destructor runs\n** Under Python 2.7.4, for unknown reasons, this always segfaults when it attempts to decrement the python reference count for this object.\n** Under Python <= 2.7.3, boost::python is somehow smart enough to avoid decrementing the python reference, and so does not segfault.\n\nSee ticket #362 for more details. http://bugs.python.org/issue17703 may be related.\n\n\n== Situation 2 ==\n\n`icetray.I3Logger.global_logger = MyPythonLogger()`\n\nThe logger has been set to a Python object. The program shutdown sequence is:\n\n* Py_Finalize()\n** Note: any use of the logger at this point would segfault!\n* main() exits\n* the logger's destructor runs\n** And segfaults (on at least Python 2.7.3 and 2.7.4)\n\n=== Conclusion ===\n\nIt's probably incorrect to ever have a static boost::shared_ptr<> to an object that might have been allocated in Python. I don't know at the moment whether the logger is the only problem.\n\n",
"reporter": "sjackso",
"cc": "",
"resolution": "fixed",
"_ts": "1372424629000000",
"component": "IceTray",
"summary": "Static shared_ptr<>s can crash on exit in python 2.7.4",
"priority": "normal",
"keywords": "",
"time": "2013-05-07T22:43:36",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
defect
|
static shared ptr s can crash on exit in python trac migrated from json status closed changetime description icetray has the following declaration in private cxx n n static icetray global logger n n is a typedef for boost shared ptr n nthis static value can be written from python in a script n n icetray global logger icetray n nthere is a problem with setting this value from python being a static value this shared ptr will continue to exist after main exits and will run its destructor as part of dynamic library shutdown that means the logger will be deleted after py finalize run there are two possible situations for python scripts that write this value n n situation n n icetray global logger icetray n nthe logger has been set to a c object that was allocated through python the program shutdown sequence is n n py finalize n main exits n the logger s destructor runs n under python for unknown reasons this always segfaults when it attempts to decrement the python reference count for this object n under python to an object that might have been allocated in python i don t know at the moment whether the logger is the only problem n n reporter sjackso cc resolution fixed ts component icetray summary static shared ptr s can crash on exit in python priority normal keywords time milestone owner type defect
| 1
|
52,425
| 3,022,947,220
|
IssuesEvent
|
2015-08-01 00:33:24
|
jpchanson/BeSeenium
|
https://api.github.com/repos/jpchanson/BeSeenium
|
closed
|
Clean up ActionController
|
Core functionality Medium Priority Refactorisation
|
Look through ActionController, make sure that it only contains necessary general code, possibly refactor class.
|
1.0
|
Clean up ActionController - Look through ActionController, make sure that it only contains necessary general code, possibly refactor class.
|
non_defect
|
clean up actioncontroller look through actioncontroller make sure that it only contains necessary general code possibly refactor class
| 0
|
234,065
| 19,093,007,705
|
IssuesEvent
|
2021-11-29 14:06:29
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
[CI] SnapshotLifecycleRestIT.testBasicTimeBasedRetention failing
|
>test-failure :Data Management/ILM+SLM Team:Data Management
|
<!--
Please fill out the following information, and ensure you have attempted
to reproduce locally
-->
**Build scan**: https://gradle-enterprise.elastic.co/s/sluoev3ttfqtm/tests/:x-pack:plugin:ilm:qa:multi-node:javaRestTest/org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT/testBasicTimeBasedRetention?top-execution=1
**Repro line**: `./gradlew ':x-pack:plugin:ilm:qa:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT.testBasicTimeBasedRetention" -Dtests.seed=9C07192018733E66 -Dtests.locale=nl-BE -Dtests.timezone=America/Guatemala -Druntime.java=11`
**Reproduces locally?**: didn't try
**Applicable branches**: master
**Failure history**: https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT&tests.test=testBasicTimeBasedRetention
<!--
Link to build stats and possible indication of when this started failing and how often it fails
<https://build-stats.elastic.co/app/kibana>
-->
**Failure excerpt**:
```
AssertionError: expected snapshot to exist but it does not: {"error":{"root_cause":[{"type":"snapshot_missing_exception","reason":"[time-based-retention-repo:snap-jtopkav2rvqjzxd0eeljtw] is missing"}],"type":"snapshot_missing_exception","reason":"[time-based-retention-repo:snap-jtopkav2rvqjzxd0eeljtw] is missing"},"status":404} |
```
|
1.0
|
[CI] SnapshotLifecycleRestIT.testBasicTimeBasedRetention failing - <!--
Please fill out the following information, and ensure you have attempted
to reproduce locally
-->
**Build scan**: https://gradle-enterprise.elastic.co/s/sluoev3ttfqtm/tests/:x-pack:plugin:ilm:qa:multi-node:javaRestTest/org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT/testBasicTimeBasedRetention?top-execution=1
**Repro line**: `./gradlew ':x-pack:plugin:ilm:qa:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT.testBasicTimeBasedRetention" -Dtests.seed=9C07192018733E66 -Dtests.locale=nl-BE -Dtests.timezone=America/Guatemala -Druntime.java=11`
**Reproduces locally?**: didn't try
**Applicable branches**: master
**Failure history**: https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.slm.SnapshotLifecycleRestIT&tests.test=testBasicTimeBasedRetention
<!--
Link to build stats and possible indication of when this started failing and how often it fails
<https://build-stats.elastic.co/app/kibana>
-->
**Failure excerpt**:
```
AssertionError: expected snapshot to exist but it does not: {"error":{"root_cause":[{"type":"snapshot_missing_exception","reason":"[time-based-retention-repo:snap-jtopkav2rvqjzxd0eeljtw] is missing"}],"type":"snapshot_missing_exception","reason":"[time-based-retention-repo:snap-jtopkav2rvqjzxd0eeljtw] is missing"},"status":404} |
```
|
non_defect
|
snapshotlifecyclerestit testbasictimebasedretention failing please fill out the following information and ensure you have attempted to reproduce locally build scan repro line gradlew x pack plugin ilm qa multi node javaresttest tests org elasticsearch xpack slm snapshotlifecyclerestit testbasictimebasedretention dtests seed dtests locale nl be dtests timezone america guatemala druntime java reproduces locally didn t try applicable branches master failure history link to build stats and possible indication of when this started failing and how often it fails failure excerpt assertionerror expected snapshot to exist but it does not error root cause is missing type snapshot missing exception reason is missing status
| 0
|
57,150
| 11,714,075,447
|
IssuesEvent
|
2020-03-09 11:35:02
|
kazuwjnlab/cvpaper
|
https://api.github.com/repos/kazuwjnlab/cvpaper
|
opened
|
[cvpaper] CVPR2019 #698 Memory-Attended Recurrent Network for Video Captioning
|
Image Captioning RNN Video Captioning multi decoder
|
## \#698 [Memory-Attended Recurrent Network for Video Captioning](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pei_Memory-Attended_Recurrent_Network_for_Video_Captioning_CVPR_2019_paper.pdf)
Wenjie Pei, Jiyuan Zhang, Xiangrong Wang, Lei Ke, Xiaoyong Shen, Yu-Wing Tai
### どんな論文か?
Video captioningの分野において、contextを考慮したキャプション生成を行うモデルMARNを提案する。
### 新規性
Vocabulary内の各単語に対して画像情報・埋め込み・その他の追加情報を与えたAttended memory decoderを用意し、decode時にencoder・Attention-based RNN decoder・Attended memory decoderとの間でattentionを計算する。
### 結果
MSR-VTT[Xu+, CVPR-2016]におけるcaptioning taskでMETEOR, ROUGE-L, CIDErの指標でS2VT[Venugopalan+, ICCV-2015]やRecNet[Wang+, CVPR-2018]よりも優れたaccuracyを達成した。
### その他(なぜ通ったか?等)
Contextの活用を目的としてattentionの改良を図ったモデルにおいて、encoderに工夫をするケースは多く見られる([Bawden+, NAACL-2018], [Kim+, CVPR-2019])が、decoderに注目した論文は少ない。またvocabulary単位で様々な情報を与える手法が独創的である。

Video captioning, Image captioning, RNN, multi decoder,
|
1.0
|
[cvpaper] CVPR2019 #698 Memory-Attended Recurrent Network for Video Captioning - ## \#698 [Memory-Attended Recurrent Network for Video Captioning](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pei_Memory-Attended_Recurrent_Network_for_Video_Captioning_CVPR_2019_paper.pdf)
Wenjie Pei, Jiyuan Zhang, Xiangrong Wang, Lei Ke, Xiaoyong Shen, Yu-Wing Tai
### どんな論文か?
Video captioningの分野において、contextを考慮したキャプション生成を行うモデルMARNを提案する。
### 新規性
Vocabulary内の各単語に対して画像情報・埋め込み・その他の追加情報を与えたAttended memory decoderを用意し、decode時にencoder・Attention-based RNN decoder・Attended memory decoderとの間でattentionを計算する。
### 結果
MSR-VTT[Xu+, CVPR-2016]におけるcaptioning taskでMETEOR, ROUGE-L, CIDErの指標でS2VT[Venugopalan+, ICCV-2015]やRecNet[Wang+, CVPR-2018]よりも優れたaccuracyを達成した。
### その他(なぜ通ったか?等)
Contextの活用を目的としてattentionの改良を図ったモデルにおいて、encoderに工夫をするケースは多く見られる([Bawden+, NAACL-2018], [Kim+, CVPR-2019])が、decoderに注目した論文は少ない。またvocabulary単位で様々な情報を与える手法が独創的である。

Video captioning, Image captioning, RNN, multi decoder,
|
non_defect
|
memory attended recurrent network for video captioning wenjie pei jiyuan zhang xiangrong wang lei ke xiaoyong shen yu wing tai どんな論文か? video captioningの分野において、contextを考慮したキャプション生成を行うモデルmarnを提案する。 新規性 vocabulary内の各単語に対して画像情報・埋め込み・その他の追加情報を与えたattended memory decoderを用意し、decode時にencoder・attention based rnn decoder・attended memory decoderとの間でattentionを計算する。 結果 msr vtt におけるcaptioning taskでmeteor rouge l やrecnet よりも優れたaccuracyを達成した。 その他(なぜ通ったか?等) contextの活用を目的としてattentionの改良を図ったモデルにおいて、encoderに工夫をするケースは多く見られる が、decoderに注目した論文は少ない。またvocabulary単位で様々な情報を与える手法が独創的である。 video captioning image captioning rnn multi decoder
| 0
|
71,548
| 23,689,067,356
|
IssuesEvent
|
2022-08-29 09:08:49
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
App crashed
|
T-Defect Z-branch
|
### Steps to reproduce
crashed when creating a new space:
new space -> private - > me only -> loading -> crash
### Outcome
#### What did you expect?
Finish creating the space
#### What happened instead?
Crash
### Your phone model
Xiaomi 11
### Operating system version
Android 12
### Application version and app store
Element dbg - testing version
### Homeserver
matrix.org
### Will you send logs?
Yes
|
1.0
|
App crashed - ### Steps to reproduce
crashed when creating a new space:
new space -> private - > me only -> loading -> crash
### Outcome
#### What did you expect?
Finish creating the space
#### What happened instead?
Crash
### Your phone model
Xiaomi 11
### Operating system version
Android 12
### Application version and app store
Element dbg - testing version
### Homeserver
matrix.org
### Will you send logs?
Yes
|
defect
|
app crashed steps to reproduce crashed when creating a new space new space private me only loading crash outcome what did you expect finish creating the space what happened instead crash your phone model xiaomi operating system version android application version and app store element dbg testing version homeserver matrix org will you send logs yes
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.