Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
44,687 | 12,325,641,114 | IssuesEvent | 2020-05-13 15:19:54 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | CMS Files/EFS Backup failing | Content export Defect DevOps Unplanned work | Last 2 builds failed with:
> fatal: [localhost]: FAILED! => {"changed": false, "msg": "Error mounting /efs: mount.nfs4: Connection timed out\n"}
http://jenkins.vfs.va.gov/job/cms/job/cms-efs-backup-prod/

- [x] Determine if flicker, run again
- [x] Create devops PR to:
1. Retain 7 days of builds, not a limit of 5 (current)
1. Notify #cms-team Slack channel on failure
- [x] Merge PR https://github.com/department-of-veterans-affairs/devops/pull/6641 | 1.0 | CMS Files/EFS Backup failing - Last 2 builds failed with:
> fatal: [localhost]: FAILED! => {"changed": false, "msg": "Error mounting /efs: mount.nfs4: Connection timed out\n"}
http://jenkins.vfs.va.gov/job/cms/job/cms-efs-backup-prod/

- [x] Determine if flicker, run again
- [x] Create devops PR to:
1. Retain 7 days of builds, not a limit of 5 (current)
1. Notify #cms-team Slack channel on failure
- [x] Merge PR https://github.com/department-of-veterans-affairs/devops/pull/6641 | defect | cms files efs backup failing last builds failed with fatal failed changed false msg error mounting efs mount connection timed out n determine if flicker run again create devops pr to retain days of builds not a limit of current notify cms team slack channel on failure merge pr | 1 |
74,825 | 25,347,594,354 | IssuesEvent | 2022-11-19 11:38:29 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | Cannot close BOINC when in fullscreen | C: Manager P: Undetermined T: Defect Validate | **Describe the bug**
Cannot close the BOINC Manager window when in fullscreen (at least on macOS, I don't know about other OS's). It works when BOINC isn't in fullscreen though.
**Steps To Reproduce**
1. Open BOINC Manager and press the fullscreen button (the green one at the top-left of the window)
2. Try to close the window (either by pressing the red X button at the top-left of the window or by pressing Command+W)
3. You will be switched to another window/app
4. Go back to see that BOINC is actually still open
5. Try to close the window again
6. This time, nothing happens, and you're not switched to another window/app
**Expected behavior**
BOINC manager should just be closed.
**System Information**
- OS: macOS 12.6
- BOINC Version: 7.20.2 | 1.0 | Cannot close BOINC when in fullscreen - **Describe the bug**
Cannot close the BOINC Manager window when in fullscreen (at least on macOS, I don't know about other OS's). It works when BOINC isn't in fullscreen though.
**Steps To Reproduce**
1. Open BOINC Manager and press the fullscreen button (the green one at the top-left of the window)
2. Try to close the window (either by pressing the red X button at the top-left of the window or by pressing Command+W)
3. You will be switched to another window/app
4. Go back to see that BOINC is actually still open
5. Try to close the window again
6. This time, nothing happens, and you're not switched to another window/app
**Expected behavior**
BOINC manager should just be closed.
**System Information**
- OS: macOS 12.6
- BOINC Version: 7.20.2 | defect | cannot close boinc when in fullscreen describe the bug cannot close the boinc manager window when in fullscreen at least on macos i don t know about other os s it works when boinc isn t in fullscreen though steps to reproduce open boinc manager and press the fullscreen button the green one at the top left of the window try to close the window either by pressing the red x button at the top left of the window or by pressing command w you will be switched to another window app go back to see that boinc is actually still open try to close the window again this time nothing happens and you re not switched to another window app expected behavior boinc manager should just be closed system information os macos boinc version | 1 |
808,232 | 30,051,276,438 | IssuesEvent | 2023-06-28 00:49:45 | apcountryman/picolibrary | https://api.github.com/repos/apcountryman/picolibrary | opened | Add Adafruit PID781 custom character ID | priority-normal status-awaiting_development type-feature | Add Adafruit PID781 custom character ID (`::picolibrary::Adafruit::PID781::Custom_Character_ID`).
- [ ] The `Custom_Character_ID` enum class should be defined in the `include/picolibrary/adafruit/pid781.h`/`source/picolibrary/adafruit/pid781.cc` header/source file pair
- [ ] The `Custom_Character_ID` enum class should have an underlying type of `std::uint8_t`
- [ ] The `Custom_Character_ID` enum class should have the following enumerators:
- [ ] `_0 = 0`: 0
- [ ] `_1 = 1`: 1
- [ ] `_2 = 2`: 2
- [ ] `_3 = 3`: 3
- [ ] `_4 = 4`: 4
- [ ] `_5 = 5`: 5
- [ ] `_6 = 6`: 6
- [ ] `_7 = 7`: 7
- [ ] A `std::ostream` insertion operator should be defined for the `Custom_Character_ID` enum class to support automated testing
- [ ] The `Custom_Character_ID` enum class `std::ostream` insertion operator should be defined in the `include/picolibrary/testing/automated/adafruit/pid781.h`/`source/picolibrary/testing/automated/adafruit/pid781.cc` header/source file pair
- [ ] `::picolibrary::Output_Formatter` should be specialized for `Custom_Character_ID`
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should be defined in the `include/picolibrary/adafruit/pid781.h`/`source/picolibrary/adafruit/pid781.cc` header/source file pair
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should have automated tests
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should support the following operations:
- [ ] `constexpr Output_Formatter() noexcept = default;`
- [ ] `constexpr Output_Formatter( Output_Formatter && source ) noexcept = default;`
- [ ] `constexpr Output_Formatter( Output_Formatter const & original ) noexcept = default;`
- [ ] `~Output_Formatter() noexcept = default;`
- [ ] `constexpr auto operator=( Output_Formatter && expression ) noexcept -> Output_Formatter & = default;`
- [ ] `constexpr auto operator=( Output_Formatter const & expression ) noexcept -> Output_Formatter & = default;`
- [ ] `auto print( Reliable_Output_Stream & stream, Adafruit::PID781::Custom_Character_ID custom_character_id ) const noexcept -> std::size_t;`: Write the formatted custom character ID to the stream
- [ ] Update documentation | 1.0 | Add Adafruit PID781 custom character ID - Add Adafruit PID781 custom character ID (`::picolibrary::Adafruit::PID781::Custom_Character_ID`).
- [ ] The `Custom_Character_ID` enum class should be defined in the `include/picolibrary/adafruit/pid781.h`/`source/picolibrary/adafruit/pid781.cc` header/source file pair
- [ ] The `Custom_Character_ID` enum class should have an underlying type of `std::uint8_t`
- [ ] The `Custom_Character_ID` enum class should have the following enumerators:
- [ ] `_0 = 0`: 0
- [ ] `_1 = 1`: 1
- [ ] `_2 = 2`: 2
- [ ] `_3 = 3`: 3
- [ ] `_4 = 4`: 4
- [ ] `_5 = 5`: 5
- [ ] `_6 = 6`: 6
- [ ] `_7 = 7`: 7
- [ ] A `std::ostream` insertion operator should be defined for the `Custom_Character_ID` enum class to support automated testing
- [ ] The `Custom_Character_ID` enum class `std::ostream` insertion operator should be defined in the `include/picolibrary/testing/automated/adafruit/pid781.h`/`source/picolibrary/testing/automated/adafruit/pid781.cc` header/source file pair
- [ ] `::picolibrary::Output_Formatter` should be specialized for `Custom_Character_ID`
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should be defined in the `include/picolibrary/adafruit/pid781.h`/`source/picolibrary/adafruit/pid781.cc` header/source file pair
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should have automated tests
- [ ] The `::picolibrary::Output_Formatter<Adafruit::PID781::Custom_Character_ID>` specialization should support the following operations:
- [ ] `constexpr Output_Formatter() noexcept = default;`
- [ ] `constexpr Output_Formatter( Output_Formatter && source ) noexcept = default;`
- [ ] `constexpr Output_Formatter( Output_Formatter const & original ) noexcept = default;`
- [ ] `~Output_Formatter() noexcept = default;`
- [ ] `constexpr auto operator=( Output_Formatter && expression ) noexcept -> Output_Formatter & = default;`
- [ ] `constexpr auto operator=( Output_Formatter const & expression ) noexcept -> Output_Formatter & = default;`
- [ ] `auto print( Reliable_Output_Stream & stream, Adafruit::PID781::Custom_Character_ID custom_character_id ) const noexcept -> std::size_t;`: Write the formatted custom character ID to the stream
- [ ] Update documentation | non_defect | add adafruit custom character id add adafruit custom character id picolibrary adafruit custom character id the custom character id enum class should be defined in the include picolibrary adafruit h source picolibrary adafruit cc header source file pair the custom character id enum class should have an underlying type of std t the custom character id enum class should have the following enumerators a std ostream insertion operator should be defined for the custom character id enum class to support automated testing the custom character id enum class std ostream insertion operator should be defined in the include picolibrary testing automated adafruit h source picolibrary testing automated adafruit cc header source file pair picolibrary output formatter should be specialized for custom character id the picolibrary output formatter specialization should be defined in the include picolibrary adafruit h source picolibrary adafruit cc header source file pair the picolibrary output formatter specialization should have automated tests the picolibrary output formatter specialization should support the following operations constexpr output formatter noexcept default constexpr output formatter output formatter source noexcept default constexpr output formatter output formatter const original noexcept default output formatter noexcept default constexpr auto operator output formatter expression noexcept output formatter default constexpr auto operator output formatter const expression noexcept output formatter default auto print reliable output stream stream adafruit custom character id custom character id const noexcept std size t write the formatted custom character id to the stream update documentation | 0 |
39,104 | 9,204,023,190 | IssuesEvent | 2019-03-08 05:30:59 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Only one ModuleDependencyAttribute is used | defect in-progress | I'm using a module with automatic dependency resolution. However, there are dependencies that can't be resolved by the compiler as they come from a view. When I try to add those dependencies by hand adding several ModuleDependencyAttributes, it only takes the first one in account, ignoring the rest.
This happens in v17.6.0 and it is a big showstopper for me.
### Steps To Reproduce
https://deck.net/084ed83fae9118f4d0630436a5fb8a3f
```csharp
public class Program
{
public static void Main()
{
Module1.Test1();
}
}
[Module(ModuleType.AMD, "Module1", true, ExportAsNamespace = "Mod1")]
[Name("Module1")]
[ModuleDependency("Module2")]
[ModuleDependency("Module3")]
[ModuleDependency("Module4")]
public static class Module1
{
public extern static void Test1();
}
```
### Expected Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
require(["Module1"], function (Mod1) {
Bridge.define("Demo.Program", {
$metadata : function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":System.Void}]}; },
main: function Main () {
Module1.Test1();
}
});
Bridge.init();
});
define("Module1", ["Module2", "Module3", "Module4"], function (Module2, Module3, Module4) {
var Module1 = { };
Bridge.define("Module1", {
$metadata : function () { return {"att":1048961,"a":2,"s":true,"m":[{"a":2,"n":"Test1","is":true,"t":8,"sn":"Test1","rt":System.Void}]}; },
$scope: Module1
});
return Module1;
});
});
```
### Actual Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
require(["Module1"], function (Mod1) {
Bridge.define("Demo.Program", {
$metadata : function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":System.Void}]}; },
main: function Main () {
Module1.Test1();
}
});
Bridge.init();
});
define("Module1", ["Module2"], function (Module2) {
var Module1 = { };
Bridge.define("Module1", {
$metadata : function () { return {"att":1048961,"a":2,"s":true,"m":[{"a":2,"n":"Test1","is":true,"t":8,"sn":"Test1","rt":System.Void}]}; },
$scope: Module1
});
return Module1;
});
});
```
| 1.0 | Only one ModuleDependencyAttribute is used - I'm using a module with automatic dependency resolution. However, there are dependencies that can't be resolved by the compiler as they come from a view. When I try to add those dependencies by hand adding several ModuleDependencyAttributes, it only takes the first one in account, ignoring the rest.
This happens in v17.6.0 and it is a big showstopper for me.
### Steps To Reproduce
https://deck.net/084ed83fae9118f4d0630436a5fb8a3f
```csharp
public class Program
{
public static void Main()
{
Module1.Test1();
}
}
[Module(ModuleType.AMD, "Module1", true, ExportAsNamespace = "Mod1")]
[Name("Module1")]
[ModuleDependency("Module2")]
[ModuleDependency("Module3")]
[ModuleDependency("Module4")]
public static class Module1
{
public extern static void Test1();
}
```
### Expected Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
require(["Module1"], function (Mod1) {
Bridge.define("Demo.Program", {
$metadata : function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":System.Void}]}; },
main: function Main () {
Module1.Test1();
}
});
Bridge.init();
});
define("Module1", ["Module2", "Module3", "Module4"], function (Module2, Module3, Module4) {
var Module1 = { };
Bridge.define("Module1", {
$metadata : function () { return {"att":1048961,"a":2,"s":true,"m":[{"a":2,"n":"Test1","is":true,"t":8,"sn":"Test1","rt":System.Void}]}; },
$scope: Module1
});
return Module1;
});
});
```
### Actual Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
require(["Module1"], function (Mod1) {
Bridge.define("Demo.Program", {
$metadata : function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":System.Void}]}; },
main: function Main () {
Module1.Test1();
}
});
Bridge.init();
});
define("Module1", ["Module2"], function (Module2) {
var Module1 = { };
Bridge.define("Module1", {
$metadata : function () { return {"att":1048961,"a":2,"s":true,"m":[{"a":2,"n":"Test1","is":true,"t":8,"sn":"Test1","rt":System.Void}]}; },
$scope: Module1
});
return Module1;
});
});
```
| defect | only one moduledependencyattribute is used i m using a module with automatic dependency resolution however there are dependencies that can t be resolved by the compiler as they come from a view when i try to add those dependencies by hand adding several moduledependencyattributes it only takes the first one in account ignoring the rest this happens in and it is a big showstopper for me steps to reproduce csharp public class program public static void main public static class public extern static void expected result js bridge assembly demo function asm globals use strict require function bridge define demo program metadata function return att a m main function main bridge init define function var bridge define metadata function return att a s true m scope return actual result js bridge assembly demo function asm globals use strict require function bridge define demo program metadata function return att a m main function main bridge init define function var bridge define metadata function return att a s true m scope return | 1 |
823,329 | 30,990,595,491 | IssuesEvent | 2023-08-09 04:06:31 | mpv-player/mpv | https://api.github.com/repos/mpv-player/mpv | closed | issues with hardwaredecoding of h265 | os:win priority:stale | ### Important Information
- mpv version: 0.33.0-184-g03b9f8e323
- Windows Version: Windows 10 21H1 and pretty much any earlier version of Windows 10
- Source of the mpv binary: shinchiro builds
- Video card and drivers: Intel HD520, all tested drivers since the issue started, also "known good" drivers that worked before
- affected mpv versions: all tested versions since over a year were affected
### Reproduction steps
Play pretty much any h265-encoded video file, enable hardware acceleration and maybe seek around a bit in the video.
### Expected behavior
Videos should play normally, mpv should stay responsive, in case of driver hiccups at least "recover" and allow for a graceful shutdown.
### Actual behavior
I've observed several outcomes without any clear pattern.
- More often than not mpv just straight up freezes.
- Sometimes a file can be played for a while without any apparent issue. Trying to seek, switch to fullscreen or resize the window often results in a freeze.
- Sometimes audio continues to play in the background, sometimes not.
- Sometimes there is is a message in the logs, which reads: `[vo/gpu/d3d11] Failed to create Texture2D`, but that does not always happen!
- Sometimes mpv continues to eat up a whole CPU core while playing dead, sometimes it does not.
Disabling hardware decoding via hotkey doesn't work either in this state. The only way to recover is to kill mpv using some sort of task manager on Windows. Ever since the issues started, I disabled hardware decoding for h265 and then everything just works fine. I know you (used to?) advice against hardware decoding, still there seems to be something at fault and worthy to report. h264, vp9 and everything else I threw at mpv work like a charm with (and without, if not available) hardware acceleration.
### Log file
A log can be found here: [https://bin.disroot.org/?50bb3a475c449723#EZTmcxzHVxEPktE95wZWTMNHx3iiFfTaE7cNTBdMeyXg](https://bin.disroot.org/?50bb3a475c449723#EZTmcxzHVxEPktE95wZWTMNHx3iiFfTaE7cNTBdMeyXg)
Note that I tried to resize the mpv window before killing mpv. This was to show that mpv was still somehow "alive" (producing logs) but not responding to any hotkeys or other mouse actions.
If you need any more information, I'll try to provide them. Thank you!
| 1.0 | issues with hardwaredecoding of h265 - ### Important Information
- mpv version: 0.33.0-184-g03b9f8e323
- Windows Version: Windows 10 21H1 and pretty much any earlier version of Windows 10
- Source of the mpv binary: shinchiro builds
- Video card and drivers: Intel HD520, all tested drivers since the issue started, also "known good" drivers that worked before
- affected mpv versions: all tested versions since over a year were affected
### Reproduction steps
Play pretty much any h265-encoded video file, enable hardware acceleration and maybe seek around a bit in the video.
### Expected behavior
Videos should play normally, mpv should stay responsive, in case of driver hiccups at least "recover" and allow for a graceful shutdown.
### Actual behavior
I've observed several outcomes without any clear pattern.
- More often than not mpv just straight up freezes.
- Sometimes a file can be played for a while without any apparent issue. Trying to seek, switch to fullscreen or resize the window often results in a freeze.
- Sometimes audio continues to play in the background, sometimes not.
- Sometimes there is is a message in the logs, which reads: `[vo/gpu/d3d11] Failed to create Texture2D`, but that does not always happen!
- Sometimes mpv continues to eat up a whole CPU core while playing dead, sometimes it does not.
Disabling hardware decoding via hotkey doesn't work either in this state. The only way to recover is to kill mpv using some sort of task manager on Windows. Ever since the issues started, I disabled hardware decoding for h265 and then everything just works fine. I know you (used to?) advice against hardware decoding, still there seems to be something at fault and worthy to report. h264, vp9 and everything else I threw at mpv work like a charm with (and without, if not available) hardware acceleration.
### Log file
A log can be found here: [https://bin.disroot.org/?50bb3a475c449723#EZTmcxzHVxEPktE95wZWTMNHx3iiFfTaE7cNTBdMeyXg](https://bin.disroot.org/?50bb3a475c449723#EZTmcxzHVxEPktE95wZWTMNHx3iiFfTaE7cNTBdMeyXg)
Note that I tried to resize the mpv window before killing mpv. This was to show that mpv was still somehow "alive" (producing logs) but not responding to any hotkeys or other mouse actions.
If you need any more information, I'll try to provide them. Thank you!
| non_defect | issues with hardwaredecoding of important information mpv version windows version windows and pretty much any earlier version of windows source of the mpv binary shinchiro builds video card and drivers intel all tested drivers since the issue started also known good drivers that worked before affected mpv versions all tested versions since over a year were affected reproduction steps play pretty much any encoded video file enable hardware acceleration and maybe seek around a bit in the video expected behavior videos should play normally mpv should stay responsive in case of driver hiccups at least recover and allow for a graceful shutdown actual behavior i ve observed several outcomes without any clear pattern more often than not mpv just straight up freezes sometimes a file can be played for a while without any apparent issue trying to seek switch to fullscreen or resize the window often results in a freeze sometimes audio continues to play in the background sometimes not sometimes there is is a message in the logs which reads failed to create but that does not always happen sometimes mpv continues to eat up a whole cpu core while playing dead sometimes it does not disabling hardware decoding via hotkey doesn t work either in this state the only way to recover is to kill mpv using some sort of task manager on windows ever since the issues started i disabled hardware decoding for and then everything just works fine i know you used to advice against hardware decoding still there seems to be something at fault and worthy to report and everything else i threw at mpv work like a charm with and without if not available hardware acceleration log file a log can be found here note that i tried to resize the mpv window before killing mpv this was to show that mpv was still somehow alive producing logs but not responding to any hotkeys or other mouse actions if you need any more information i ll try to provide them thank you | 0 |
212,920 | 16,504,392,018 | IssuesEvent | 2021-05-25 17:26:36 | 007-hd/intro-to-github | https://api.github.com/repos/007-hd/intro-to-github | closed | Use github as a project management system! | bug documentation | make a todo list
- [x] you can make a todo list by
- [x] adding a dasn
- [x] a space
- [x] and open square bracket
- [x] a space
- [x] and a close square bracket | 1.0 | Use github as a project management system! - make a todo list
- [x] you can make a todo list by
- [x] adding a dasn
- [x] a space
- [x] and open square bracket
- [x] a space
- [x] and a close square bracket | non_defect | use github as a project management system make a todo list you can make a todo list by adding a dasn a space and open square bracket a space and a close square bracket | 0 |
269,941 | 20,512,676,881 | IssuesEvent | 2022-03-01 08:34:30 | Best-engineer/codestates-nonmajor | https://api.github.com/repos/Best-engineer/codestates-nonmajor | closed | // | documentation duplicate | ### ISSUE
- Group: `client`, `server`, `sr`
- Type: `bug`, `feature`, `delete`
- Detail: fix actions from client redux
### TODO
1. [ ] Job1
2. [ ] Job2
3. [ ] Job3
### Estimated time
> Pick one
### `0.5h`
### `1h`
### `1.5h`
### `2h`
### `2.5h`
### `3h`
### Labels
- Estimated time: `E: 1h`
- Group : `client`, `server`
- Sprint: `Sprint__NUMBER__`
- Urgency: `High`, `Middle`, `Low`
| 1.0 | // - ### ISSUE
- Group: `client`, `server`, `sr`
- Type: `bug`, `feature`, `delete`
- Detail: fix actions from client redux
### TODO
1. [ ] Job1
2. [ ] Job2
3. [ ] Job3
### Estimated time
> Pick one
### `0.5h`
### `1h`
### `1.5h`
### `2h`
### `2.5h`
### `3h`
### Labels
- Estimated time: `E: 1h`
- Group : `client`, `server`
- Sprint: `Sprint__NUMBER__`
- Urgency: `High`, `Middle`, `Low`
| non_defect | issue group client server sr type bug feature delete detail fix actions from client redux todo estimated time pick one labels estimated time e group client server sprint sprint number urgency high middle low | 0 |
8,393 | 2,611,495,764 | IssuesEvent | 2015-02-27 05:35:22 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Errors in campaign mission 6 | auto-migrated Priority-Low Type-Defect | ```
Was roping around collecting crates near the upper left.
As I picked up a utility crate in one position, another immediately appeared at
the same place, so I sat there on rope collecting 5 or so crates. At the end of
the turn, I got the "well this was pointless" message while still on rope, then
the robot began his speech.
When it was over, was spammed endlessly with:
Lua: Error while calling onGameTick: ...ta//Missions/Campaign/A Classic
Fairytale/dragon.lua:371: bad argument #1 to 'unpack' (table expected, got nil)
It also happened on another turn while I was on rope, although it eventually
ended that time.
```
Original issue reported on code.google.com by `kyberneticist@gmail.com` on 1 Sep 2012 at 7:57
* Blocking: #494 | 1.0 | Errors in campaign mission 6 - ```
Was roping around collecting crates near the upper left.
As I picked up a utility crate in one position, another immediately appeared at
the same place, so I sat there on rope collecting 5 or so crates. At the end of
the turn, I got the "well this was pointless" message while still on rope, then
the robot began his speech.
When it was over, was spammed endlessly with:
Lua: Error while calling onGameTick: ...ta//Missions/Campaign/A Classic
Fairytale/dragon.lua:371: bad argument #1 to 'unpack' (table expected, got nil)
It also happened on another turn while I was on rope, although it eventually
ended that time.
```
Original issue reported on code.google.com by `kyberneticist@gmail.com` on 1 Sep 2012 at 7:57
* Blocking: #494 | defect | errors in campaign mission was roping around collecting crates near the upper left as i picked up a utility crate in one position another immediately appeared at the same place so i sat there on rope collecting or so crates at the end of the turn i got the well this was pointless message while still on rope then the robot began his speech when it was over was spammed endlessly with lua error while calling ongametick ta missions campaign a classic fairytale dragon lua bad argument to unpack table expected got nil it also happened on another turn while i was on rope although it eventually ended that time original issue reported on code google com by kyberneticist gmail com on sep at blocking | 1 |
23,419 | 3,814,346,995 | IssuesEvent | 2016-03-28 12:44:37 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | NoClassDefFoundError: javax/cache/integration/CacheLoaderException | Team: Client Team: Core Type: Defect | Hi,
We are trying to update hazelcast from 3.4 to 3.6.1, but we are repeatedly getting this issue.
`Factory method [public static com.hazelcast.core.HazelcastInstance com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(com.hazelcast.config.Config)] threw exception; nested exception is java.lang.NoClassDefFoundError: javax/cache/integration/CacheLoaderException`
How can we resolve this issue? | 1.0 | NoClassDefFoundError: javax/cache/integration/CacheLoaderException - Hi,
We are trying to update hazelcast from 3.4 to 3.6.1, but we are repeatedly getting this issue.
`Factory method [public static com.hazelcast.core.HazelcastInstance com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(com.hazelcast.config.Config)] threw exception; nested exception is java.lang.NoClassDefFoundError: javax/cache/integration/CacheLoaderException`
How can we resolve this issue? | defect | noclassdeffounderror javax cache integration cacheloaderexception hi we are trying to update hazelcast from to but we are repeatedly getting this issue factory method threw exception nested exception is java lang noclassdeffounderror javax cache integration cacheloaderexception how can we resolve this issue | 1 |
140,622 | 11,353,727,747 | IssuesEvent | 2020-01-24 16:06:49 | IntellectualSites/PlotSquared | https://api.github.com/repos/IntellectualSites/PlotSquared | opened | Schematic based World Generation is not working correctly | [?] Testing Required | <!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! -->
<!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### -->
<!--- # NOTICE:
```diff
! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered
! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for
! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8
! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support
! due to lack of developer interest and time. Additionally, NukkitX has not had feature
! updates since the Better Together, which prevents some PlotSquared features from ever
! functioning. Contributions are always welcome however!
```
**Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)**
**Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.**
**[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels!
-->
# Bug Report Template:
<!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.-->
## Required Information section:
> ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information
### Server config info (/plot debugpaste / file paste links):
<!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here -->
<!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml -->
I atteached the files because the first thing doesnt worked:
[latest.log](https://github.com/IntellectualSites/PlotSquared/files/4109359/latest.log)
[2020-01-23-9.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109363/2020-01-23-9.log.gz)
[2020-01-23-10.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109364/2020-01-23-10.log.gz)
[yml.zip](https://github.com/IntellectualSites/PlotSquared/files/4109371/yml.zip)
<!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. -->
### Server type:
**Select one**
<!-- Select the type you are reporting the issue for (put an "X" between of brackets): -->
- [X] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)*
- [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED*
- [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED*
### Minecraft Version:
**Select one**
<!-- Select the type you are reporting the issue for (put an "X" between of brackets):
The maintained versions are 1.14.4 and 1.15.x -->
- [] Minecraft 1.15
- [X] Minecraft 1.14.4
- [] Minecraft 1.13.2
- [] Minecraft Java Edition *other versions, please specify*:
- [] Minecraft Bedrock Edition *specify version*:
### Server build info:
<!--- Run /version in-game or in console & paste the full output here: -->
```
This server is running CraftBukkit version git-Spigot-cbd1a1b-009d8af (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT)
[16:31:29] [Server thread/INFO]: This is a final build for 1.14.4. Please see https://www.spigotmc.org/go/1.14.4 for details about upgrading.
```
### WorldEdit/FAWE versions:
<!--- Specify which plugin you are using, and add its version -->
- [X] FAWE *FastAsyncWorldEdit-1.15-45*:
- [] WorldEdit *version*:
### Description of the problem:
<!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. -->
<!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc -->
If I want to use the schematic based world gen. with the plot.schem in the right folder for creating a new world, an Error occurs. After the Server has been loaded completely, the world is genarated and I can also enter it, but then there are always half or broken plots.
### How to replicate:
<!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that -->
I have done everything like it is described in the wiki, but than this Error occurs. The marked space for the plot schematic is about 200 Blocks in booth x and y direction. Maybe its too big? I also checked the worlds.yml but the Plot size is set to 200.
## Additional Information:
> The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues.
Sometimes the schematic isnยดt placed completely if I use FAWE, mostly if I dont use //paste -a. Than I have to paste it again, before it works.
### Other plugins being used on the server:
<!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list -->
Plugins (19): NightVision, LuckPerms, WorldEdit, Spawntp, FastAsyncWorldEdit, SuperTrails, VoxelSniper, ClearLag, Vault, F3NPerm, Citizens, Essentials, ChatEx, CommandPanels, PlotSquared, Multiverse-Core, Multiverse-Inventories, Multiverse-SignPortals, ItemJoin
### Relevant console output, log lines, and/or screenshots:
<!--- Please use in-line code insertion
```
https://pastebin.com/Q5NqC5Ey
```
for short (20 lines or less) text blobs, or a paste service for large blobs -->
### Additional relevant comments/remarks:
<!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc -->
At the moment the Server is running on localhost
# AFFIRMATION OF COMPLETION:
<!-- Make sure you have completed the following steps (put an "X" between of brackets): -->
- [X] I included all information required in the sections above
- [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue)
- [X] I made sure I am using an up-to-date version of PlotSquared
- [X] I made sure the bug/error is not caused by any other plugin
- [] I didn't read but checked everything above.
| 1.0 | Schematic based World Generation is not working correctly - <!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! -->
<!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### -->
<!--- # NOTICE:
```diff
! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered
! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for
! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8
! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support
! due to lack of developer interest and time. Additionally, NukkitX has not had feature
! updates since the Better Together, which prevents some PlotSquared features from ever
! functioning. Contributions are always welcome however!
```
**Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)**
**Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.**
**[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels!
-->
# Bug Report Template:
<!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.-->
## Required Information section:
> ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information
### Server config info (/plot debugpaste / file paste links):
<!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here -->
<!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml -->
I atteached the files because the first thing doesnt worked:
[latest.log](https://github.com/IntellectualSites/PlotSquared/files/4109359/latest.log)
[2020-01-23-9.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109363/2020-01-23-9.log.gz)
[2020-01-23-10.log.gz](https://github.com/IntellectualSites/PlotSquared/files/4109364/2020-01-23-10.log.gz)
[yml.zip](https://github.com/IntellectualSites/PlotSquared/files/4109371/yml.zip)
<!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. -->
### Server type:
**Select one**
<!-- Select the type you are reporting the issue for (put an "X" between of brackets): -->
- [X] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)*
- [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED*
- [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED*
### Minecraft Version:
**Select one**
<!-- Select the type you are reporting the issue for (put an "X" between of brackets):
The maintained versions are 1.14.4 and 1.15.x -->
- [] Minecraft 1.15
- [X] Minecraft 1.14.4
- [] Minecraft 1.13.2
- [] Minecraft Java Edition *other versions, please specify*:
- [] Minecraft Bedrock Edition *specify version*:
### Server build info:
<!--- Run /version in-game or in console & paste the full output here: -->
```
This server is running CraftBukkit version git-Spigot-cbd1a1b-009d8af (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT)
[16:31:29] [Server thread/INFO]: This is a final build for 1.14.4. Please see https://www.spigotmc.org/go/1.14.4 for details about upgrading.
```
### WorldEdit/FAWE versions:
<!--- Specify which plugin you are using, and add its version -->
- [X] FAWE *FastAsyncWorldEdit-1.15-45*:
- [] WorldEdit *version*:
### Description of the problem:
<!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. -->
<!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc -->
If I want to use the schematic based world gen. with the plot.schem in the right folder for creating a new world, an Error occurs. After the Server has been loaded completely, the world is genarated and I can also enter it, but then there are always half or broken plots.
### How to replicate:
<!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that -->
I have done everything like it is described in the wiki, but than this Error occurs. The marked space for the plot schematic is about 200 Blocks in booth x and y direction. Maybe its too big? I also checked the worlds.yml but the Plot size is set to 200.
## Additional Information:
> The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues.
Sometimes the schematic isnยดt placed completely if I use FAWE, mostly if I dont use //paste -a. Than I have to paste it again, before it works.
### Other plugins being used on the server:
<!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list -->
Plugins (19): NightVision, LuckPerms, WorldEdit, Spawntp, FastAsyncWorldEdit, SuperTrails, VoxelSniper, ClearLag, Vault, F3NPerm, Citizens, Essentials, ChatEx, CommandPanels, PlotSquared, Multiverse-Core, Multiverse-Inventories, Multiverse-SignPortals, ItemJoin
### Relevant console output, log lines, and/or screenshots:
<!--- Please use in-line code insertion
```
https://pastebin.com/Q5NqC5Ey
```
for short (20 lines or less) text blobs, or a paste service for large blobs -->
### Additional relevant comments/remarks:
<!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc -->
At the moment the Server is running on localhost
# AFFIRMATION OF COMPLETION:
<!-- Make sure you have completed the following steps (put an "X" between of brackets): -->
- [X] I included all information required in the sections above
- [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue)
- [X] I made sure I am using an up-to-date version of PlotSquared
- [X] I made sure the bug/error is not caused by any other plugin
- [] I didn't read but checked everything above.
| non_defect | schematic based world generation is not working correctly notice diff plotsquared for minecraft java edition versions between through to are considered legacy and will receive limited to no support please consider upgrading to for future support plugins exist for which bring back behaviors found in all versions of plotsquared for sponge and nukkit x will receive limited to no support due to lack of developer interest and time additionally nukkitx has not had feature updates since the better together which prevents some plotsquared features from ever functioning contributions are always welcome however feature requests suggestions are to be submitted at the code contributions are to be done through tagging the specific issue ticket s if applicable and please for the love of the little sanity we have left use the correct channels bug report template required information section all fields in this section are required and must contain appropriate information server config info plot debugpaste file paste links i atteached the files because the first thing doesnt worked server type select one spigot paper craftbukkit should not be used re test with spigot first sponge note not actively maintained nukkitx note not actively maintained minecraft version select one select the type you are reporting the issue for put an x between of brackets the maintained versions are and x minecraft minecraft minecraft minecraft java edition other versions please specify minecraft bedrock edition specify version server build info this server is running craftbukkit version git spigot mc implementing api version snapshot this is a final build for please see for details about upgrading worldedit fawe versions fawe fastasyncworldedit worldedit version description of the problem if i want to use the schematic based world gen with the plot schem in the right folder for creating a new world an error occurs after the server has been loaded completely the world is genarated and i can also enter it but then there are always half or broken plots how to replicate i have done everything like it is described in the wiki but than this error occurs the marked space for the plot schematic is about blocks in booth x and y direction maybe its too big i also checked the worlds yml but the plot size is set to additional information the information here is optional for you to provide however it may help us to more readily diagnose any compatibility and bug issues sometimes the schematic isnยดt placed completely if i use fawe mostly if i dont use paste a than i have to paste it again before it works other plugins being used on the server plugins nightvision luckperms worldedit spawntp fastasyncworldedit supertrails voxelsniper clearlag vault citizens essentials chatex commandpanels plotsquared multiverse core multiverse inventories multiverse signportals itemjoin relevant console output log lines and or screenshots please use in line code insertion for short lines or less text blobs or a paste service for large blobs additional relevant comments remarks at the moment the server is running on localhost affirmation of completion i included all information required in the sections above i made sure there are no duplicates of this report i made sure i am using an up to date version of plotsquared i made sure the bug error is not caused by any other plugin i didn t read but checked everything above | 0 |
93,675 | 27,013,079,089 | IssuesEvent | 2023-02-10 16:54:36 | microsoft/onnxruntime | https://api.github.com/repos/microsoft/onnxruntime | opened | TensorRT Execution Build Fails on Jetson Jetpack 4.6.1 | build | ### Describe the issue
ORT fails to build on Jetson Jetpack 4.6.1
I am using the 4.6.1 docker image `nvcr.io/nvidia/l4t-ml:r32.7.1-py3`
I am following the guide published by onnxruntime https://onnxruntime.ai/docs/build/eps.html#nvidia-jetson-tx1tx2nanoxavier
I suspect that the repository has outgrown this documentation. Is there new documentation?
notably this Microsoft prebuilt image for Jetson `mcr.microsoft.com/azureml/onnxruntime:v.1.4.0-jetpack4.4-l4t-base-r32.4.3` does not include TRT execution bindings. Is there a new supported image?
Thank you for your assistance.
### Urgency
Yes, this issue is blocking our production roadmap
### Target platform
Jetson Xaiver NX, Jetpack 4.6.1
### Build script
FROM nvcr.io/nvidia/l4t-ml:r32.7.1-py3
ARG DEBIAN_FRONTEND=noninteractive
RUN apt-get update
RUN apt-get install lshw -y
RUN apt-get install git -y
RUN git clone --recursive https://github.com/microsoft/onnxruntime
RUN export CUDACXX="/usr/local/cuda/bin/nvcc"
RUN apt-get update -y
RUN apt-get install python3-pip -y
RUN apt-get install python3-matplotlib -y
RUN apt-get install gfortran -y
RUN apt-get install -y build-essential libatlas-base-dev
RUN apt-get install ffmpeg libsm6 libxext6 -y
RUN apt install -y --no-install-recommends build-essential software-properties-common libopenblas-dev
RUN wget https://github.com/Kitware/CMake/releases/download/v3.23.0-rc1/cmake-3.23.0-rc1-linux-aarch64.tar.gz
RUN tar xf cmake-3.23.0-rc1-linux-aarch64.tar.gz
RUN export PATH="/cmake-3.20.0-rc1-linux-aarch64/bin:$PATH"
RUN apt remove cmake -y
RUN pwd
RUN ln -s /cmake-3.23.0-rc1-linux-aarch64/bin/* /usr/local/bin
RUN which cmake
RUN cmake --version
ENV PATH="/usr/local/cuda/bin:${PATH}"
ENV LD_LIBRARY_PATH="/usr/local/cuda/lib64:${LD_LIBRARY_PATH}"
RUN pip3 install --upgrade pip
RUN pip3 uninstall torch -y
RUN pip3 install torch==1.10.0
RUN echo "$PATH" && echo "$LD_LIBRARY_PATH"
WORKDIR /onnxruntime
RUN ./build.sh --config Release --update --build --parallel --build_wheel --use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu --tensorrt_home /usr/lib/aarch64-linux-gnu
### Error / output
[ 13%] Building CXX object _deps/protobuf-build/CMakeFiles/libprotobuf.dir/__/protobuf-src/src/google/protobuf/wrappers.pb.cc.o
[ 13%] Linking CXX static library libprotobuf.a
[ 13%] Built target libprotobuf
Makefile:165: recipe for target 'all' failed
make: *** [all] Error 2
Traceback (most recent call last):
File "/onnxruntime/tools/ci_build/build.py", line 2741, in <module>
sys.exit(main())
File "/onnxruntime/tools/ci_build/build.py", line 2638, in main
build_targets(args, cmake_path, build_dir, configs, num_parallel_jobs, args.target)
File "/onnxruntime/tools/ci_build/build.py", line 1399, in build_targets
run_subprocess(cmd_args, env=env)
File "/onnxruntime/tools/ci_build/build.py", line 766, in run_subprocess
return run(*args, cwd=cwd, capture_stdout=capture_stdout, shell=shell, env=my_env)
File "/onnxruntime/tools/python/util/run.py", line 57, in run
shell=shell,
File "/usr/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['/usr/local/bin/cmake', '--build', '/onnxruntime/build/Linux/Release', '--config', 'Release', '--', '-j6']' returned non-zero exit status 2.
The command '/bin/sh -c ./onnxruntime/build.sh --config Release --update --build --parallel --build_wheel --use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu --tensorrt_home /usr/lib/aarch64-linux-gnu' returned a non-zero code: 1
### Visual Studio Version
_No response_
### GCC / Compiler Version
_No response_ | 1.0 | TensorRT Execution Build Fails on Jetson Jetpack 4.6.1 - ### Describe the issue
ORT fails to build on Jetson Jetpack 4.6.1
I am using the 4.6.1 docker image `nvcr.io/nvidia/l4t-ml:r32.7.1-py3`
I am following the guide published by onnxruntime https://onnxruntime.ai/docs/build/eps.html#nvidia-jetson-tx1tx2nanoxavier
I suspect that the repository has outgrown this documentation. Is there new documentation?
notably this Microsoft prebuilt image for Jetson `mcr.microsoft.com/azureml/onnxruntime:v.1.4.0-jetpack4.4-l4t-base-r32.4.3` does not include TRT execution bindings. Is there a new supported image?
Thank you for your assistance.
### Urgency
Yes, this issue is blocking our production roadmap
### Target platform
Jetson Xaiver NX, Jetpack 4.6.1
### Build script
FROM nvcr.io/nvidia/l4t-ml:r32.7.1-py3
ARG DEBIAN_FRONTEND=noninteractive
RUN apt-get update
RUN apt-get install lshw -y
RUN apt-get install git -y
RUN git clone --recursive https://github.com/microsoft/onnxruntime
RUN export CUDACXX="/usr/local/cuda/bin/nvcc"
RUN apt-get update -y
RUN apt-get install python3-pip -y
RUN apt-get install python3-matplotlib -y
RUN apt-get install gfortran -y
RUN apt-get install -y build-essential libatlas-base-dev
RUN apt-get install ffmpeg libsm6 libxext6 -y
RUN apt install -y --no-install-recommends build-essential software-properties-common libopenblas-dev
RUN wget https://github.com/Kitware/CMake/releases/download/v3.23.0-rc1/cmake-3.23.0-rc1-linux-aarch64.tar.gz
RUN tar xf cmake-3.23.0-rc1-linux-aarch64.tar.gz
RUN export PATH="/cmake-3.20.0-rc1-linux-aarch64/bin:$PATH"
RUN apt remove cmake -y
RUN pwd
RUN ln -s /cmake-3.23.0-rc1-linux-aarch64/bin/* /usr/local/bin
RUN which cmake
RUN cmake --version
ENV PATH="/usr/local/cuda/bin:${PATH}"
ENV LD_LIBRARY_PATH="/usr/local/cuda/lib64:${LD_LIBRARY_PATH}"
RUN pip3 install --upgrade pip
RUN pip3 uninstall torch -y
RUN pip3 install torch==1.10.0
RUN echo "$PATH" && echo "$LD_LIBRARY_PATH"
WORKDIR /onnxruntime
RUN ./build.sh --config Release --update --build --parallel --build_wheel --use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu --tensorrt_home /usr/lib/aarch64-linux-gnu
### Error / output
[ 13%] Building CXX object _deps/protobuf-build/CMakeFiles/libprotobuf.dir/__/protobuf-src/src/google/protobuf/wrappers.pb.cc.o
[ 13%] Linking CXX static library libprotobuf.a
[ 13%] Built target libprotobuf
Makefile:165: recipe for target 'all' failed
make: *** [all] Error 2
Traceback (most recent call last):
File "/onnxruntime/tools/ci_build/build.py", line 2741, in <module>
sys.exit(main())
File "/onnxruntime/tools/ci_build/build.py", line 2638, in main
build_targets(args, cmake_path, build_dir, configs, num_parallel_jobs, args.target)
File "/onnxruntime/tools/ci_build/build.py", line 1399, in build_targets
run_subprocess(cmd_args, env=env)
File "/onnxruntime/tools/ci_build/build.py", line 766, in run_subprocess
return run(*args, cwd=cwd, capture_stdout=capture_stdout, shell=shell, env=my_env)
File "/onnxruntime/tools/python/util/run.py", line 57, in run
shell=shell,
File "/usr/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['/usr/local/bin/cmake', '--build', '/onnxruntime/build/Linux/Release', '--config', 'Release', '--', '-j6']' returned non-zero exit status 2.
The command '/bin/sh -c ./onnxruntime/build.sh --config Release --update --build --parallel --build_wheel --use_tensorrt --cuda_home /usr/local/cuda --cudnn_home /usr/lib/aarch64-linux-gnu --tensorrt_home /usr/lib/aarch64-linux-gnu' returned a non-zero code: 1
### Visual Studio Version
_No response_
### GCC / Compiler Version
_No response_ | non_defect | tensorrt execution build fails on jetson jetpack describe the issue ort fails to build on jetson jetpack i am using the docker image nvcr io nvidia ml i am following the guide published by onnxruntime i suspect that the repository has outgrown this documentation is there new documentation notably this microsoft prebuilt image for jetson mcr microsoft com azureml onnxruntime v base does not include trt execution bindings is there a new supported image thank you for your assistance urgency yes this issue is blocking our production roadmap target platform jetson xaiver nx jetpack build script from nvcr io nvidia ml arg debian frontend noninteractive run apt get update run apt get install lshw y run apt get install git y run git clone recursive run export cudacxx usr local cuda bin nvcc run apt get update y run apt get install pip y run apt get install matplotlib y run apt get install gfortran y run apt get install y build essential libatlas base dev run apt get install ffmpeg y run apt install y no install recommends build essential software properties common libopenblas dev run wget run tar xf cmake linux tar gz run export path cmake linux bin path run apt remove cmake y run pwd run ln s cmake linux bin usr local bin run which cmake run cmake version env path usr local cuda bin path env ld library path usr local cuda ld library path run install upgrade pip run uninstall torch y run install torch run echo path echo ld library path workdir onnxruntime run build sh config release update build parallel build wheel use tensorrt cuda home usr local cuda cudnn home usr lib linux gnu tensorrt home usr lib linux gnu error output building cxx object deps protobuf build cmakefiles libprotobuf dir protobuf src src google protobuf wrappers pb cc o linking cxx static library libprotobuf a built target libprotobuf makefile recipe for target all failed make error traceback most recent call last file onnxruntime tools ci build build py line in sys exit main file onnxruntime tools ci build build py line in main build targets args cmake path build dir configs num parallel jobs args target file onnxruntime tools ci build build py line in build targets run subprocess cmd args env env file onnxruntime tools ci build build py line in run subprocess return run args cwd cwd capture stdout capture stdout shell shell env my env file onnxruntime tools python util run py line in run shell shell file usr lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status the command bin sh c onnxruntime build sh config release update build parallel build wheel use tensorrt cuda home usr local cuda cudnn home usr lib linux gnu tensorrt home usr lib linux gnu returned a non zero code visual studio version no response gcc compiler version no response | 0 |
44,839 | 12,403,328,383 | IssuesEvent | 2020-05-21 13:43:52 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | [DOCUMENTS]: Documents MUST be accessible, Covid Links and Resources PDF | 508-defect-2 508-issue-documents 508/Accessibility needs-grooming vsa vsa-public-websites | # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
```diff
! This is really a 508-defect-1, but since we need to coordinate with another
! team to find out who is responsible for creating the accessible PDF, I am
! setting this as a 2.
```
**Feedback framework**
- **โ๏ธ Must** for if the feedback must be applied
- **โ ๏ธShould** if the feedback is best practice
- **โ๏ธ Consider** for suggestions/enhancements
## Description
Documents **must** be accessible. The [Covid-19 Links and Resources PDF](https://www.va.gov/covid-19-links-and-resources.pdf) fails an accessibility check, and must be remediated so that all the content is accessible for screen readers and other assistive technology.
If this PDF is created by another team, they may need to remediate the PDF. I am not clear on the responsibility assignment. Depending on what application is used to create the PDF, there are export options that will support accessibility.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to access all of the content of PDFs, just as sighted users may.
## Environment
* Operating System: all
* Browser: all
* Server destination: production
## Steps to Recreate
1. Access the PDF at https://www.va.gov/covid-19-links-and-resources.pdf
1. Open the PDF in Adobe Acrobat Pro
1. Navigate to the Tools tab, and scroll to the bottom
1. In the second to the last row, select Accessibility
1. Run a Full Check
1. Verify there are Failed items, such as Tagged PDF, Title, in the Document; additional issues in Page Content, etc.
## Possible Fixes (optional)
Using the Adobe Acrobat Pro Accessibility Checker, follow the steps listed in the first link below to evaluate and remediate the PDF. Frankly, the first few times can really try your patience! After a while, it becomes clearer, and this is a valuable skill to add to your resume.
## WCAG or Vendor Guidance (optional)
* [Recommended: Using the Acrobat Pro DC Accessibility Checker](https://www.adobe.com/accessibility/products/acrobat/using-acrobat-pro-accessibility-checker.html)
* [HHS.gov - PDF File 508 Checklist (WCAG 2.0 Refresh)](https://www.hhs.gov/web/section-508/making-files-accessible/checklist/pdf/index.html)
* [Make Sure PDF Documents are Tagged for Accessibility, Deque University](https://dequeuniversity.com/tips/tagged-pdf-documents)
* [Requirements for an Accessible PDF: Part 1 (addl parts linked), Deque](https://www.deque.com/blog/requirements-accessible-pdf/)
* [PDF Accessibility: Everything You Need to Know, Deque](https://www.deque.com/blog/pdf-accessibility/)
* [W3C PDF Techniques for WCAG 2.0](https://www.w3.org/TR/WCAG20-TECHS/pdf)
* [Adobe PDF Accessibility Overview](https://www.adobe.com/accessibility/pdf.html)
## Screenshots or Trace Logs

| 1.0 | [DOCUMENTS]: Documents MUST be accessible, Covid Links and Resources PDF - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
```diff
! This is really a 508-defect-1, but since we need to coordinate with another
! team to find out who is responsible for creating the accessible PDF, I am
! setting this as a 2.
```
**Feedback framework**
- **โ๏ธ Must** for if the feedback must be applied
- **โ ๏ธShould** if the feedback is best practice
- **โ๏ธ Consider** for suggestions/enhancements
## Description
Documents **must** be accessible. The [Covid-19 Links and Resources PDF](https://www.va.gov/covid-19-links-and-resources.pdf) fails an accessibility check, and must be remediated so that all the content is accessible for screen readers and other assistive technology.
If this PDF is created by another team, they may need to remediate the PDF. I am not clear on the responsibility assignment. Depending on what application is used to create the PDF, there are export options that will support accessibility.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to access all of the content of PDFs, just as sighted users may.
## Environment
* Operating System: all
* Browser: all
* Server destination: production
## Steps to Recreate
1. Access the PDF at https://www.va.gov/covid-19-links-and-resources.pdf
1. Open the PDF in Adobe Acrobat Pro
1. Navigate to the Tools tab, and scroll to the bottom
1. In the second to the last row, select Accessibility
1. Run a Full Check
1. Verify there are Failed items, such as Tagged PDF, Title, in the Document; additional issues in Page Content, etc.
## Possible Fixes (optional)
Using the Adobe Acrobat Pro Accessibility Checker, follow the steps listed in the first link below to evaluate and remediate the PDF. Frankly, the first few times can really try your patience! After a while, it becomes clearer, and this is a valuable skill to add to your resume.
## WCAG or Vendor Guidance (optional)
* [Recommended: Using the Acrobat Pro DC Accessibility Checker](https://www.adobe.com/accessibility/products/acrobat/using-acrobat-pro-accessibility-checker.html)
* [HHS.gov - PDF File 508 Checklist (WCAG 2.0 Refresh)](https://www.hhs.gov/web/section-508/making-files-accessible/checklist/pdf/index.html)
* [Make Sure PDF Documents are Tagged for Accessibility, Deque University](https://dequeuniversity.com/tips/tagged-pdf-documents)
* [Requirements for an Accessible PDF: Part 1 (addl parts linked), Deque](https://www.deque.com/blog/requirements-accessible-pdf/)
* [PDF Accessibility: Everything You Need to Know, Deque](https://www.deque.com/blog/pdf-accessibility/)
* [W3C PDF Techniques for WCAG 2.0](https://www.w3.org/TR/WCAG20-TECHS/pdf)
* [Adobe PDF Accessibility Overview](https://www.adobe.com/accessibility/pdf.html)
## Screenshots or Trace Logs

| defect | documents must be accessible covid links and resources pdf diff this is really a defect but since we need to coordinate with another team to find out who is responsible for creating the accessible pdf i am setting this as a feedback framework โ๏ธ must for if the feedback must be applied โ ๏ธshould if the feedback is best practice โ๏ธ consider for suggestions enhancements description documents must be accessible the fails an accessibility check and must be remediated so that all the content is accessible for screen readers and other assistive technology if this pdf is created by another team they may need to remediate the pdf i am not clear on the responsibility assignment depending on what application is used to create the pdf there are export options that will support accessibility point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to access all of the content of pdfs just as sighted users may environment operating system all browser all server destination production steps to recreate access the pdf at open the pdf in adobe acrobat pro navigate to the tools tab and scroll to the bottom in the second to the last row select accessibility run a full check verify there are failed items such as tagged pdf title in the document additional issues in page content etc possible fixes optional using the adobe acrobat pro accessibility checker follow the steps listed in the first link below to evaluate and remediate the pdf frankly the first few times can really try your patience after a while it becomes clearer and this is a valuable skill to add to your resume wcag or vendor guidance optional screenshots or trace logs | 1 |
44,289 | 9,559,099,352 | IssuesEvent | 2019-05-03 15:46:28 | trufflesuite/ganache | https://api.github.com/repos/trufflesuite/ganache | closed | System Error when running Ganache 2.0.1 on darwin | truffle-decoder | This happens when I am clicking on one of my deployed contracts.
PLATFORM: darwin
GANACHE VERSION: 2.0.1
EXCEPTION:
```
TypeError: Cannot read property 'nodes' of undefined
at getStateVariables (/node_modules/truffle-decoder/dist/allocate/storage.js:146:24)
at vars.concat.linearizedBaseContractsFromBase.map (/node_modules/truffle-decoder/dist/allocate/storage.js:154:73)
at Array.map (native)
at allocateContract (/node_modules/truffle-decoder/dist/allocate/storage.js:154:61)
at Object.getStorageAllocations (/node_modules/truffle-decoder/dist/allocate/storage.js:28:23)
at TruffleContractDecoder.init (/node_modules/truffle-decoder/dist/interface/contract-decoder.js:85:43)
at Object.getContractState (/src/truffle-integration/decode.js:18:11)
at process.<anonymous> (/src/truffle-integration/index.js:76:31)
at emitTwo (events.js:125:13)
at process.emit (events.js:213:7)
at emit (internal/child_process.js:768:12)
at _combinedTickCallback (internal/process/next_tick.js:141:11)
at process._tickCallback (internal/process/next_tick.js:180:9)
```
APPLICATION LOG:
```
T+39957526ms: net_version
T+39957697ms: eth_sendTransaction
T+39957823ms: Transaction: 0x7a6fb1fd07eece85737553d247a6d373b746df178f77c1cf7ed14a6b3ae18a02
T+39957823ms: Contract created: 0x9db9528836587517cd6b90026b430f6e58063f64
T+39957823ms: Gas usage: 1250492
T+39957823ms: Block Number: 12
T+39957823ms: Block Time: Wed Apr 24 2019 09:20:34 GMT+0300 (EEST)
T+39957823ms: Runtime Error: revert
T+39957823ms: eth_call
T+39957925ms: eth_getBlockByNumber
T+39957925ms: eth_getBlockByNumber
T+39958030ms: eth_getBlockByNumber
T+39958191ms: eth_getBlockByNumber
T+39958191ms: eth_getBlockByNumber
T+40053858ms: net_version
T+40054236ms: eth_accounts
T+40054826ms: eth_getBlockByNumber
T+40054826ms: eth_accounts
T+40055240ms: eth_getBlockByNumber
T+40055240ms: eth_getBlockByNumber
T+40055785ms: eth_getBlockByNumber
T+40055785ms: eth_estimateGas
T+40056099ms: eth_getBlockByNumber
T+40056099ms: eth_blockNumber
T+40056500ms: eth_sendTransaction
T+40121694ms: Transaction: 0x6d3674c82a35d941f4a49ff409d2d1dc1c28a1f016283a6a592fd8914d6538df
T+40121694ms: Contract created: 0x9f2b851d683abe6d66bb552a50014ffbb94c60e7
T+40121694ms: Gas usage: 277462
T+40121694ms: Block Number: 13
T+40121694ms: Block Time: Wed Apr 24 2019 09:23:06 GMT+0300 (EEST)
T+40121694ms: eth_getTransactionReceipt
T+40121694ms: eth_getCode
T+40121694ms: eth_getTransactionByHash
T+40121878ms: eth_getBlockByNumber
T+40121878ms: eth_getBalance
T+40121878ms: eth_getBlockByNumber
T+40122016ms: eth_getBlockByNumber
T+40122150ms: eth_sendTransaction
T+40122352ms: Transaction: 0xa903f8b1bfac5ddc702356a9eb001fd7a5b6a87826151fbe114b25471cfff478
T+40122352ms: Gas usage: 42008
T+40122352ms: Block Number: 14
T+40122352ms: Block Time: Wed Apr 24 2019 09:23:06 GMT+0300 (EEST)
T+40122465ms: eth_getTransactionReceipt
T+40122575ms: eth_getBlockByNumber
T+40122705ms: eth_accounts
T+40122933ms: eth_getBlockByNumber
T+40123046ms: eth_getBlockByNumber
T+40123172ms: eth_getBlockByNumber
T+40123317ms: eth_estimateGas
T+40123502ms: eth_getBlockByNumber
T+40123502ms: eth_blockNumber
T+40123502ms: eth_sendTransaction
T+40123625ms: Transaction: 0x2fac23f33da86c7506a4d8539efd29a88baae0222f43b88f8b7c022cb1259d5c
T+40123625ms: Contract created: 0xc93fc6808cff5dddf88e46b78a5119b3db1d24e1
T+40123625ms: Gas usage: 953902
T+40123625ms: Block Number: 15
T+40123625ms: Block Time: Wed Apr 24 2019 09:23:07 GMT+0300 (EEST)
T+40123625ms: eth_getTransactionReceipt
T+40123625ms: eth_getCode
T+40123729ms: eth_getTransactionByHash
T+40123729ms: eth_getBlockByNumber
T+40123864ms: eth_getBalance
T+40124081ms: eth_getBlockByNumber
T+40124081ms: eth_getBlockByNumber
T+40124081ms: eth_sendTransaction
T+40124203ms: Transaction: 0x5ea3bcec61d6dd35b94b6a4da55e6459770985313719b955f50c88492a00d370
T+40124203ms: Gas usage: 27008
T+40124203ms: Block Number: 16
T+40124203ms: Block Time: Wed Apr 24 2019 09:23:07 GMT+0300 (EEST)
T+40124357ms: eth_getTransactionReceipt
T+40124486ms: eth_getBlockByNumber
T+40124668ms: eth_accounts
T+40124790ms: eth_getBlockByNumber
T+40124790ms: net_version
T+40124790ms: eth_getBlockByNumber
T+40124790ms: eth_getBlockByNumber
T+40124927ms: net_version
T+40125108ms: eth_estimateGas
T+40125294ms: eth_getBlockByNumber
T+40125410ms: eth_blockNumber
T+40125549ms: net_version
T+40125664ms: eth_sendTransaction
T+40125914ms: Transaction: 0xe20dd446b22263db722723767505115a1ce3f80fd516de00c068086d6fd08e11
T+40125914ms: Contract created: 0x4c6bafda49317d7451f00f11d5420a5d954e68c5
T+40125914ms: Gas usage: 4573266
T+40125914ms: Block Number: 17
T+40125914ms: Block Time: Wed Apr 24 2019 09:23:08 GMT+0300 (EEST)
T+40125914ms: eth_getTransactionReceipt
T+40125914ms: eth_getCode
T+40125914ms: eth_getTransactionByHash
T+40125914ms: eth_getBlockByNumber
T+40125914ms: eth_getBalance
T+40126054ms: eth_getBlockByNumber
T+40126245ms: eth_getBlockByNumber
T+40126449ms: eth_sendTransaction
T+40126647ms: Transaction: 0xe4b2bd6d6957c2c0bfbeb725f30a2f83bca66a3db320a943b5caa642bcbe93e4
T+40126647ms: Gas usage: 27008
T+40126647ms: Block Number: 18
T+40126647ms: Block Time: Wed Apr 24 2019 09:23:08 GMT+0300 (EEST)
T+40126780ms: eth_getTransactionReceipt
T+40126905ms: eth_getBlockByNumber
T+40126905ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40160025ms: eth_unsubscribe
T+40160025ms: eth_unsubscribe
T+40160190ms: eth_unsubscribe
T+40160190ms: eth_unsubscribe
``` | 1.0 | System Error when running Ganache 2.0.1 on darwin - This happens when I am clicking on one of my deployed contracts.
PLATFORM: darwin
GANACHE VERSION: 2.0.1
EXCEPTION:
```
TypeError: Cannot read property 'nodes' of undefined
at getStateVariables (/node_modules/truffle-decoder/dist/allocate/storage.js:146:24)
at vars.concat.linearizedBaseContractsFromBase.map (/node_modules/truffle-decoder/dist/allocate/storage.js:154:73)
at Array.map (native)
at allocateContract (/node_modules/truffle-decoder/dist/allocate/storage.js:154:61)
at Object.getStorageAllocations (/node_modules/truffle-decoder/dist/allocate/storage.js:28:23)
at TruffleContractDecoder.init (/node_modules/truffle-decoder/dist/interface/contract-decoder.js:85:43)
at Object.getContractState (/src/truffle-integration/decode.js:18:11)
at process.<anonymous> (/src/truffle-integration/index.js:76:31)
at emitTwo (events.js:125:13)
at process.emit (events.js:213:7)
at emit (internal/child_process.js:768:12)
at _combinedTickCallback (internal/process/next_tick.js:141:11)
at process._tickCallback (internal/process/next_tick.js:180:9)
```
APPLICATION LOG:
```
T+39957526ms: net_version
T+39957697ms: eth_sendTransaction
T+39957823ms: Transaction: 0x7a6fb1fd07eece85737553d247a6d373b746df178f77c1cf7ed14a6b3ae18a02
T+39957823ms: Contract created: 0x9db9528836587517cd6b90026b430f6e58063f64
T+39957823ms: Gas usage: 1250492
T+39957823ms: Block Number: 12
T+39957823ms: Block Time: Wed Apr 24 2019 09:20:34 GMT+0300 (EEST)
T+39957823ms: Runtime Error: revert
T+39957823ms: eth_call
T+39957925ms: eth_getBlockByNumber
T+39957925ms: eth_getBlockByNumber
T+39958030ms: eth_getBlockByNumber
T+39958191ms: eth_getBlockByNumber
T+39958191ms: eth_getBlockByNumber
T+40053858ms: net_version
T+40054236ms: eth_accounts
T+40054826ms: eth_getBlockByNumber
T+40054826ms: eth_accounts
T+40055240ms: eth_getBlockByNumber
T+40055240ms: eth_getBlockByNumber
T+40055785ms: eth_getBlockByNumber
T+40055785ms: eth_estimateGas
T+40056099ms: eth_getBlockByNumber
T+40056099ms: eth_blockNumber
T+40056500ms: eth_sendTransaction
T+40121694ms: Transaction: 0x6d3674c82a35d941f4a49ff409d2d1dc1c28a1f016283a6a592fd8914d6538df
T+40121694ms: Contract created: 0x9f2b851d683abe6d66bb552a50014ffbb94c60e7
T+40121694ms: Gas usage: 277462
T+40121694ms: Block Number: 13
T+40121694ms: Block Time: Wed Apr 24 2019 09:23:06 GMT+0300 (EEST)
T+40121694ms: eth_getTransactionReceipt
T+40121694ms: eth_getCode
T+40121694ms: eth_getTransactionByHash
T+40121878ms: eth_getBlockByNumber
T+40121878ms: eth_getBalance
T+40121878ms: eth_getBlockByNumber
T+40122016ms: eth_getBlockByNumber
T+40122150ms: eth_sendTransaction
T+40122352ms: Transaction: 0xa903f8b1bfac5ddc702356a9eb001fd7a5b6a87826151fbe114b25471cfff478
T+40122352ms: Gas usage: 42008
T+40122352ms: Block Number: 14
T+40122352ms: Block Time: Wed Apr 24 2019 09:23:06 GMT+0300 (EEST)
T+40122465ms: eth_getTransactionReceipt
T+40122575ms: eth_getBlockByNumber
T+40122705ms: eth_accounts
T+40122933ms: eth_getBlockByNumber
T+40123046ms: eth_getBlockByNumber
T+40123172ms: eth_getBlockByNumber
T+40123317ms: eth_estimateGas
T+40123502ms: eth_getBlockByNumber
T+40123502ms: eth_blockNumber
T+40123502ms: eth_sendTransaction
T+40123625ms: Transaction: 0x2fac23f33da86c7506a4d8539efd29a88baae0222f43b88f8b7c022cb1259d5c
T+40123625ms: Contract created: 0xc93fc6808cff5dddf88e46b78a5119b3db1d24e1
T+40123625ms: Gas usage: 953902
T+40123625ms: Block Number: 15
T+40123625ms: Block Time: Wed Apr 24 2019 09:23:07 GMT+0300 (EEST)
T+40123625ms: eth_getTransactionReceipt
T+40123625ms: eth_getCode
T+40123729ms: eth_getTransactionByHash
T+40123729ms: eth_getBlockByNumber
T+40123864ms: eth_getBalance
T+40124081ms: eth_getBlockByNumber
T+40124081ms: eth_getBlockByNumber
T+40124081ms: eth_sendTransaction
T+40124203ms: Transaction: 0x5ea3bcec61d6dd35b94b6a4da55e6459770985313719b955f50c88492a00d370
T+40124203ms: Gas usage: 27008
T+40124203ms: Block Number: 16
T+40124203ms: Block Time: Wed Apr 24 2019 09:23:07 GMT+0300 (EEST)
T+40124357ms: eth_getTransactionReceipt
T+40124486ms: eth_getBlockByNumber
T+40124668ms: eth_accounts
T+40124790ms: eth_getBlockByNumber
T+40124790ms: net_version
T+40124790ms: eth_getBlockByNumber
T+40124790ms: eth_getBlockByNumber
T+40124927ms: net_version
T+40125108ms: eth_estimateGas
T+40125294ms: eth_getBlockByNumber
T+40125410ms: eth_blockNumber
T+40125549ms: net_version
T+40125664ms: eth_sendTransaction
T+40125914ms: Transaction: 0xe20dd446b22263db722723767505115a1ce3f80fd516de00c068086d6fd08e11
T+40125914ms: Contract created: 0x4c6bafda49317d7451f00f11d5420a5d954e68c5
T+40125914ms: Gas usage: 4573266
T+40125914ms: Block Number: 17
T+40125914ms: Block Time: Wed Apr 24 2019 09:23:08 GMT+0300 (EEST)
T+40125914ms: eth_getTransactionReceipt
T+40125914ms: eth_getCode
T+40125914ms: eth_getTransactionByHash
T+40125914ms: eth_getBlockByNumber
T+40125914ms: eth_getBalance
T+40126054ms: eth_getBlockByNumber
T+40126245ms: eth_getBlockByNumber
T+40126449ms: eth_sendTransaction
T+40126647ms: Transaction: 0xe4b2bd6d6957c2c0bfbeb725f30a2f83bca66a3db320a943b5caa642bcbe93e4
T+40126647ms: Gas usage: 27008
T+40126647ms: Block Number: 18
T+40126647ms: Block Time: Wed Apr 24 2019 09:23:08 GMT+0300 (EEST)
T+40126780ms: eth_getTransactionReceipt
T+40126905ms: eth_getBlockByNumber
T+40126905ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40127067ms: eth_getBlockByNumber
T+40160025ms: eth_unsubscribe
T+40160025ms: eth_unsubscribe
T+40160190ms: eth_unsubscribe
T+40160190ms: eth_unsubscribe
``` | non_defect | system error when running ganache on darwin this happens when i am clicking on one of my deployed contracts platform darwin ganache version exception typeerror cannot read property nodes of undefined at getstatevariables node modules truffle decoder dist allocate storage js at vars concat linearizedbasecontractsfrombase map node modules truffle decoder dist allocate storage js at array map native at allocatecontract node modules truffle decoder dist allocate storage js at object getstorageallocations node modules truffle decoder dist allocate storage js at trufflecontractdecoder init node modules truffle decoder dist interface contract decoder js at object getcontractstate src truffle integration decode js at process src truffle integration index js at emittwo events js at process emit events js at emit internal child process js at combinedtickcallback internal process next tick js at process tickcallback internal process next tick js application log t net version t eth sendtransaction t transaction t contract created t gas usage t block number t block time wed apr gmt eest t runtime error revert t eth call t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t net version t eth accounts t eth getblockbynumber t eth accounts t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth estimategas t eth getblockbynumber t eth blocknumber t eth sendtransaction t transaction t contract created t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getcode t eth gettransactionbyhash t eth getblockbynumber t eth getbalance t eth getblockbynumber t eth getblockbynumber t eth sendtransaction t transaction t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getblockbynumber t eth accounts t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth estimategas t eth getblockbynumber t eth blocknumber t eth sendtransaction t transaction t contract created t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getcode t eth gettransactionbyhash t eth getblockbynumber t eth getbalance t eth getblockbynumber t eth getblockbynumber t eth sendtransaction t transaction t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getblockbynumber t eth accounts t eth getblockbynumber t net version t eth getblockbynumber t eth getblockbynumber t net version t eth estimategas t eth getblockbynumber t eth blocknumber t net version t eth sendtransaction t transaction t contract created t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getcode t eth gettransactionbyhash t eth getblockbynumber t eth getbalance t eth getblockbynumber t eth getblockbynumber t eth sendtransaction t transaction t gas usage t block number t block time wed apr gmt eest t eth gettransactionreceipt t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth getblockbynumber t eth unsubscribe t eth unsubscribe t eth unsubscribe t eth unsubscribe | 0 |
38,537 | 8,881,055,314 | IssuesEvent | 2019-01-14 08:57:21 | helight/xgcom | https://api.github.com/repos/helight/xgcom | closed | ็จไบarchlinux ็ PKGBUILD ๆๅ
ๆไปถ | Priority-Medium Type-Defect auto-migrated | ```
# Contributor: GunsNRose<3444542@163.com>
pkgname=xgcom
pkgver=11
pkgrel=1
pkgdesc="A gui tools to help guys develop series port, like minicom"
arch=('i686')
groups=('network')
url="http://code.google.com/p/xgcom"
license=('GPL2')
depends=('make' 'autoconf' 'automake' 'glib2' 'vte' 'gtk2')
makedepends=('subversion')
noextract=()
md5sums=() #generate with 'makepkg -g'
_svntrunk=http://xgcom.googlecode.com/svn/trunk
_svnmod=xgcom
build() {
cd "$srcdir"
if [ -d $_svnmod/.svn ]; then
(cd $_svnmod && svn up -r $pkgver)
else
svn co $_svntrunk --config-dir ./ -r $pkgver $_svnmod
fi
msg "SVN checkout done or server timeout"
msg "Starting make..."
rm -rf "$srcdir/$_svnmod-build"
cp -r "$srcdir/$_svnmod" "$srcdir/$_svnmod-build"
cd "$srcdir/$_svnmod-build"
#
# BUILD
#
./autogen.sh
./configure --prefix=/usr
make || return 1
make DESTDIR="$pkgdir/" install
#make INSTALL_ROOT=$pkgdir install || return 1
}
```
Original issue reported on code.google.com by `zlb...@gmail.com` on 1 Sep 2009 at 3:39
| 1.0 | ็จไบarchlinux ็ PKGBUILD ๆๅ
ๆไปถ - ```
# Contributor: GunsNRose<3444542@163.com>
pkgname=xgcom
pkgver=11
pkgrel=1
pkgdesc="A gui tools to help guys develop series port, like minicom"
arch=('i686')
groups=('network')
url="http://code.google.com/p/xgcom"
license=('GPL2')
depends=('make' 'autoconf' 'automake' 'glib2' 'vte' 'gtk2')
makedepends=('subversion')
noextract=()
md5sums=() #generate with 'makepkg -g'
_svntrunk=http://xgcom.googlecode.com/svn/trunk
_svnmod=xgcom
build() {
cd "$srcdir"
if [ -d $_svnmod/.svn ]; then
(cd $_svnmod && svn up -r $pkgver)
else
svn co $_svntrunk --config-dir ./ -r $pkgver $_svnmod
fi
msg "SVN checkout done or server timeout"
msg "Starting make..."
rm -rf "$srcdir/$_svnmod-build"
cp -r "$srcdir/$_svnmod" "$srcdir/$_svnmod-build"
cd "$srcdir/$_svnmod-build"
#
# BUILD
#
./autogen.sh
./configure --prefix=/usr
make || return 1
make DESTDIR="$pkgdir/" install
#make INSTALL_ROOT=$pkgdir install || return 1
}
```
Original issue reported on code.google.com by `zlb...@gmail.com` on 1 Sep 2009 at 3:39
| defect | ็จไบarchlinux ็ pkgbuild ๆๅ
ๆไปถ contributor gunsnrose pkgname xgcom pkgver pkgrel pkgdesc a gui tools to help guys develop series port like minicom arch groups network url license depends make autoconf automake vte makedepends subversion noextract generate with makepkg g svntrunk svnmod xgcom build cd srcdir if then cd svnmod svn up r pkgver else svn co svntrunk config dir r pkgver svnmod fi msg svn checkout done or server timeout msg starting make rm rf srcdir svnmod build cp r srcdir svnmod srcdir svnmod build cd srcdir svnmod build build autogen sh configure prefix usr make return make destdir pkgdir install make install root pkgdir install return original issue reported on code google com by zlb gmail com on sep at | 1 |
70,541 | 8,557,898,554 | IssuesEvent | 2018-11-08 16:46:57 | kubernetes-incubator/service-catalog | https://api.github.com/repos/kubernetes-incubator/service-catalog | closed | Fix svcat's terminology for name, k8's name, uuid, and id to match the server | kind/design svcat | When I first wrote svcat, I didn't fully understand the differences between k8's name, external name, k8s uid, etc. This has resulted in mixed terminology used in svcat and the server.
Here is the translation:
| svcat | api |
| name | external name |
| uuid | k8's name |
We want to continue to use name for the external name (otherwise everything is going to be unreadable), however using uuid for the k8's name is super confusing because k8's resources already has a uid.
It's not clear if anyone uses the ability of svcat to retrieve by either external name or k8's name? I would like to remove the `--uuid` filter flag entirely.
If people aren't ok with that, then please suggest an alternative name for that flag. ๐ | 1.0 | Fix svcat's terminology for name, k8's name, uuid, and id to match the server - When I first wrote svcat, I didn't fully understand the differences between k8's name, external name, k8s uid, etc. This has resulted in mixed terminology used in svcat and the server.
Here is the translation:
| svcat | api |
| name | external name |
| uuid | k8's name |
We want to continue to use name for the external name (otherwise everything is going to be unreadable), however using uuid for the k8's name is super confusing because k8's resources already has a uid.
It's not clear if anyone uses the ability of svcat to retrieve by either external name or k8's name? I would like to remove the `--uuid` filter flag entirely.
If people aren't ok with that, then please suggest an alternative name for that flag. ๐ | non_defect | fix svcat s terminology for name s name uuid and id to match the server when i first wrote svcat i didn t fully understand the differences between s name external name uid etc this has resulted in mixed terminology used in svcat and the server here is the translation svcat api name external name uuid s name we want to continue to use name for the external name otherwise everything is going to be unreadable however using uuid for the s name is super confusing because s resources already has a uid it s not clear if anyone uses the ability of svcat to retrieve by either external name or s name i would like to remove the uuid filter flag entirely if people aren t ok with that then please suggest an alternative name for that flag ๐ | 0 |
17,819 | 9,920,913,228 | IssuesEvent | 2019-06-30 13:50:55 | JurassiCraftTeam/JurassiCraft2 | https://api.github.com/repos/JurassiCraftTeam/JurassiCraft2 | closed | Major performance lag (over-spawning?) | 1.12.2 enhancement performance severe | In trying to put together a server, I have been troubleshooting quite a bit to improve performance. Currently we have an unplayable amount of lag - everything stutters, blocks broken reappear instantly and take a moment to realize they've been broken, and the server sends out a ton of "Can't keep up" messages. And the server's got pretty decent CPU and memory, too.
My collection of mods includes JurassiCraft (2.1.17), which we have been playing with the 'Dinosaur Spawning' option enabled (and 'Allow Carcass Spawning' disabled).
I pulled out JurassiCraft and dropped it on a new server as the only mod (except for LLibrary, of course). The performance was still just as bad, so it does seem to be an issue with JurassiCraft. I also noticed there were quite a lot of dinosaurs spawning (Triceratops in a Mesa biome - I'll attach a screenshot of one of the spots where they seem especially thick). I don't know if the lag is due to uncontrolled spawning, or if the dino code itself needs a bit more optimizing.

| True | Major performance lag (over-spawning?) - In trying to put together a server, I have been troubleshooting quite a bit to improve performance. Currently we have an unplayable amount of lag - everything stutters, blocks broken reappear instantly and take a moment to realize they've been broken, and the server sends out a ton of "Can't keep up" messages. And the server's got pretty decent CPU and memory, too.
My collection of mods includes JurassiCraft (2.1.17), which we have been playing with the 'Dinosaur Spawning' option enabled (and 'Allow Carcass Spawning' disabled).
I pulled out JurassiCraft and dropped it on a new server as the only mod (except for LLibrary, of course). The performance was still just as bad, so it does seem to be an issue with JurassiCraft. I also noticed there were quite a lot of dinosaurs spawning (Triceratops in a Mesa biome - I'll attach a screenshot of one of the spots where they seem especially thick). I don't know if the lag is due to uncontrolled spawning, or if the dino code itself needs a bit more optimizing.

| non_defect | major performance lag over spawning in trying to put together a server i have been troubleshooting quite a bit to improve performance currently we have an unplayable amount of lag everything stutters blocks broken reappear instantly and take a moment to realize they ve been broken and the server sends out a ton of can t keep up messages and the server s got pretty decent cpu and memory too my collection of mods includes jurassicraft which we have been playing with the dinosaur spawning option enabled and allow carcass spawning disabled i pulled out jurassicraft and dropped it on a new server as the only mod except for llibrary of course the performance was still just as bad so it does seem to be an issue with jurassicraft i also noticed there were quite a lot of dinosaurs spawning triceratops in a mesa biome i ll attach a screenshot of one of the spots where they seem especially thick i don t know if the lag is due to uncontrolled spawning or if the dino code itself needs a bit more optimizing | 0 |
75,377 | 25,795,915,704 | IssuesEvent | 2022-12-10 15:01:47 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | No ringing on incoming Jitsi group call | T-Defect | ### Steps to reproduce
Start a Jitsi group call in a room.
### Outcome
#### What did you expect?
The room participants should have their phone ringing.
#### What happened instead?
Room participants (Android Element) do not get a notification and mobile phones do not ring.
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
Element 1.5.8
### Homeserver
Synapse 1.73.0+focal1
### Will you send logs?
No
### Are you willing to provide a PR?
No | 1.0 | No ringing on incoming Jitsi group call - ### Steps to reproduce
Start a Jitsi group call in a room.
### Outcome
#### What did you expect?
The room participants should have their phone ringing.
#### What happened instead?
Room participants (Android Element) do not get a notification and mobile phones do not ring.
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
Element 1.5.8
### Homeserver
Synapse 1.73.0+focal1
### Will you send logs?
No
### Are you willing to provide a PR?
No | defect | no ringing on incoming jitsi group call steps to reproduce start a jitsi group call in a room outcome what did you expect the room participants should have their phone ringing what happened instead room participants android element do not get a notification and mobile phones do not ring your phone model no response operating system version no response application version and app store element homeserver synapse will you send logs no are you willing to provide a pr no | 1 |
67,401 | 20,961,609,033 | IssuesEvent | 2022-03-27 21:48:48 | abedmaatalla/sipdroid | https://api.github.com/repos/abedmaatalla/sipdroid | closed | incoming sound delay | Priority-Medium Type-Defect auto-migrated | ```
What steps will reproduce the problem?
1. A(sipdroid) call to B(ipphone)
2. B pickup and saying
3. A can't hear (3 second)
4. sound is good after 3 second
What is the expected output? What do you see instead?
What version of the product are you using? On what device/operating system?
1. A(sipdroid) => android 4.1.2 , device => LG-FL180L
2. B(ipphone) => moimstone ip255s
Which SIP server are you using? What happens with PBXes?
asterisk 11.4.0
Which type of network are you using?
WIFI => sipdroid
100Mbyte lan => ipphone
Please provide any additional information below.
- samsung device is good, only LG device is bad
- sipdroid ver => 3.0 Beta
```
Original issue reported on code.google.com by `ssik...@gmail.com` on 25 Jun 2013 at 2:11
| 1.0 | incoming sound delay - ```
What steps will reproduce the problem?
1. A(sipdroid) call to B(ipphone)
2. B pickup and saying
3. A can't hear (3 second)
4. sound is good after 3 second
What is the expected output? What do you see instead?
What version of the product are you using? On what device/operating system?
1. A(sipdroid) => android 4.1.2 , device => LG-FL180L
2. B(ipphone) => moimstone ip255s
Which SIP server are you using? What happens with PBXes?
asterisk 11.4.0
Which type of network are you using?
WIFI => sipdroid
100Mbyte lan => ipphone
Please provide any additional information below.
- samsung device is good, only LG device is bad
- sipdroid ver => 3.0 Beta
```
Original issue reported on code.google.com by `ssik...@gmail.com` on 25 Jun 2013 at 2:11
| defect | incoming sound delay what steps will reproduce the problem a sipdroid call to b ipphone b pickup and saying a can t hear second sound is good after second what is the expected output what do you see instead what version of the product are you using on what device operating system a sipdroid android device lg b ipphone moimstone which sip server are you using what happens with pbxes asterisk which type of network are you using wifi sipdroid lan ipphone please provide any additional information below samsung device is good only lg device is bad sipdroid ver beta original issue reported on code google com by ssik gmail com on jun at | 1 |
175,335 | 14,524,721,368 | IssuesEvent | 2020-12-14 11:51:05 | MakeContributions/markdown-dungeon | https://api.github.com/repos/MakeContributions/markdown-dungeon | closed | Link not found | bug documentation enhancement good first issue | I started the game, i ended up in a page
where it shows like this

When i click the link it took me to 404 page of github.
Please take a note. | 1.0 | Link not found - I started the game, i ended up in a page
where it shows like this

When i click the link it took me to 404 page of github.
Please take a note. | non_defect | link not found i started the game i ended up in a page where it shows like this when i click the link it took me to page of github please take a note | 0 |
60,668 | 17,023,488,407 | IssuesEvent | 2021-07-03 02:17:10 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [roads] leisure=track should not automatically be area | Component: mapnik Priority: major Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 2.23pm, Thursday, 1st October 2009]**
Closed ways with leisure=track should not automatically be rendered as areas, but only if they have area=yes set.
See #1900 for the same bug in osmarender. | 1.0 | [roads] leisure=track should not automatically be area - **[Submitted to the original trac issue database at 2.23pm, Thursday, 1st October 2009]**
Closed ways with leisure=track should not automatically be rendered as areas, but only if they have area=yes set.
See #1900 for the same bug in osmarender. | defect | leisure track should not automatically be area closed ways with leisure track should not automatically be rendered as areas but only if they have area yes set see for the same bug in osmarender | 1 |
343,142 | 30,653,301,380 | IssuesEvent | 2023-07-25 10:20:30 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_numpy_logic.test_jax_isnan | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5486981878"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5493324879"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5484385318"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5477523919"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5479218847"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_numpy_logic.test_jax_isnan - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5486981878"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5493324879"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5484385318"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5477523919"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5479218847"><img src=https://img.shields.io/badge/-success-success></a>
| non_defect | fix jax numpy logic test jax isnan tensorflow a href src torch a href src numpy a href src jax a href src paddle a href src | 0 |
52,726 | 13,224,979,687 | IssuesEvent | 2020-08-17 20:14:38 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Evaluate impact of unfiltered events (Trac #244) | Migrated from Trac defect jeb + pnf | From pnf review
5) As discussed in Sec. 6.4 PnF is challenged by ``Big Eventsโโ that can be processed with delays larger than 10 s due to sophisticated reconstructions. Technical feasibility of online processing are reviewed and approved by the TFT board. It is recommended that for such proposals, including online ToO proposals, detailed testing and feedback is provided. Such proposals should use aggressive pre-filtering to ensure that the rate of these events is kept low. This is what has been achieved until now. If this will not be possible in the future, technical solutions could be found (for example the online alert processing might be decoupled from PnF, with the results of the online reconstruction not recorded in main filtered data stream), but this would require considerable currently unbudgeted development work from PnF and SPADE.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/244">https://code.icecube.wisc.edu/projects/icecube/ticket/244</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-25T13:53:17",
"_ts": "1337953997000000",
"description": "From pnf review\n\n5) As discussed in Sec. 6.4 PnF is challenged by ``Big Events\u2019\u2019 that can be processed with delays larger than 10 s due to sophisticated reconstructions. Technical feasibility of online processing are reviewed and approved by the TFT board. It is recommended that for such proposals, including online ToO proposals, detailed testing and feedback is provided. Such proposals should use aggressive pre-filtering to ensure that the rate of these events is kept low. This is what has been achieved until now. If this will not be possible in the future, technical solutions could be found (for example the online alert processing might be decoupled from PnF, with the results of the online reconstruction not recorded in main filtered data stream), but this would require considerable currently unbudgeted development work from PnF and SPADE.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2011-02-19T16:00:43",
"component": "jeb + pnf",
"summary": "Evaluate impact of unfiltered events",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Evaluate impact of unfiltered events (Trac #244) - From pnf review
5) As discussed in Sec. 6.4 PnF is challenged by ``Big Eventsโโ that can be processed with delays larger than 10 s due to sophisticated reconstructions. Technical feasibility of online processing are reviewed and approved by the TFT board. It is recommended that for such proposals, including online ToO proposals, detailed testing and feedback is provided. Such proposals should use aggressive pre-filtering to ensure that the rate of these events is kept low. This is what has been achieved until now. If this will not be possible in the future, technical solutions could be found (for example the online alert processing might be decoupled from PnF, with the results of the online reconstruction not recorded in main filtered data stream), but this would require considerable currently unbudgeted development work from PnF and SPADE.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/244">https://code.icecube.wisc.edu/projects/icecube/ticket/244</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-25T13:53:17",
"_ts": "1337953997000000",
"description": "From pnf review\n\n5) As discussed in Sec. 6.4 PnF is challenged by ``Big Events\u2019\u2019 that can be processed with delays larger than 10 s due to sophisticated reconstructions. Technical feasibility of online processing are reviewed and approved by the TFT board. It is recommended that for such proposals, including online ToO proposals, detailed testing and feedback is provided. Such proposals should use aggressive pre-filtering to ensure that the rate of these events is kept low. This is what has been achieved until now. If this will not be possible in the future, technical solutions could be found (for example the online alert processing might be decoupled from PnF, with the results of the online reconstruction not recorded in main filtered data stream), but this would require considerable currently unbudgeted development work from PnF and SPADE.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2011-02-19T16:00:43",
"component": "jeb + pnf",
"summary": "Evaluate impact of unfiltered events",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
| defect | evaluate impact of unfiltered events trac from pnf review as discussed in sec pnf is challenged by big eventsโโ that can be processed with delays larger than s due to sophisticated reconstructions technical feasibility of online processing are reviewed and approved by the tft board it is recommended that for such proposals including online too proposals detailed testing and feedback is provided such proposals should use aggressive pre filtering to ensure that the rate of these events is kept low this is what has been achieved until now if this will not be possible in the future technical solutions could be found for example the online alert processing might be decoupled from pnf with the results of the online reconstruction not recorded in main filtered data stream but this would require considerable currently unbudgeted development work from pnf and spade migrated from json status closed changetime ts description from pnf review n as discussed in sec pnf is challenged by big events that can be processed with delays larger than s due to sophisticated reconstructions technical feasibility of online processing are reviewed and approved by the tft board it is recommended that for such proposals including online too proposals detailed testing and feedback is provided such proposals should use aggressive pre filtering to ensure that the rate of these events is kept low this is what has been achieved until now if this will not be possible in the future technical solutions could be found for example the online alert processing might be decoupled from pnf with the results of the online reconstruction not recorded in main filtered data stream but this would require considerable currently unbudgeted development work from pnf and spade reporter blaufuss cc resolution fixed time component jeb pnf summary evaluate impact of unfiltered events priority normal keywords milestone owner tschmidt type defect | 1 |
218,805 | 24,400,845,102 | IssuesEvent | 2022-10-05 01:07:01 | SitecorePowerShell/Console | https://api.github.com/repos/SitecorePowerShell/Console | closed | XSS - ListView | bug area-user-interface area-security | ### Expected Behavior
No XSS
### Actual Behavior
XSS
### Steps to Reproduce the Problem
```powershell
$item = gi -path '/sitecore/content/Home'
$item.Title = '<script>alert(1)</script>'
$item | Show-ListView -Title "Test" -Property @{ Label = 'label'; Expression = { $_.Title } }
``` | True | XSS - ListView - ### Expected Behavior
No XSS
### Actual Behavior
XSS
### Steps to Reproduce the Problem
```powershell
$item = gi -path '/sitecore/content/Home'
$item.Title = '<script>alert(1)</script>'
$item | Show-ListView -Title "Test" -Property @{ Label = 'label'; Expression = { $_.Title } }
``` | non_defect | xss listview expected behavior no xss actual behavior xss steps to reproduce the problem powershell item gi path sitecore content home item title alert item show listview title test property label label expression title | 0 |
5,471 | 2,610,188,347 | IssuesEvent | 2015-02-26 18:59:40 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | ๅผ่่ฒๆๅฆไฝๆทกๅ | auto-migrated Priority-Medium Type-Defect | ```
ใๆ่ฆใ
ๆๆถๅ๏ผๅธๆๆถ้ดไธบ่ชๅทฑๅไธ๏ผๅฐฑ่ฟๆ ทๅๅๆฌข็ไบบๅฐ่ๅคฉ่๏ฟฝ๏ฟฝ
๏ฟฝๆๆถๅ๏ผๅ็ฐ่บซ่พน็ไบบ้ฝไธไบ่งฃ่ชๅทฑ๏ผ้ขๅฏน็่บซ่พน็ไบบ๏ผ็ช๏ฟฝ
๏ฟฝ๏ฟฝ่งๅพ่ฏดไธๅบ่ฏ๏ผๆๆถๅ๏ผๅจ่ชๅทฑ่ๅผฑ็ๆถๅ๏ผๆณไธไธชไบบ่บฒ่ตท
ๆฅ๏ผไธๆฟๅซไบบ็ๅฐ่ชๅทฑ็ไผคๅฃ๏ผๆๆถๅ๏ผ็ช็ถๅพๆณ้็ฆป็ฐๅจ๏ฟฝ๏ฟฝ
๏ฟฝ็ๆดป๏ผๆณไธ้กพไธๅๆถๆพ่ชๅทฑ็ฎๅ็่กๆๅปๆตๆตชใๅพๅค็พๅฅณๅฏน๏ฟฝ
๏ฟฝ๏ฟฝๅฆไฝ็ฅๆ้ฝๆฏ่พๅ
ณๆณจ๏ผๆฏ็ซ่ฐไนไธๆณ่ธไธ้ฟๅพๅค้พ็็ๆ๏ผ
้ฃไนๅฐๅบๅไปไนๅฏไปฅ็ฅๆ้ฃ่ฒๆๅฆไฝๆทกๅ๏ผ
ใๅฎขๆทๆกไพใ
ใใ็ทไบบไผ้ฟ้ป่คๆๅ,
ๅคๅคฉๆฅไบๆ็นๅ ้ไบ๏ผๆไปฌๅบ่ฏฅๆไนๅป้ค่ธไธ็่ฒๆ๏ผ็ฅๆ๏ฟฝ๏ฟฝ
๏ฟฝไปไนๅฅฝ็ๆนๆณ๏ผๅฆไฝ็ฅๆๆๆๆๅฅฝ๏ผๆไน่ฝๅฝปๅบๅป้ค่ธไธ็๏ฟฝ
๏ฟฝ๏ฟฝ็น๏ผ่ฟไบ้ฎ้ข้ฝๆฏ่ฒๆๆฃ่
ๆๆณ้ฎ็๏ผๆ่ธไธ็ๆ็นๅป้คไบ
่ฟไบๅฐฑไธๆฏ้ฎ้ขไบใๅคง็ญ็ๅคฉ๏ผ่ธไธ็ฒๅบๆถฒใ้ฎ็้ไธๅฑๅฑ๏ฟฝ๏ฟฝ
๏ฟฝ้ฎไบๅ้ฎ๏ผๆนไบๅๆน๏ผ็ปไบๅฏไปฅๅบ้จไบ๏ผๅฏไธๅบๆฑ๏ผๆดไธชไบบ๏ฟฝ
๏ฟฝ๏ฟฝ่ทไธชๅฑๆ็ๅฐไธ๏ผๆๆปๆฏๅจๅซไบบๆๅฅๅๅฒ็ฌ็่กจๆ
ไธญ่ฝ่่
้๏ผไฝฟ็จไบๆ ๆฐ็ง็ฅๆไบงๅ๏ผไฝ้ป่คๆๅง็ปๆฒกๆ็ฆปๅผ้ฃๅผ ไน๏ฟฝ๏ฟฝ
๏ฟฝไฝๅฟง้็่ธใๅจ่ๅคๅง๏ผๅฅณๅญฉๅญๆๆ็ๅฟๅผๅงๆ็ป่ทๅซไบบไบค๏ฟฝ
๏ฟฝ๏ฟฝ๏ผ็ฌๆฅ็ฌๅพ๏ผๅฐฑ่ฟๆ ท่บฒ่บฒ่่็็ปๆไบๆ็ๅคงๅญฆ็ๆดปใ</br>
ใใๆไธบไบ็ฅๆ็ๆฏๅไบไธๅฐ็ๅๅคซ๏ผ็ดๅฐๆไธๆฌกๅฆนๅฆนๅบๅทฎ๏ฟฝ๏ฟฝ
๏ฟฝๆฅ๏ผๅ
ดๅฅ็ๆ็ๆ่ฏดๆพๅฐไบไธๆฌพ้ๅธธไธ้็็ฅๆไบงๅ๏ผๅซ้ป๏ฟฝ
๏ฟฝ๏ฟฝ่ๅฐๅฅน็ไธไธชๅฎขๆท็จ็ๅฐฑๆฏ่ฟไธช๏ผ็ฅๆๆๆๅพไธ้็๏ผ่ฟๆฌก
่งๅฐๅฅนๆ นๆฌ็ไธๅบๅๆฅๅฅน่ธไธ็้ฃไบ่ฒๆ๏ผ็ฎ่คไนๆ่งๆฏๅ๏ฟฝ๏ฟฝ
๏ฟฝ็ฝ็ๅพๅค๏ผไบๆฏๆ็ๆไธ็ฝๆพๅฐไปไปฌ็ๅฎ็ฝ๏ผๅจ่ฏขๅฎขๆๅๆ๏ฟฝ
๏ฟฝ๏ฟฝ่ดญไบไธคไธชๅจๆใ</br>
ใใๆ
ขๆ
ข็็จไธไธชๅจๆไนๅ๏ผๅ็ฐๆๆๅผๅงๆ
ขๆ
ข็ๅๆทก๏ผๆ๏ฟฝ๏ฟฝ
๏ฟฝ็กฎๅฎไธ้๏ผ่ไธๅฏน่บซไฝไนๆฒกๆไปปไฝ็ๅฝฑๅ๏ผๅๆฅๅๆฅ็ไนฐไบ๏ฟฝ
๏ฟฝ๏ฟฝไธชๅจๆใไธๅ
ฑไธๆฅ็จๅฎไบ3ไธชๅจๆ่ธไธๆๅบๆฌๅทฒ็ป็ไธๆธ
ๆฅ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผ่ไธ็ฎ่คไนๆฏไปฅๅๅนฒๅๅ
ๆณฝๅคไบ๏ผ่ไธ่ฟๅพๆๅผนๆง๏ผ่ฟไธช๏ฟฝ
๏ฟฝ๏ฟฝ่่ๅฐ็จ่ตทๆฅ่ฟๆฏๆบๆพๅฟ็๏ผๅฎๅ
จ่ฟๆฏๅพ้่ฆ็๏ผๅพ็ๅฅฝๅฎ
ใ
้
่ฏปไบ่ฒๆๅฆไฝๆทกๅ๏ผๅ็่ธไธๅฎนๆ้ฟๆ็ๅๅ ๏ผ
ใ่ฒๆๅฝขๆๅๅ ใ
ใใๅ
้จๅ ็ด
ใใไธใๅๅ
ใใๅฝไบบๅๅฐๅๅๆถ๏ผๅฐฑไผๅๆณ่พไธ่
บ็ด ๏ผไธบๅฏนไปๅๅ่ๅ๏ฟฝ๏ฟฝ
๏ฟฝๅคใๅฆๆ้ฟๆๅๅฐๅๅ๏ผไบบไฝๆฐ้ไปฃ่ฐข็ๅนณ่กกๅฐฑไผ้ญๅฐ็ ดๅ๏ฟฝ
๏ฟฝ๏ฟฝ็ฎ่คๆ้็่ฅๅ
ปไพๅบ่ถไบ็ผๆ
ข๏ผ่ฒ็ด ๆฏ็ป่ๅฐฑไผๅๅพๅพๆดป่ท
ใ
ใใไบใ่ทๅฐ่ๅๆณๅคฑ่ฐ
ใใ้ฟๅญ่ฏ้ๆๅซ็ๅฅณๆง่ทๅฐ่้ๆฟ็ด ๏ผไผๅบๆฟ้บฆๆๅฎ็ป่๏ฟฝ๏ฟฝ
๏ฟฝๅๆณ่ๅฝขๆไธๅๅ็ๆ็น๏ผๅ ้ฟๅญ่ฏ่ๅฝขๆ็ๆ็น๏ผ่ฝ็ถๅจ๏ฟฝ
๏ฟฝ๏ฟฝ่ฏไธญๆญๅไผๅๆญข๏ผไฝไปไผๅจ็ฎ่คไธๅ็ๅพ้ฟไธๆฎตๆถ้ดใๆๅญ
ไธญๅ ๅฅณๆง่ทๅฐ่้ๆฟ็ด ็ๅขๅ ๏ผไปๆๅญ4โ5ไธชๆๅผๅงไผๅฎนๆๅบ
็ฐๆ๏ผ่ฟๆถๅๅบ็ฐ็ๆ็นๅจไบงๅๅคง้จๅไผๆถๅคฑใๅฏๆฏ๏ผๆฐ้๏ฟฝ๏ฟฝ
๏ฟฝ่ฐขไธๆญฃๅธธใ่่ค่ฃธ้ฒๅจๅผบ็็็ดซๅค็บฟไธใ็ฒพ็ฅไธๅๅฐๅๅ็ญ๏ฟฝ
๏ฟฝ๏ฟฝๅ ๏ผ้ฝไผไฝฟๆๅ ๆทฑใๆๆถๆฐ้ฟๅบ็ๆ๏ผไบงๅไนไธไผๆถๅคฑ๏ผๆ
ไปฅ้่ฆๆดๅ ๆณจๆใ
ใใไธใๆฐ้ไปฃ่ฐข็ผๆ
ข
ใใ่็ๆฐ้ไปฃ่ฐขๅ่ฝไธๆญฃๅธธๆๅตๅทขๅ่ฝๅ้ๆถไนไผๅบ็ฐๆ๏ฟฝ๏ฟฝ
๏ฟฝๅ ไธบๆฐ้ไปฃ่ฐขไธ้กบ็
ใๆๅ
ๅๆณๅคฑ่ฐ๏ผไฝฟ่บซไฝๅคไบๆๆ็ถๆ๏ฟฝ
๏ฟฝ๏ฟฝ๏ผไป่ๅ ๅง่ฒ็ด ้ฎ้ขใๆไปฌๅธธ่ฏด็ไพฟ็งไผๅฝขๆๆ๏ผๅ
ถๅฎๅฐฑๆฏ
ๅ
ๅๆณๅคฑ่ฐๅฏผ่ด่ฟๆไฝ่ดจ่ๅฝขๆ็ใๅฆๅค๏ผ่บซไฝ็ถๆไธๆญฃๅธธ๏ฟฝ๏ฟฝ
๏ฟฝๆถๅ๏ผ็ดซๅค็บฟ็็
งๅฐไนไผๅ ้ๆ็ๅฝขๆใ
ใใๅใ้่ฏฏ็ไฝฟ็จๅๅฆๅ
ใใไฝฟ็จไบไธ้ๅ่ชๅทฑ็ฎ่ค็ๅๅฆๅ๏ผไผๅฏผ่ด็ฎ่ค่ฟๆใๅจ๏ฟฝ๏ฟฝ
๏ฟฝ็็่ฟ็จไธญๅฆ่ฟ้็
งๅฐๅฐ็ดซๅค็บฟ๏ผ็ฎ่คไผไธบไบๆตๅพกๅค็็ไพต๏ฟฝ
๏ฟฝ๏ฟฝ๏ผๅจๆ็็็้จไฝ่้้บฆๆๅฎ่ฒ็ด ๏ผ่ฟๆ ทไผๅบ็ฐ่ฒ็ด ๆฒ็็
้ฎ้ขใ
ใใๅค้จๅ ็ด
ใใไธใ็ดซๅค็บฟ
ใใ็
งๅฐ็ดซๅค็บฟ็ๆถๅ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผไผๅจๅบๅบๅฑไบง๏ฟฝ๏ฟฝ
๏ฟฝๅพๅค้บฆๆๅฎ่ฒ็ด ใๆไปฅไธบไบไฟๆค็ฎ่ค๏ผไผๅจๆๆ้จไฝ่้ๆด๏ฟฝ
๏ฟฝ๏ฟฝ็่ฒ็ด ใ็ปๅธธ่ฃธ้ฒๅจๅผบ็็้ณๅ
ๅบไธไธไป
ไฟ่ฟ็ฎ่ค็่ๅ๏ผ
่ฟไผๅผ่ตท้ปๆใ้ๆ็ญ่ฒ็ด ๆฒ็็็ฎ่ค็พๆฃใ
ใใไบใไธ่ฏ็ๆธ
ๆดไน ๆฏ
ใใๅ ๅผบ็็ๆธ
ๆดไน ๆฏไฝฟ็ฎ่คๅๅพๆๆ๏ผ่ฟๆ ทไผๅบๆฟ็ฎ่คใ๏ฟฝ๏ฟฝ
๏ฟฝ็ฎ่คๆๆๆถ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผ้ป่ฒ็ด ็ป่ไผๅๆณๅพๅค้บฆ๏ฟฝ
๏ฟฝ๏ฟฝๅฎ่ฒ็ด ๏ผๅฝ่ฒ็ด ่ฟๅฉๆถๅฐฑๅบ็ฐไบๆใ็็ต็ญ็ฎ่ค่ฒ็ด ๆฒ็็
้ฎ้ขใ
ใใไธใ้ไผ ๅบๅ
ใใ็ถๆฏไธญๆ้ฟๆ็๏ผๅๆฌไบบ้ฟๆ็ๆฆ็ๅฐฑๅพ้ซ๏ผ่ฟ็งๆ
ๅต๏ฟฝ๏ฟฝ
๏ฟฝไธๅฎ็จๅบฆไธๅฐฑๅฏๅคๅฎๆฏ้ไผ ๅบๅ ็ไฝ็จใๆไปฅๅฎถ้็นๅซๆฏ้ฟ๏ฟฝ
๏ฟฝ๏ฟฝๆ้ฟๆ็ไบบ๏ผ่ฆๆณจๆ้ฟๅ
ๅผๅ้ฟๆ็้่ฆๅ ็ด ไนไธโโ็ดซๅค
็บฟ็
งๅฐ๏ผ่ฟๆฏ้ข้ฒๆๅฟ
้กปๆณจๆ็ใ
ใๆ็้ฎๅธฎไฝ ่งฃๅณใ
ใใ1,้ป่่ๅฐ็ฒพๅๆถฒ็็ๆๆๆๅ?็็ๅฏไปฅๆ่ธไธ็้ป่ค๏ฟฝ๏ฟฝ
๏ฟฝๅปๆๅ?
ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒDNA็ฒพๅ่ฝๅคๆๆ็ไฟฎๅคๅจๅด้พไปฅ่งฆ๏ฟฝ๏ฟฝ
๏ฟฝ็่ฒๆ๏ผๅ
ถ็ฌๆ็็บณ่ฑๆๅไธบ็ฎ่ค็็พ็ฝไธ้ไธฝ๏ผๆไพไบๅฟ
๏ฟฝ
๏ฟฝ๏ฟฝๅฏๅฐ็่ฅๅ
ป็ฉ่ดจ๏ผๅฏไปฅๆๆ็ๅป้ค้ป่คๆ๏ผ้ป่คๆ๏ผ้ป่คๆ
๏ผ่ด่ถๆ๏ผๆๆใๅฆๅจ ๆ็ญใๅฎๅฎๅฎๅ
จ็ช็ ดไบไผ ็ป็็พ่คๆถ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผๅฎๅฆๅจ็ฎ่คไธญๆณจๅ
ฅไบไธๆฏๅ
ผๅ
ทๆดปๅใๅ็ใๆปๅ
ป็ญๅๆ็๏ฟฝ
๏ฟฝ๏ฟฝๅฐพ้
๏ผๅๆถไธบ่ธ้จๆไพๅคง้ๆๆบ็ปด็็ด ็ฒพๅ๏ผ่ธ้จ็ๆนๅๆพ
่ๆ่งใ่ชไบงๅไธๅธไปฅๆฅ๏ผ่้กพๅฎข็บท็บทไป็ปๆฐ้กพๅฎข๏ผ71%็ๆฐ๏ฟฝ๏ฟฝ
๏ฟฝๅฎข้ฝๆฏ้่ฟ่้กพๅฎขไป็ป่ๆฅ๏ผๅฃ็ข็ฑๆญค่ๆฅ!
ใใ2๏ผๆ็จ้ป่่ๅฐ็พ็ฝ๏ผไผไผค่บซไฝๅ?ๆๅฏไฝ็จๅ?
ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒๅบ็จไบ็ฒพ็บฏๅคๅ้
ๆนๅ้ขๅ
็ๅ็ฑป๏ฟฝ๏ฟฝ
๏ฟฝๆ็งๆ๏ผๅนถๅฐโDNA็พ่ค็ณป็ปโ็ๆณๅบ็จๅฐไบ่ฏฅไบงๅไธญ๏ผ่ฝๅฝป๏ฟฝ
๏ฟฝ๏ฟฝ็ฅ้ค้ป่คๆ๏ผ่ด่ถๆ๏ผๅฆๅจ ๆ๏ผๆๆ๏ผ้ป่คๆ๏ผ่ๅนดๆ๏ผๆ
ๆๆทกๅ้ป่คๆ่ณๆฅ่ฟ่ค่ฒใ้ป่่ๅฐ้่ฟๆณๅฝใ็พๅฝใๅฐๆนพ๏ฟฝ๏ฟฝ
๏ฟฝๅฐ็ไธๅฎถ้ๅๅไฝ๏ผ่ถ
่ฟ10ๅนด็็ ็ฉถไปฅๅ
จๆฐ็DNA่่คไฟฎๅคๆ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผๆๆไผ ็ปๅๅญฆๆค่ค็ๅฟต๏ผไธๆ่ฟฝๅฏปๅ็ฐ็ ด่ฏๅคง่ช็ถ็็พไธฝ๏ฟฝ
๏ฟฝ๏ฟฝ่ฟน๏ผไปคๆฏไธไฝ็ฑ็พ็ๅฅณๆง้ฝ่ฝไบซๅๅฐ็งๆๅๆฐๆๅธฆๆฅ็่ช็ถ
ไน็พใ
ไธไธบไบๆดฒๅฅณๆง่ค่ดจ็ ๅถ๏ผ็ฒพๅฟๅตๆคๅฅณๆง็พไธฝ๏ผๅคๅนดๆฅ๏ผไธบๆฐ๏ฟฝ๏ฟฝ
๏ฟฝ็พไธ่ฎก็ๅฅณๆง่งฃ้คไบ้ป่คๆๅฐๆฐใๆทฑๅพๅนฟๅคงๅฅณๆงๆๅ็ไฟก่ต!
ใใ3๏ผๅป้ค้ป่คๆไนๅ๏ผไผๅๅผนๅ?
ใใ็ญ๏ผๅพๅคๆพ็ป้ฟไบ้ป่คๆ็ไบบๅฃซ๏ผ่ชไป้ๆฉไบ้ป่่ๅฐ๏ฟฝ๏ฟฝ
๏ฟฝ็ฝ๏ผๅฐฑไธๅณๆฐธ้ธใ่ฟๆฌพ็ฅๆไบงๅๆฏ็ป่ฟๆฐๅไฝๆๅจ็ฅๆไธๅฎถ๏ฟฝ
๏ฟฝ๏ฟฝๆฎๆ็ๅฝขๆๅๅ ็ฒพๅฟ็ ๅถ่ๆ็จไบๅฎ่ฏด่ฏ๏ผ่ฎฉๆถ่ดน่
ๆๅใ
ๆ ็ซๆๅจๅ็!ๆไปฌ็ๅพๅคๆฐๅฎขๆท้ฝๆฏ่ๅฎขๆทไป็ป่ๆฅ๏ผ่ฏท้ฎ๏ฟฝ
๏ฟฝ๏ฟฝๅฆๆๆๆไธๅฅฝ๏ผไผๆๅฎขๆท่ฝฌไป็ปๅ?
ใใ4๏ผไฝ ไปฌ็ไปทๆ ผๆ็น่ดต๏ผ่ฝไธ่ฝไพฟๅฎไธ็น?
ใใ็ญ๏ผๅฆๆๆจไฝฟ็จ่ฅฟ่ฏๆๅฐ้่ฆ2000ๅ
๏ผ็
ๆ็่ฏๆๅฐ้่ฆ3
000ๅ
๏ผๅๆๆฏๆๅฐๆฏ5000ๅ
๏ผ่่ฟไบๆฏซๆ ็้ฎ๏ผไธไผๅฏนๅฝปๅบๅป๏ฟฝ
๏ฟฝ๏ฟฝไฝ ็ๆ็นๆไปปไฝๅธฎๅฉ!ไธๅไปท้ฑ๏ผไธไปฝไปทๅผ๏ผๆไปฌ็ฐๅจๅ็๏ฟฝ๏ฟฝ
๏ฟฝๆฏไธไธชๅฃ็ข๏ผไธไธชๅ็๏ผไปท้ฑๅนถไธ้ซใๅฆๆ่ฑ่ฟ็น้ฑๆไฝ ็๏ฟฝ
๏ฟฝ๏ฟฝ่คๆๅฝปๅบๅป้ค๏ผไฝ ่ฟไผ่งๅพ่ดตๅ?ไฝ ่ฟไผๅๅป่ฑ้ฃไนๅคๅคๆ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผไธไฝๆๆฒกๅปๆ๏ผ่ฟๆ่ชๅทฑ็็ฎ่คๅผ็่ถๆฅ่ถ็ณๅ
ใใ5๏ผๆ้ๅ็จ้ป่่ๅฐ็ฒพๅๆถฒๅ?
ใใ็ญ๏ผ้ป่่ๅฐ้็จไบบ็พค๏ผ
ใใ1ใ็็็ดไนฑๅผ่ตท็้ป่คๆไบบ็พค
ใใ2ใ็่ฒๅผ่ตท็ๅฆๅจ ๆไบบ็พค
ใใ3ใๅนด็บชๅข้ฟๅผ่ตท็่ๅนดๆไบบ็พค
ใใ4ใๅๅฆๅ่ฒ็ด ๆฒ็งฏใ่พๅฐๆไบบ็พค
ใใ5ใ้ฟๆๆฅ็
งๅผ่ตท็ๆฅๆๆไบบ็พค
ใใ6ใ่่คๆๆทกๆฅ้็พ็ฝ็ไบบ็พค
ใ็ฅๆๅฐๆนๆณใ
่ฒๆๅฆไฝๆทกๅ๏ผๅๆถไธบๆจๅไบซ็ฅๆๅฐๆนๆณ
1.ๅฐไบบๅ่็ๆตใ็ปด็็ด CใEๆณจๅฐๆถฒๅๅๅฐๆถๅจๆดๅ็้ข้จ๏ผ๏ฟฝ
๏ฟฝ๏ฟฝๆณๅฝๅ่็ฒ250ๅ
ๅๆธฉๆฐด่ฐๆ็ณ็ถ๏ผ่ฟ
้ๆท็้ข้จ๏ผๅ่็ฒไผ
่ช่กๅ็กฌใๅ็ญใๅทๅด๏ผ30ๅ้ๅๅปๆ๏ผๅฏๆๆๅ้้ข้จๆใ
2ใๅฌ็ๅญไป15ใๆฉ็ฎ6ใๆก่ฑ12๏ผๆททๅ็ ๆ็ปๆซ๏ผ้ฅญๅ็จ็ฑณๆฑค๏ฟฝ๏ฟฝ
๏ฟฝๆ๏ผไธๆฅไธๆฌก๏ผ่ฟๆๆฐๆ๏ผ้ข้จไผๅๅพ็ฝๅซฉ่ๅ
ๆปใ
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:26 | 1.0 | ๅผ่่ฒๆๅฆไฝๆทกๅ - ```
ใๆ่ฆใ
ๆๆถๅ๏ผๅธๆๆถ้ดไธบ่ชๅทฑๅไธ๏ผๅฐฑ่ฟๆ ทๅๅๆฌข็ไบบๅฐ่ๅคฉ่๏ฟฝ๏ฟฝ
๏ฟฝๆๆถๅ๏ผๅ็ฐ่บซ่พน็ไบบ้ฝไธไบ่งฃ่ชๅทฑ๏ผ้ขๅฏน็่บซ่พน็ไบบ๏ผ็ช๏ฟฝ
๏ฟฝ๏ฟฝ่งๅพ่ฏดไธๅบ่ฏ๏ผๆๆถๅ๏ผๅจ่ชๅทฑ่ๅผฑ็ๆถๅ๏ผๆณไธไธชไบบ่บฒ่ตท
ๆฅ๏ผไธๆฟๅซไบบ็ๅฐ่ชๅทฑ็ไผคๅฃ๏ผๆๆถๅ๏ผ็ช็ถๅพๆณ้็ฆป็ฐๅจ๏ฟฝ๏ฟฝ
๏ฟฝ็ๆดป๏ผๆณไธ้กพไธๅๆถๆพ่ชๅทฑ็ฎๅ็่กๆๅปๆตๆตชใๅพๅค็พๅฅณๅฏน๏ฟฝ
๏ฟฝ๏ฟฝๅฆไฝ็ฅๆ้ฝๆฏ่พๅ
ณๆณจ๏ผๆฏ็ซ่ฐไนไธๆณ่ธไธ้ฟๅพๅค้พ็็ๆ๏ผ
้ฃไนๅฐๅบๅไปไนๅฏไปฅ็ฅๆ้ฃ่ฒๆๅฆไฝๆทกๅ๏ผ
ใๅฎขๆทๆกไพใ
ใใ็ทไบบไผ้ฟ้ป่คๆๅ,
ๅคๅคฉๆฅไบๆ็นๅ ้ไบ๏ผๆไปฌๅบ่ฏฅๆไนๅป้ค่ธไธ็่ฒๆ๏ผ็ฅๆ๏ฟฝ๏ฟฝ
๏ฟฝไปไนๅฅฝ็ๆนๆณ๏ผๅฆไฝ็ฅๆๆๆๆๅฅฝ๏ผๆไน่ฝๅฝปๅบๅป้ค่ธไธ็๏ฟฝ
๏ฟฝ๏ฟฝ็น๏ผ่ฟไบ้ฎ้ข้ฝๆฏ่ฒๆๆฃ่
ๆๆณ้ฎ็๏ผๆ่ธไธ็ๆ็นๅป้คไบ
่ฟไบๅฐฑไธๆฏ้ฎ้ขไบใๅคง็ญ็ๅคฉ๏ผ่ธไธ็ฒๅบๆถฒใ้ฎ็้ไธๅฑๅฑ๏ฟฝ๏ฟฝ
๏ฟฝ้ฎไบๅ้ฎ๏ผๆนไบๅๆน๏ผ็ปไบๅฏไปฅๅบ้จไบ๏ผๅฏไธๅบๆฑ๏ผๆดไธชไบบ๏ฟฝ
๏ฟฝ๏ฟฝ่ทไธชๅฑๆ็ๅฐไธ๏ผๆๆปๆฏๅจๅซไบบๆๅฅๅๅฒ็ฌ็่กจๆ
ไธญ่ฝ่่
้๏ผไฝฟ็จไบๆ ๆฐ็ง็ฅๆไบงๅ๏ผไฝ้ป่คๆๅง็ปๆฒกๆ็ฆปๅผ้ฃๅผ ไน๏ฟฝ๏ฟฝ
๏ฟฝไฝๅฟง้็่ธใๅจ่ๅคๅง๏ผๅฅณๅญฉๅญๆๆ็ๅฟๅผๅงๆ็ป่ทๅซไบบไบค๏ฟฝ
๏ฟฝ๏ฟฝ๏ผ็ฌๆฅ็ฌๅพ๏ผๅฐฑ่ฟๆ ท่บฒ่บฒ่่็็ปๆไบๆ็ๅคงๅญฆ็ๆดปใ</br>
ใใๆไธบไบ็ฅๆ็ๆฏๅไบไธๅฐ็ๅๅคซ๏ผ็ดๅฐๆไธๆฌกๅฆนๅฆนๅบๅทฎ๏ฟฝ๏ฟฝ
๏ฟฝๆฅ๏ผๅ
ดๅฅ็ๆ็ๆ่ฏดๆพๅฐไบไธๆฌพ้ๅธธไธ้็็ฅๆไบงๅ๏ผๅซ้ป๏ฟฝ
๏ฟฝ๏ฟฝ่ๅฐๅฅน็ไธไธชๅฎขๆท็จ็ๅฐฑๆฏ่ฟไธช๏ผ็ฅๆๆๆๅพไธ้็๏ผ่ฟๆฌก
่งๅฐๅฅนๆ นๆฌ็ไธๅบๅๆฅๅฅน่ธไธ็้ฃไบ่ฒๆ๏ผ็ฎ่คไนๆ่งๆฏๅ๏ฟฝ๏ฟฝ
๏ฟฝ็ฝ็ๅพๅค๏ผไบๆฏๆ็ๆไธ็ฝๆพๅฐไปไปฌ็ๅฎ็ฝ๏ผๅจ่ฏขๅฎขๆๅๆ๏ฟฝ
๏ฟฝ๏ฟฝ่ดญไบไธคไธชๅจๆใ</br>
ใใๆ
ขๆ
ข็็จไธไธชๅจๆไนๅ๏ผๅ็ฐๆๆๅผๅงๆ
ขๆ
ข็ๅๆทก๏ผๆ๏ฟฝ๏ฟฝ
๏ฟฝ็กฎๅฎไธ้๏ผ่ไธๅฏน่บซไฝไนๆฒกๆไปปไฝ็ๅฝฑๅ๏ผๅๆฅๅๆฅ็ไนฐไบ๏ฟฝ
๏ฟฝ๏ฟฝไธชๅจๆใไธๅ
ฑไธๆฅ็จๅฎไบ3ไธชๅจๆ่ธไธๆๅบๆฌๅทฒ็ป็ไธๆธ
ๆฅ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผ่ไธ็ฎ่คไนๆฏไปฅๅๅนฒๅๅ
ๆณฝๅคไบ๏ผ่ไธ่ฟๅพๆๅผนๆง๏ผ่ฟไธช๏ฟฝ
๏ฟฝ๏ฟฝ่่ๅฐ็จ่ตทๆฅ่ฟๆฏๆบๆพๅฟ็๏ผๅฎๅ
จ่ฟๆฏๅพ้่ฆ็๏ผๅพ็ๅฅฝๅฎ
ใ
้
่ฏปไบ่ฒๆๅฆไฝๆทกๅ๏ผๅ็่ธไธๅฎนๆ้ฟๆ็ๅๅ ๏ผ
ใ่ฒๆๅฝขๆๅๅ ใ
ใใๅ
้จๅ ็ด
ใใไธใๅๅ
ใใๅฝไบบๅๅฐๅๅๆถ๏ผๅฐฑไผๅๆณ่พไธ่
บ็ด ๏ผไธบๅฏนไปๅๅ่ๅ๏ฟฝ๏ฟฝ
๏ฟฝๅคใๅฆๆ้ฟๆๅๅฐๅๅ๏ผไบบไฝๆฐ้ไปฃ่ฐข็ๅนณ่กกๅฐฑไผ้ญๅฐ็ ดๅ๏ฟฝ
๏ฟฝ๏ฟฝ็ฎ่คๆ้็่ฅๅ
ปไพๅบ่ถไบ็ผๆ
ข๏ผ่ฒ็ด ๆฏ็ป่ๅฐฑไผๅๅพๅพๆดป่ท
ใ
ใใไบใ่ทๅฐ่ๅๆณๅคฑ่ฐ
ใใ้ฟๅญ่ฏ้ๆๅซ็ๅฅณๆง่ทๅฐ่้ๆฟ็ด ๏ผไผๅบๆฟ้บฆๆๅฎ็ป่๏ฟฝ๏ฟฝ
๏ฟฝๅๆณ่ๅฝขๆไธๅๅ็ๆ็น๏ผๅ ้ฟๅญ่ฏ่ๅฝขๆ็ๆ็น๏ผ่ฝ็ถๅจ๏ฟฝ
๏ฟฝ๏ฟฝ่ฏไธญๆญๅไผๅๆญข๏ผไฝไปไผๅจ็ฎ่คไธๅ็ๅพ้ฟไธๆฎตๆถ้ดใๆๅญ
ไธญๅ ๅฅณๆง่ทๅฐ่้ๆฟ็ด ็ๅขๅ ๏ผไปๆๅญ4โ5ไธชๆๅผๅงไผๅฎนๆๅบ
็ฐๆ๏ผ่ฟๆถๅๅบ็ฐ็ๆ็นๅจไบงๅๅคง้จๅไผๆถๅคฑใๅฏๆฏ๏ผๆฐ้๏ฟฝ๏ฟฝ
๏ฟฝ่ฐขไธๆญฃๅธธใ่่ค่ฃธ้ฒๅจๅผบ็็็ดซๅค็บฟไธใ็ฒพ็ฅไธๅๅฐๅๅ็ญ๏ฟฝ
๏ฟฝ๏ฟฝๅ ๏ผ้ฝไผไฝฟๆๅ ๆทฑใๆๆถๆฐ้ฟๅบ็ๆ๏ผไบงๅไนไธไผๆถๅคฑ๏ผๆ
ไปฅ้่ฆๆดๅ ๆณจๆใ
ใใไธใๆฐ้ไปฃ่ฐข็ผๆ
ข
ใใ่็ๆฐ้ไปฃ่ฐขๅ่ฝไธๆญฃๅธธๆๅตๅทขๅ่ฝๅ้ๆถไนไผๅบ็ฐๆ๏ฟฝ๏ฟฝ
๏ฟฝๅ ไธบๆฐ้ไปฃ่ฐขไธ้กบ็
ใๆๅ
ๅๆณๅคฑ่ฐ๏ผไฝฟ่บซไฝๅคไบๆๆ็ถๆ๏ฟฝ
๏ฟฝ๏ฟฝ๏ผไป่ๅ ๅง่ฒ็ด ้ฎ้ขใๆไปฌๅธธ่ฏด็ไพฟ็งไผๅฝขๆๆ๏ผๅ
ถๅฎๅฐฑๆฏ
ๅ
ๅๆณๅคฑ่ฐๅฏผ่ด่ฟๆไฝ่ดจ่ๅฝขๆ็ใๅฆๅค๏ผ่บซไฝ็ถๆไธๆญฃๅธธ๏ฟฝ๏ฟฝ
๏ฟฝๆถๅ๏ผ็ดซๅค็บฟ็็
งๅฐไนไผๅ ้ๆ็ๅฝขๆใ
ใใๅใ้่ฏฏ็ไฝฟ็จๅๅฆๅ
ใใไฝฟ็จไบไธ้ๅ่ชๅทฑ็ฎ่ค็ๅๅฆๅ๏ผไผๅฏผ่ด็ฎ่ค่ฟๆใๅจ๏ฟฝ๏ฟฝ
๏ฟฝ็็่ฟ็จไธญๅฆ่ฟ้็
งๅฐๅฐ็ดซๅค็บฟ๏ผ็ฎ่คไผไธบไบๆตๅพกๅค็็ไพต๏ฟฝ
๏ฟฝ๏ฟฝ๏ผๅจๆ็็็้จไฝ่้้บฆๆๅฎ่ฒ็ด ๏ผ่ฟๆ ทไผๅบ็ฐ่ฒ็ด ๆฒ็็
้ฎ้ขใ
ใใๅค้จๅ ็ด
ใใไธใ็ดซๅค็บฟ
ใใ็
งๅฐ็ดซๅค็บฟ็ๆถๅ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผไผๅจๅบๅบๅฑไบง๏ฟฝ๏ฟฝ
๏ฟฝๅพๅค้บฆๆๅฎ่ฒ็ด ใๆไปฅไธบไบไฟๆค็ฎ่ค๏ผไผๅจๆๆ้จไฝ่้ๆด๏ฟฝ
๏ฟฝ๏ฟฝ็่ฒ็ด ใ็ปๅธธ่ฃธ้ฒๅจๅผบ็็้ณๅ
ๅบไธไธไป
ไฟ่ฟ็ฎ่ค็่ๅ๏ผ
่ฟไผๅผ่ตท้ปๆใ้ๆ็ญ่ฒ็ด ๆฒ็็็ฎ่ค็พๆฃใ
ใใไบใไธ่ฏ็ๆธ
ๆดไน ๆฏ
ใใๅ ๅผบ็็ๆธ
ๆดไน ๆฏไฝฟ็ฎ่คๅๅพๆๆ๏ผ่ฟๆ ทไผๅบๆฟ็ฎ่คใ๏ฟฝ๏ฟฝ
๏ฟฝ็ฎ่คๆๆๆถ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผ้ป่ฒ็ด ็ป่ไผๅๆณๅพๅค้บฆ๏ฟฝ
๏ฟฝ๏ฟฝๅฎ่ฒ็ด ๏ผๅฝ่ฒ็ด ่ฟๅฉๆถๅฐฑๅบ็ฐไบๆใ็็ต็ญ็ฎ่ค่ฒ็ด ๆฒ็็
้ฎ้ขใ
ใใไธใ้ไผ ๅบๅ
ใใ็ถๆฏไธญๆ้ฟๆ็๏ผๅๆฌไบบ้ฟๆ็ๆฆ็ๅฐฑๅพ้ซ๏ผ่ฟ็งๆ
ๅต๏ฟฝ๏ฟฝ
๏ฟฝไธๅฎ็จๅบฆไธๅฐฑๅฏๅคๅฎๆฏ้ไผ ๅบๅ ็ไฝ็จใๆไปฅๅฎถ้็นๅซๆฏ้ฟ๏ฟฝ
๏ฟฝ๏ฟฝๆ้ฟๆ็ไบบ๏ผ่ฆๆณจๆ้ฟๅ
ๅผๅ้ฟๆ็้่ฆๅ ็ด ไนไธโโ็ดซๅค
็บฟ็
งๅฐ๏ผ่ฟๆฏ้ข้ฒๆๅฟ
้กปๆณจๆ็ใ
ใๆ็้ฎๅธฎไฝ ่งฃๅณใ
ใใ1,้ป่่ๅฐ็ฒพๅๆถฒ็็ๆๆๆๅ?็็ๅฏไปฅๆ่ธไธ็้ป่ค๏ฟฝ๏ฟฝ
๏ฟฝๅปๆๅ?
ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒDNA็ฒพๅ่ฝๅคๆๆ็ไฟฎๅคๅจๅด้พไปฅ่งฆ๏ฟฝ๏ฟฝ
๏ฟฝ็่ฒๆ๏ผๅ
ถ็ฌๆ็็บณ่ฑๆๅไธบ็ฎ่ค็็พ็ฝไธ้ไธฝ๏ผๆไพไบๅฟ
๏ฟฝ
๏ฟฝ๏ฟฝๅฏๅฐ็่ฅๅ
ป็ฉ่ดจ๏ผๅฏไปฅๆๆ็ๅป้ค้ป่คๆ๏ผ้ป่คๆ๏ผ้ป่คๆ
๏ผ่ด่ถๆ๏ผๆๆใๅฆๅจ ๆ็ญใๅฎๅฎๅฎๅ
จ็ช็ ดไบไผ ็ป็็พ่คๆถ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผๅฎๅฆๅจ็ฎ่คไธญๆณจๅ
ฅไบไธๆฏๅ
ผๅ
ทๆดปๅใๅ็ใๆปๅ
ป็ญๅๆ็๏ฟฝ
๏ฟฝ๏ฟฝๅฐพ้
๏ผๅๆถไธบ่ธ้จๆไพๅคง้ๆๆบ็ปด็็ด ็ฒพๅ๏ผ่ธ้จ็ๆนๅๆพ
่ๆ่งใ่ชไบงๅไธๅธไปฅๆฅ๏ผ่้กพๅฎข็บท็บทไป็ปๆฐ้กพๅฎข๏ผ71%็ๆฐ๏ฟฝ๏ฟฝ
๏ฟฝๅฎข้ฝๆฏ้่ฟ่้กพๅฎขไป็ป่ๆฅ๏ผๅฃ็ข็ฑๆญค่ๆฅ!
ใใ2๏ผๆ็จ้ป่่ๅฐ็พ็ฝ๏ผไผไผค่บซไฝๅ?ๆๅฏไฝ็จๅ?
ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒๅบ็จไบ็ฒพ็บฏๅคๅ้
ๆนๅ้ขๅ
็ๅ็ฑป๏ฟฝ๏ฟฝ
๏ฟฝๆ็งๆ๏ผๅนถๅฐโDNA็พ่ค็ณป็ปโ็ๆณๅบ็จๅฐไบ่ฏฅไบงๅไธญ๏ผ่ฝๅฝป๏ฟฝ
๏ฟฝ๏ฟฝ็ฅ้ค้ป่คๆ๏ผ่ด่ถๆ๏ผๅฆๅจ ๆ๏ผๆๆ๏ผ้ป่คๆ๏ผ่ๅนดๆ๏ผๆ
ๆๆทกๅ้ป่คๆ่ณๆฅ่ฟ่ค่ฒใ้ป่่ๅฐ้่ฟๆณๅฝใ็พๅฝใๅฐๆนพ๏ฟฝ๏ฟฝ
๏ฟฝๅฐ็ไธๅฎถ้ๅๅไฝ๏ผ่ถ
่ฟ10ๅนด็็ ็ฉถไปฅๅ
จๆฐ็DNA่่คไฟฎๅคๆ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผๆๆไผ ็ปๅๅญฆๆค่ค็ๅฟต๏ผไธๆ่ฟฝๅฏปๅ็ฐ็ ด่ฏๅคง่ช็ถ็็พไธฝ๏ฟฝ
๏ฟฝ๏ฟฝ่ฟน๏ผไปคๆฏไธไฝ็ฑ็พ็ๅฅณๆง้ฝ่ฝไบซๅๅฐ็งๆๅๆฐๆๅธฆๆฅ็่ช็ถ
ไน็พใ
ไธไธบไบๆดฒๅฅณๆง่ค่ดจ็ ๅถ๏ผ็ฒพๅฟๅตๆคๅฅณๆง็พไธฝ๏ผๅคๅนดๆฅ๏ผไธบๆฐ๏ฟฝ๏ฟฝ
๏ฟฝ็พไธ่ฎก็ๅฅณๆง่งฃ้คไบ้ป่คๆๅฐๆฐใๆทฑๅพๅนฟๅคงๅฅณๆงๆๅ็ไฟก่ต!
ใใ3๏ผๅป้ค้ป่คๆไนๅ๏ผไผๅๅผนๅ?
ใใ็ญ๏ผๅพๅคๆพ็ป้ฟไบ้ป่คๆ็ไบบๅฃซ๏ผ่ชไป้ๆฉไบ้ป่่ๅฐ๏ฟฝ๏ฟฝ
๏ฟฝ็ฝ๏ผๅฐฑไธๅณๆฐธ้ธใ่ฟๆฌพ็ฅๆไบงๅๆฏ็ป่ฟๆฐๅไฝๆๅจ็ฅๆไธๅฎถ๏ฟฝ
๏ฟฝ๏ฟฝๆฎๆ็ๅฝขๆๅๅ ็ฒพๅฟ็ ๅถ่ๆ็จไบๅฎ่ฏด่ฏ๏ผ่ฎฉๆถ่ดน่
ๆๅใ
ๆ ็ซๆๅจๅ็!ๆไปฌ็ๅพๅคๆฐๅฎขๆท้ฝๆฏ่ๅฎขๆทไป็ป่ๆฅ๏ผ่ฏท้ฎ๏ฟฝ
๏ฟฝ๏ฟฝๅฆๆๆๆไธๅฅฝ๏ผไผๆๅฎขๆท่ฝฌไป็ปๅ?
ใใ4๏ผไฝ ไปฌ็ไปทๆ ผๆ็น่ดต๏ผ่ฝไธ่ฝไพฟๅฎไธ็น?
ใใ็ญ๏ผๅฆๆๆจไฝฟ็จ่ฅฟ่ฏๆๅฐ้่ฆ2000ๅ
๏ผ็
ๆ็่ฏๆๅฐ้่ฆ3
000ๅ
๏ผๅๆๆฏๆๅฐๆฏ5000ๅ
๏ผ่่ฟไบๆฏซๆ ็้ฎ๏ผไธไผๅฏนๅฝปๅบๅป๏ฟฝ
๏ฟฝ๏ฟฝไฝ ็ๆ็นๆไปปไฝๅธฎๅฉ!ไธๅไปท้ฑ๏ผไธไปฝไปทๅผ๏ผๆไปฌ็ฐๅจๅ็๏ฟฝ๏ฟฝ
๏ฟฝๆฏไธไธชๅฃ็ข๏ผไธไธชๅ็๏ผไปท้ฑๅนถไธ้ซใๅฆๆ่ฑ่ฟ็น้ฑๆไฝ ็๏ฟฝ
๏ฟฝ๏ฟฝ่คๆๅฝปๅบๅป้ค๏ผไฝ ่ฟไผ่งๅพ่ดตๅ?ไฝ ่ฟไผๅๅป่ฑ้ฃไนๅคๅคๆ๏ฟฝ๏ฟฝ
๏ฟฝ๏ผไธไฝๆๆฒกๅปๆ๏ผ่ฟๆ่ชๅทฑ็็ฎ่คๅผ็่ถๆฅ่ถ็ณๅ
ใใ5๏ผๆ้ๅ็จ้ป่่ๅฐ็ฒพๅๆถฒๅ?
ใใ็ญ๏ผ้ป่่ๅฐ้็จไบบ็พค๏ผ
ใใ1ใ็็็ดไนฑๅผ่ตท็้ป่คๆไบบ็พค
ใใ2ใ็่ฒๅผ่ตท็ๅฆๅจ ๆไบบ็พค
ใใ3ใๅนด็บชๅข้ฟๅผ่ตท็่ๅนดๆไบบ็พค
ใใ4ใๅๅฆๅ่ฒ็ด ๆฒ็งฏใ่พๅฐๆไบบ็พค
ใใ5ใ้ฟๆๆฅ็
งๅผ่ตท็ๆฅๆๆไบบ็พค
ใใ6ใ่่คๆๆทกๆฅ้็พ็ฝ็ไบบ็พค
ใ็ฅๆๅฐๆนๆณใ
่ฒๆๅฆไฝๆทกๅ๏ผๅๆถไธบๆจๅไบซ็ฅๆๅฐๆนๆณ
1.ๅฐไบบๅ่็ๆตใ็ปด็็ด CใEๆณจๅฐๆถฒๅๅๅฐๆถๅจๆดๅ็้ข้จ๏ผ๏ฟฝ
๏ฟฝ๏ฟฝๆณๅฝๅ่็ฒ250ๅ
ๅๆธฉๆฐด่ฐๆ็ณ็ถ๏ผ่ฟ
้ๆท็้ข้จ๏ผๅ่็ฒไผ
่ช่กๅ็กฌใๅ็ญใๅทๅด๏ผ30ๅ้ๅๅปๆ๏ผๅฏๆๆๅ้้ข้จๆใ
2ใๅฌ็ๅญไป15ใๆฉ็ฎ6ใๆก่ฑ12๏ผๆททๅ็ ๆ็ปๆซ๏ผ้ฅญๅ็จ็ฑณๆฑค๏ฟฝ๏ฟฝ
๏ฟฝๆ๏ผไธๆฅไธๆฌก๏ผ่ฟๆๆฐๆ๏ผ้ข้จไผๅๅพ็ฝๅซฉ่ๅ
ๆปใ
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:26 | defect | ๅผ่่ฒๆๅฆไฝๆทกๅ ใๆ่ฆใ ๆๆถๅ๏ผๅธๆๆถ้ดไธบ่ชๅทฑๅไธ๏ผๅฐฑ่ฟๆ ทๅๅๆฌข็ไบบๅฐ่ๅคฉ่๏ฟฝ๏ฟฝ ๏ฟฝๆๆถๅ๏ผๅ็ฐ่บซ่พน็ไบบ้ฝไธไบ่งฃ่ชๅทฑ๏ผ้ขๅฏน็่บซ่พน็ไบบ๏ผ็ช๏ฟฝ ๏ฟฝ๏ฟฝ่งๅพ่ฏดไธๅบ่ฏ๏ผๆๆถๅ๏ผๅจ่ชๅทฑ่ๅผฑ็ๆถๅ๏ผๆณไธไธชไบบ่บฒ่ตท ๆฅ๏ผไธๆฟๅซไบบ็ๅฐ่ชๅทฑ็ไผคๅฃ๏ผๆๆถๅ๏ผ็ช็ถๅพๆณ้็ฆป็ฐๅจ๏ฟฝ๏ฟฝ ๏ฟฝ็ๆดป๏ผๆณไธ้กพไธๅๆถๆพ่ชๅทฑ็ฎๅ็่กๆๅปๆตๆตชใๅพๅค็พๅฅณๅฏน๏ฟฝ ๏ฟฝ๏ฟฝๅฆไฝ็ฅๆ้ฝๆฏ่พๅ
ณๆณจ๏ผๆฏ็ซ่ฐไนไธๆณ่ธไธ้ฟๅพๅค้พ็็ๆ๏ผ ้ฃไนๅฐๅบๅไปไนๅฏไปฅ็ฅๆ้ฃ่ฒๆๅฆไฝๆทกๅ๏ผ ใๅฎขๆทๆกไพใ ใใ็ทไบบไผ้ฟ้ป่คๆๅ ๅคๅคฉๆฅไบๆ็นๅ ้ไบ๏ผๆไปฌๅบ่ฏฅๆไนๅป้ค่ธไธ็่ฒๆ๏ผ็ฅๆ๏ฟฝ๏ฟฝ ๏ฟฝไปไนๅฅฝ็ๆนๆณ๏ผๅฆไฝ็ฅๆๆๆๆๅฅฝ๏ผๆไน่ฝๅฝปๅบๅป้ค่ธไธ็๏ฟฝ ๏ฟฝ๏ฟฝ็น๏ผ่ฟไบ้ฎ้ข้ฝๆฏ่ฒๆๆฃ่
ๆๆณ้ฎ็๏ผๆ่ธไธ็ๆ็นๅป้คไบ ่ฟไบๅฐฑไธๆฏ้ฎ้ขไบใๅคง็ญ็ๅคฉ๏ผ่ธไธ็ฒๅบๆถฒใ้ฎ็้ไธๅฑๅฑ๏ฟฝ๏ฟฝ ๏ฟฝ้ฎไบๅ้ฎ๏ผๆนไบๅๆน๏ผ็ปไบๅฏไปฅๅบ้จไบ๏ผๅฏไธๅบๆฑ๏ผๆดไธชไบบ๏ฟฝ ๏ฟฝ๏ฟฝ่ทไธชๅฑๆ็ๅฐไธ๏ผๆๆปๆฏๅจๅซไบบๆๅฅๅๅฒ็ฌ็่กจๆ
ไธญ่ฝ่่ ้๏ผไฝฟ็จไบๆ ๆฐ็ง็ฅๆไบงๅ๏ผไฝ้ป่คๆๅง็ปๆฒกๆ็ฆปๅผ้ฃๅผ ไน๏ฟฝ๏ฟฝ ๏ฟฝไฝๅฟง้็่ธใๅจ่ๅคๅง๏ผๅฅณๅญฉๅญๆๆ็ๅฟๅผๅงๆ็ป่ทๅซไบบไบค๏ฟฝ ๏ฟฝ๏ฟฝ๏ผ็ฌๆฅ็ฌๅพ๏ผๅฐฑ่ฟๆ ท่บฒ่บฒ่่็็ปๆไบๆ็ๅคงๅญฆ็ๆดปใ ใใๆไธบไบ็ฅๆ็ๆฏๅไบไธๅฐ็ๅๅคซ๏ผ็ดๅฐๆไธๆฌกๅฆนๅฆนๅบๅทฎ๏ฟฝ๏ฟฝ ๏ฟฝๆฅ๏ผๅ
ดๅฅ็ๆ็ๆ่ฏดๆพๅฐไบไธๆฌพ้ๅธธไธ้็็ฅๆไบงๅ๏ผๅซ้ป๏ฟฝ ๏ฟฝ๏ฟฝ่ๅฐๅฅน็ไธไธชๅฎขๆท็จ็ๅฐฑๆฏ่ฟไธช๏ผ็ฅๆๆๆๅพไธ้็๏ผ่ฟๆฌก ่งๅฐๅฅนๆ นๆฌ็ไธๅบๅๆฅๅฅน่ธไธ็้ฃไบ่ฒๆ๏ผ็ฎ่คไนๆ่งๆฏๅ๏ฟฝ๏ฟฝ ๏ฟฝ็ฝ็ๅพๅค๏ผไบๆฏๆ็ๆไธ็ฝๆพๅฐไปไปฌ็ๅฎ็ฝ๏ผๅจ่ฏขๅฎขๆๅๆ๏ฟฝ ๏ฟฝ๏ฟฝ่ดญไบไธคไธชๅจๆใ ใใๆ
ขๆ
ข็็จไธไธชๅจๆไนๅ๏ผๅ็ฐๆๆๅผๅงๆ
ขๆ
ข็ๅๆทก๏ผๆ๏ฟฝ๏ฟฝ ๏ฟฝ็กฎๅฎไธ้๏ผ่ไธๅฏน่บซไฝไนๆฒกๆไปปไฝ็ๅฝฑๅ๏ผๅๆฅๅๆฅ็ไนฐไบ๏ฟฝ ๏ฟฝ๏ฟฝไธชๅจๆใ ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผ่ไธ็ฎ่คไนๆฏไปฅๅๅนฒๅๅ
ๆณฝๅคไบ๏ผ่ไธ่ฟๅพๆๅผนๆง๏ผ่ฟไธช๏ฟฝ ๏ฟฝ๏ฟฝ่่ๅฐ็จ่ตทๆฅ่ฟๆฏๆบๆพๅฟ็๏ผๅฎๅ
จ่ฟๆฏๅพ้่ฆ็๏ผๅพ็ๅฅฝๅฎ ใ ้
่ฏปไบ่ฒๆๅฆไฝๆทกๅ๏ผๅ็่ธไธๅฎนๆ้ฟๆ็ๅๅ ๏ผ ใ่ฒๆๅฝขๆๅๅ ใ ใใๅ
้จๅ ็ด ใใไธใๅๅ ใใๅฝไบบๅๅฐๅๅๆถ๏ผๅฐฑไผๅๆณ่พไธ่
บ็ด ๏ผไธบๅฏนไปๅๅ่ๅ๏ฟฝ๏ฟฝ ๏ฟฝๅคใๅฆๆ้ฟๆๅๅฐๅๅ๏ผไบบไฝๆฐ้ไปฃ่ฐข็ๅนณ่กกๅฐฑไผ้ญๅฐ็ ดๅ๏ฟฝ ๏ฟฝ๏ฟฝ็ฎ่คๆ้็่ฅๅ
ปไพๅบ่ถไบ็ผๆ
ข๏ผ่ฒ็ด ๆฏ็ป่ๅฐฑไผๅๅพๅพๆดป่ท ใ ใใไบใ่ทๅฐ่ๅๆณๅคฑ่ฐ ใใ้ฟๅญ่ฏ้ๆๅซ็ๅฅณๆง่ทๅฐ่้ๆฟ็ด ๏ผไผๅบๆฟ้บฆๆๅฎ็ป่๏ฟฝ๏ฟฝ ๏ฟฝๅๆณ่ๅฝขๆไธๅๅ็ๆ็น๏ผๅ ้ฟๅญ่ฏ่ๅฝขๆ็ๆ็น๏ผ่ฝ็ถๅจ๏ฟฝ ๏ฟฝ๏ฟฝ่ฏไธญๆญๅไผๅๆญข๏ผไฝไปไผๅจ็ฎ่คไธๅ็ๅพ้ฟไธๆฎตๆถ้ดใๆๅญ ไธญๅ ๅฅณๆง่ทๅฐ่้ๆฟ็ด ็ๅขๅ ๏ผ โ ็ฐๆ๏ผ่ฟๆถๅๅบ็ฐ็ๆ็นๅจไบงๅๅคง้จๅไผๆถๅคฑใๅฏๆฏ๏ผๆฐ้๏ฟฝ๏ฟฝ ๏ฟฝ่ฐขไธๆญฃๅธธใ่่ค่ฃธ้ฒๅจๅผบ็็็ดซๅค็บฟไธใ็ฒพ็ฅไธๅๅฐๅๅ็ญ๏ฟฝ ๏ฟฝ๏ฟฝๅ ๏ผ้ฝไผไฝฟๆๅ ๆทฑใๆๆถๆฐ้ฟๅบ็ๆ๏ผไบงๅไนไธไผๆถๅคฑ๏ผๆ ไปฅ้่ฆๆดๅ ๆณจๆใ ใใไธใๆฐ้ไปฃ่ฐข็ผๆ
ข ใใ่็ๆฐ้ไปฃ่ฐขๅ่ฝไธๆญฃๅธธๆๅตๅทขๅ่ฝๅ้ๆถไนไผๅบ็ฐๆ๏ฟฝ๏ฟฝ ๏ฟฝๅ ไธบๆฐ้ไปฃ่ฐขไธ้กบ็
ใๆๅ
ๅๆณๅคฑ่ฐ๏ผไฝฟ่บซไฝๅคไบๆๆ็ถๆ๏ฟฝ ๏ฟฝ๏ฟฝ๏ผไป่ๅ ๅง่ฒ็ด ้ฎ้ขใๆไปฌๅธธ่ฏด็ไพฟ็งไผๅฝขๆๆ๏ผๅ
ถๅฎๅฐฑๆฏ ๅ
ๅๆณๅคฑ่ฐๅฏผ่ด่ฟๆไฝ่ดจ่ๅฝขๆ็ใๅฆๅค๏ผ่บซไฝ็ถๆไธๆญฃๅธธ๏ฟฝ๏ฟฝ ๏ฟฝๆถๅ๏ผ็ดซๅค็บฟ็็
งๅฐไนไผๅ ้ๆ็ๅฝขๆใ ใใๅใ้่ฏฏ็ไฝฟ็จๅๅฆๅ ใใไฝฟ็จไบไธ้ๅ่ชๅทฑ็ฎ่ค็ๅๅฆๅ๏ผไผๅฏผ่ด็ฎ่ค่ฟๆใๅจ๏ฟฝ๏ฟฝ ๏ฟฝ็็่ฟ็จไธญๅฆ่ฟ้็
งๅฐๅฐ็ดซๅค็บฟ๏ผ็ฎ่คไผไธบไบๆตๅพกๅค็็ไพต๏ฟฝ ๏ฟฝ๏ฟฝ๏ผๅจๆ็็็้จไฝ่้้บฆๆๅฎ่ฒ็ด ๏ผ่ฟๆ ทไผๅบ็ฐ่ฒ็ด ๆฒ็็ ้ฎ้ขใ ใใๅค้จๅ ็ด ใใไธใ็ดซๅค็บฟ ใใ็
งๅฐ็ดซๅค็บฟ็ๆถๅ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผไผๅจๅบๅบๅฑไบง๏ฟฝ๏ฟฝ ๏ฟฝๅพๅค้บฆๆๅฎ่ฒ็ด ใๆไปฅไธบไบไฟๆค็ฎ่ค๏ผไผๅจๆๆ้จไฝ่้ๆด๏ฟฝ ๏ฟฝ๏ฟฝ็่ฒ็ด ใ็ปๅธธ่ฃธ้ฒๅจๅผบ็็้ณๅ
ๅบไธไธไป
ไฟ่ฟ็ฎ่ค็่ๅ๏ผ ่ฟไผๅผ่ตท้ปๆใ้ๆ็ญ่ฒ็ด ๆฒ็็็ฎ่ค็พๆฃใ ใใไบใไธ่ฏ็ๆธ
ๆดไน ๆฏ ใใๅ ๅผบ็็ๆธ
ๆดไน ๆฏไฝฟ็ฎ่คๅๅพๆๆ๏ผ่ฟๆ ทไผๅบๆฟ็ฎ่คใ๏ฟฝ๏ฟฝ ๏ฟฝ็ฎ่คๆๆๆถ๏ผไบบไฝไธบไบไฟๆค็ฎ่ค๏ผ้ป่ฒ็ด ็ป่ไผๅๆณๅพๅค้บฆ๏ฟฝ ๏ฟฝ๏ฟฝๅฎ่ฒ็ด ๏ผๅฝ่ฒ็ด ่ฟๅฉๆถๅฐฑๅบ็ฐไบๆใ็็ต็ญ็ฎ่ค่ฒ็ด ๆฒ็็ ้ฎ้ขใ ใใไธใ้ไผ ๅบๅ ใใ็ถๆฏไธญๆ้ฟๆ็๏ผๅๆฌไบบ้ฟๆ็ๆฆ็ๅฐฑๅพ้ซ๏ผ่ฟ็งๆ
ๅต๏ฟฝ๏ฟฝ ๏ฟฝไธๅฎ็จๅบฆไธๅฐฑๅฏๅคๅฎๆฏ้ไผ ๅบๅ ็ไฝ็จใๆไปฅๅฎถ้็นๅซๆฏ้ฟ๏ฟฝ ๏ฟฝ๏ฟฝๆ้ฟๆ็ไบบ๏ผ่ฆๆณจๆ้ฟๅ
ๅผๅ้ฟๆ็้่ฆๅ ็ด ไนไธโโ็ดซๅค ็บฟ็
งๅฐ๏ผ่ฟๆฏ้ข้ฒๆๅฟ
้กปๆณจๆ็ใ ใๆ็้ฎๅธฎไฝ ่งฃๅณใ ใใ ้ป่่ๅฐ็ฒพๅๆถฒ็็ๆๆๆๅ ็็ๅฏไปฅๆ่ธไธ็้ป่ค๏ฟฝ๏ฟฝ ๏ฟฝๅปๆๅ ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒdna็ฒพๅ่ฝๅคๆๆ็ไฟฎๅคๅจๅด้พไปฅ่งฆ๏ฟฝ๏ฟฝ ๏ฟฝ็่ฒๆ๏ผๅ
ถ็ฌๆ็็บณ่ฑๆๅไธบ็ฎ่ค็็พ็ฝไธ้ไธฝ๏ผๆไพไบๅฟ
๏ฟฝ ๏ฟฝ๏ฟฝๅฏๅฐ็่ฅๅ
ป็ฉ่ดจ๏ผๅฏไปฅๆๆ็ๅป้ค้ป่คๆ๏ผ้ป่คๆ๏ผ้ป่คๆ ๏ผ่ด่ถๆ๏ผๆๆใๅฆๅจ ๆ็ญใๅฎๅฎๅฎๅ
จ็ช็ ดไบไผ ็ป็็พ่คๆถ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผๅฎๅฆๅจ็ฎ่คไธญๆณจๅ
ฅไบไธๆฏๅ
ผๅ
ทๆดปๅใๅ็ใๆปๅ
ป็ญๅๆ็๏ฟฝ ๏ฟฝ๏ฟฝๅฐพ้
๏ผๅๆถไธบ่ธ้จๆไพๅคง้ๆๆบ็ปด็็ด ็ฒพๅ๏ผ่ธ้จ็ๆนๅๆพ ่ๆ่งใ่ชไบงๅไธๅธไปฅๆฅ๏ผ่้กพๅฎข็บท็บทไป็ปๆฐ้กพๅฎข๏ผ ็ๆฐ๏ฟฝ๏ฟฝ ๏ฟฝๅฎข้ฝๆฏ้่ฟ่้กพๅฎขไป็ป่ๆฅ๏ผๅฃ็ข็ฑๆญค่ๆฅ ใใ ๏ผๆ็จ้ป่่ๅฐ็พ็ฝ๏ผไผไผค่บซไฝๅ ๆๅฏไฝ็จๅ ใใ็ญ๏ผ้ป่่ๅฐ็ฒพๅๆถฒๅบ็จไบ็ฒพ็บฏๅคๅ้
ๆนๅ้ขๅ
็ๅ็ฑป๏ฟฝ๏ฟฝ ๏ฟฝๆ็งๆ๏ผๅนถๅฐโdna็พ่ค็ณป็ปโ็ๆณๅบ็จๅฐไบ่ฏฅไบงๅไธญ๏ผ่ฝๅฝป๏ฟฝ ๏ฟฝ๏ฟฝ็ฅ้ค้ป่คๆ๏ผ่ด่ถๆ๏ผๅฆๅจ ๆ๏ผๆๆ๏ผ้ป่คๆ๏ผ่ๅนดๆ๏ผๆ ๆๆทกๅ้ป่คๆ่ณๆฅ่ฟ่ค่ฒใ้ป่่ๅฐ้่ฟๆณๅฝใ็พๅฝใๅฐๆนพ๏ฟฝ๏ฟฝ ๏ฟฝๅฐ็ไธๅฎถ้ๅๅไฝ๏ผ ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผๆๆไผ ็ปๅๅญฆๆค่ค็ๅฟต๏ผไธๆ่ฟฝๅฏปๅ็ฐ็ ด่ฏๅคง่ช็ถ็็พไธฝ๏ฟฝ ๏ฟฝ๏ฟฝ่ฟน๏ผไปคๆฏไธไฝ็ฑ็พ็ๅฅณๆง้ฝ่ฝไบซๅๅฐ็งๆๅๆฐๆๅธฆๆฅ็่ช็ถ ไน็พใ ไธไธบไบๆดฒๅฅณๆง่ค่ดจ็ ๅถ๏ผ็ฒพๅฟๅตๆคๅฅณๆง็พไธฝ๏ผๅคๅนดๆฅ๏ผไธบๆฐ๏ฟฝ๏ฟฝ ๏ฟฝ็พไธ่ฎก็ๅฅณๆง่งฃ้คไบ้ป่คๆๅฐๆฐใๆทฑๅพๅนฟๅคงๅฅณๆงๆๅ็ไฟก่ต ใใ ๏ผๅป้ค้ป่คๆไนๅ๏ผไผๅๅผนๅ ใใ็ญ๏ผๅพๅคๆพ็ป้ฟไบ้ป่คๆ็ไบบๅฃซ๏ผ่ชไป้ๆฉไบ้ป่่ๅฐ๏ฟฝ๏ฟฝ ๏ฟฝ็ฝ๏ผๅฐฑไธๅณๆฐธ้ธใ่ฟๆฌพ็ฅๆไบงๅๆฏ็ป่ฟๆฐๅไฝๆๅจ็ฅๆไธๅฎถ๏ฟฝ ๏ฟฝ๏ฟฝๆฎๆ็ๅฝขๆๅๅ ็ฒพๅฟ็ ๅถ่ๆ็จไบๅฎ่ฏด่ฏ๏ผ่ฎฉๆถ่ดน่
ๆๅใ ๆ ็ซๆๅจๅ็ ๆไปฌ็ๅพๅคๆฐๅฎขๆท้ฝๆฏ่ๅฎขๆทไป็ป่ๆฅ๏ผ่ฏท้ฎ๏ฟฝ ๏ฟฝ๏ฟฝๅฆๆๆๆไธๅฅฝ๏ผไผๆๅฎขๆท่ฝฌไป็ปๅ ใใ ๏ผไฝ ไปฌ็ไปทๆ ผๆ็น่ดต๏ผ่ฝไธ่ฝไพฟๅฎไธ็น ใใ็ญ๏ผ ๏ผ ๏ผ ๏ผ่่ฟไบๆฏซๆ ็้ฎ๏ผไธไผๅฏนๅฝปๅบๅป๏ฟฝ ๏ฟฝ๏ฟฝไฝ ็ๆ็นๆไปปไฝๅธฎๅฉ ไธๅไปท้ฑ๏ผไธไปฝไปทๅผ๏ผๆไปฌ็ฐๅจๅ็๏ฟฝ๏ฟฝ ๏ฟฝๆฏไธไธชๅฃ็ข๏ผไธไธชๅ็๏ผไปท้ฑๅนถไธ้ซใๅฆๆ่ฑ่ฟ็น้ฑๆไฝ ็๏ฟฝ ๏ฟฝ๏ฟฝ่คๆๅฝปๅบๅป้ค๏ผไฝ ่ฟไผ่งๅพ่ดตๅ ไฝ ่ฟไผๅๅป่ฑ้ฃไนๅคๅคๆ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผไธไฝๆๆฒกๅปๆ๏ผ่ฟๆ่ชๅทฑ็็ฎ่คๅผ็่ถๆฅ่ถ็ณๅ ใใ ๏ผๆ้ๅ็จ้ป่่ๅฐ็ฒพๅๆถฒๅ ใใ็ญ๏ผ้ป่่ๅฐ้็จไบบ็พค๏ผ ใใ ใ็็็ดไนฑๅผ่ตท็้ป่คๆไบบ็พค ใใ ใ็่ฒๅผ่ตท็ๅฆๅจ ๆไบบ็พค ใใ ใๅนด็บชๅข้ฟๅผ่ตท็่ๅนดๆไบบ็พค ใใ ใๅๅฆๅ่ฒ็ด ๆฒ็งฏใ่พๅฐๆไบบ็พค ใใ ใ้ฟๆๆฅ็
งๅผ่ตท็ๆฅๆๆไบบ็พค ใใ ใ่่คๆๆทกๆฅ้็พ็ฝ็ไบบ็พค ใ็ฅๆๅฐๆนๆณใ ่ฒๆๅฆไฝๆทกๅ๏ผๅๆถไธบๆจๅไบซ็ฅๆๅฐๆนๆณ ๅฐไบบๅ่็ๆตใ็ปด็็ด cใeๆณจๅฐๆถฒๅๅๅฐๆถๅจๆดๅ็้ข้จ๏ผ๏ฟฝ ๏ฟฝ๏ฟฝ ๏ผ่ฟ
้ๆท็้ข้จ๏ผๅ่็ฒไผ ่ช่กๅ็กฌใๅ็ญใๅทๅด๏ผ ๏ผๅฏๆๆๅ้้ข้จๆใ ใ ใ ใ ๏ผๆททๅ็ ๆ็ปๆซ๏ผ้ฅญๅ็จ็ฑณๆฑค๏ฟฝ๏ฟฝ ๏ฟฝๆ๏ผไธๆฅไธๆฌก๏ผ่ฟๆๆฐๆ๏ผ้ข้จไผๅๅพ็ฝๅซฉ่ๅ
ๆปใ original issue reported on code google com by additive gmail com on jul at | 1 |
62,317 | 17,023,895,874 | IssuesEvent | 2021-07-03 04:25:34 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | 'France' relationship probably broken | Component: nominatim Priority: major Resolution: wontfix Type: defect | **[Submitted to the original trac issue database at 4.50pm, Tuesday, 4th February 2014]**
Our nominatim testcases in geocode-glib that search for "bonneville" with "fr_FR" locale are currently broken because Nominatim is not returning "country" in address object but only "country_code". The test case also breaks on the description of the place as we rely on country to construct a good description for place. | 1.0 | 'France' relationship probably broken - **[Submitted to the original trac issue database at 4.50pm, Tuesday, 4th February 2014]**
Our nominatim testcases in geocode-glib that search for "bonneville" with "fr_FR" locale are currently broken because Nominatim is not returning "country" in address object but only "country_code". The test case also breaks on the description of the place as we rely on country to construct a good description for place. | defect | france relationship probably broken our nominatim testcases in geocode glib that search for bonneville with fr fr locale are currently broken because nominatim is not returning country in address object but only country code the test case also breaks on the description of the place as we rely on country to construct a good description for place | 1 |
42,641 | 11,194,794,867 | IssuesEvent | 2020-01-03 02:58:42 | libarchive/libarchive | https://api.github.com/repos/libarchive/libarchive | closed | test_open_failure leaks memory | Type-Defect | test_open_failure leaks memory:
```
Direct leak of 32 byte(s) in 1 object(s) allocated from:
#0 0x7fdb48073ce6 in calloc (/usr/lib/x86_64-linux-gnu/libasan.so.5+0x10dce6)
#1 0x55f22bacbb84 in archive_write_client_open libarchive/archive_write.c:348
#2 0x55f22bacb5f4 in __archive_write_open_filter libarchive/archive_write.c:266
#3 0x55f22bacb4da in __archive_write_open_filter libarchive/archive_write.c:257
#4 0x55f22bacb6da in __archive_write_filters_open libarchive/archive_write.c:280
#5 0x55f22baccb56 in archive_write_open libarchive/archive_write.c:527
#6 0x55f22bbc5e86 in test_open_failure libarchive/test/test_open_failure.c:176
#7 0x55f22bb7028a in test_run test_utils/test_main.c:3548
#8 0x55f22bb71efe in main test_utils/test_main.c:4049
#9 0x7fdb47af11e2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x271e2)
Indirect leak of 10240 byte(s) in 1 object(s) allocated from:
#0 0x7fdb48073ae8 in malloc (/usr/lib/x86_64-linux-gnu/libasan.so.5+0x10dae8)
#1 0x55f22bacbb94 in archive_write_client_open libarchive/archive_write.c:349
#2 0x55f22bacb5f4 in __archive_write_open_filter libarchive/archive_write.c:266
#3 0x55f22bacb4da in __archive_write_open_filter libarchive/archive_write.c:257
#4 0x55f22bacb6da in __archive_write_filters_open libarchive/archive_write.c:280
#5 0x55f22baccb56 in archive_write_open libarchive/archive_write.c:527
#6 0x55f22bbc5e86 in test_open_failure libarchive/test/test_open_failure.c:176
#7 0x55f22bb7028a in test_run test_utils/test_main.c:3548
#8 0x55f22bb71efe in main test_utils/test_main.c:4049
#9 0x7fdb47af11e2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x271e2)
```
| 1.0 | test_open_failure leaks memory - test_open_failure leaks memory:
```
Direct leak of 32 byte(s) in 1 object(s) allocated from:
#0 0x7fdb48073ce6 in calloc (/usr/lib/x86_64-linux-gnu/libasan.so.5+0x10dce6)
#1 0x55f22bacbb84 in archive_write_client_open libarchive/archive_write.c:348
#2 0x55f22bacb5f4 in __archive_write_open_filter libarchive/archive_write.c:266
#3 0x55f22bacb4da in __archive_write_open_filter libarchive/archive_write.c:257
#4 0x55f22bacb6da in __archive_write_filters_open libarchive/archive_write.c:280
#5 0x55f22baccb56 in archive_write_open libarchive/archive_write.c:527
#6 0x55f22bbc5e86 in test_open_failure libarchive/test/test_open_failure.c:176
#7 0x55f22bb7028a in test_run test_utils/test_main.c:3548
#8 0x55f22bb71efe in main test_utils/test_main.c:4049
#9 0x7fdb47af11e2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x271e2)
Indirect leak of 10240 byte(s) in 1 object(s) allocated from:
#0 0x7fdb48073ae8 in malloc (/usr/lib/x86_64-linux-gnu/libasan.so.5+0x10dae8)
#1 0x55f22bacbb94 in archive_write_client_open libarchive/archive_write.c:349
#2 0x55f22bacb5f4 in __archive_write_open_filter libarchive/archive_write.c:266
#3 0x55f22bacb4da in __archive_write_open_filter libarchive/archive_write.c:257
#4 0x55f22bacb6da in __archive_write_filters_open libarchive/archive_write.c:280
#5 0x55f22baccb56 in archive_write_open libarchive/archive_write.c:527
#6 0x55f22bbc5e86 in test_open_failure libarchive/test/test_open_failure.c:176
#7 0x55f22bb7028a in test_run test_utils/test_main.c:3548
#8 0x55f22bb71efe in main test_utils/test_main.c:4049
#9 0x7fdb47af11e2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x271e2)
```
| defect | test open failure leaks memory test open failure leaks memory direct leak of byte s in object s allocated from in calloc usr lib linux gnu libasan so in archive write client open libarchive archive write c in archive write open filter libarchive archive write c in archive write open filter libarchive archive write c in archive write filters open libarchive archive write c in archive write open libarchive archive write c in test open failure libarchive test test open failure c in test run test utils test main c in main test utils test main c in libc start main lib linux gnu libc so indirect leak of byte s in object s allocated from in malloc usr lib linux gnu libasan so in archive write client open libarchive archive write c in archive write open filter libarchive archive write c in archive write open filter libarchive archive write c in archive write filters open libarchive archive write c in archive write open libarchive archive write c in test open failure libarchive test test open failure c in test run test utils test main c in main test utils test main c in libc start main lib linux gnu libc so | 1 |
21,921 | 14,935,183,457 | IssuesEvent | 2021-01-25 11:34:29 | airyhq/airy | https://api.github.com/repos/airyhq/airy | opened | AKHQ deployment | infrastructure | AKHQ or formerly known KafkaHQ is a useful tool for Kafka debugging that we would like to add as an optional core component | 1.0 | AKHQ deployment - AKHQ or formerly known KafkaHQ is a useful tool for Kafka debugging that we would like to add as an optional core component | non_defect | akhq deployment akhq or formerly known kafkahq is a useful tool for kafka debugging that we would like to add as an optional core component | 0 |
116,528 | 14,973,162,653 | IssuesEvent | 2021-01-28 00:26:03 | urbit/landscape | https://api.github.com/repos/urbit/landscape | closed | Clickable tags should be <button>s or <a>s | design | **Describe and screenshoot the design issue**
Certain Landscape buttons/clickables (e.g. the "Join a Group" button) aren't actually clickable tags, but are instead <div>s, <span>s, <svg>s, or other non-clickables; I assume that's for CSS reasons. The issue with that is that since only the JavaScript knows what's clickable and what isn't, vimium users like myself, or people who don't own a mouse and use qutebrowser or navigate using the Tab button, won't be able to select it. For vimium users, the error looks like this:

"Clickable" elements without that little boxed letter symbol can't be selected using vimium, Tab, or qutebrowser.
Looking thru the unselectable elements, they look to be mostly <div>s:

**To Reproduce**
Steps to reproduce/view the behavior:
1. Open a localhost Landscape instance
2. Attempt to select each clickable element using the Tab button
3. Observe which elements are and are not clickable
**Expected design behavior**
All clickable elements should be navigable without a mouse.
**Desktop (please complete the following information):**
- OS: [Artix Linux, X11]
- Browser [Brave using Vimium]
- Base hash of your urbit ship. 74m9v | 1.0 | Clickable tags should be <button>s or <a>s - **Describe and screenshoot the design issue**
Certain Landscape buttons/clickables (e.g. the "Join a Group" button) aren't actually clickable tags, but are instead <div>s, <span>s, <svg>s, or other non-clickables; I assume that's for CSS reasons. The issue with that is that since only the JavaScript knows what's clickable and what isn't, vimium users like myself, or people who don't own a mouse and use qutebrowser or navigate using the Tab button, won't be able to select it. For vimium users, the error looks like this:

"Clickable" elements without that little boxed letter symbol can't be selected using vimium, Tab, or qutebrowser.
Looking thru the unselectable elements, they look to be mostly <div>s:

**To Reproduce**
Steps to reproduce/view the behavior:
1. Open a localhost Landscape instance
2. Attempt to select each clickable element using the Tab button
3. Observe which elements are and are not clickable
**Expected design behavior**
All clickable elements should be navigable without a mouse.
**Desktop (please complete the following information):**
- OS: [Artix Linux, X11]
- Browser [Brave using Vimium]
- Base hash of your urbit ship. 74m9v | non_defect | clickable tags should be s or s describe and screenshoot the design issue certain landscape buttons clickables e g the join a group button aren t actually clickable tags but are instead s s s or other non clickables i assume that s for css reasons the issue with that is that since only the javascript knows what s clickable and what isn t vimium users like myself or people who don t own a mouse and use qutebrowser or navigate using the tab button won t be able to select it for vimium users the error looks like this clickable elements without that little boxed letter symbol can t be selected using vimium tab or qutebrowser looking thru the unselectable elements they look to be mostly s to reproduce steps to reproduce view the behavior open a localhost landscape instance attempt to select each clickable element using the tab button observe which elements are and are not clickable expected design behavior all clickable elements should be navigable without a mouse desktop please complete the following information os browser base hash of your urbit ship | 0 |
66,758 | 14,798,951,582 | IssuesEvent | 2021-01-13 01:03:53 | jtimberlake/griffin | https://api.github.com/repos/jtimberlake/griffin | opened | CVE-2020-36181 (High) detected in jackson-databind-2.8.7.jar, jackson-databind-2.9.9.3.jar | security vulnerability | ## CVE-2020-36181 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.7.jar</b>, <b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: griffin/measure/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.7/jackson-databind-2.8.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: griffin/service/hibernate_mysql_pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181>CVE-2020-36181</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"vulnerabilityIdentifier":"CVE-2020-36181","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36181 (High) detected in jackson-databind-2.8.7.jar, jackson-databind-2.9.9.3.jar - ## CVE-2020-36181 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.7.jar</b>, <b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: griffin/measure/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.7/jackson-databind-2.8.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: griffin/service/hibernate_mysql_pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181>CVE-2020-36181</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"vulnerabilityIdentifier":"CVE-2020-36181","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file griffin measure pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file griffin service hibernate mysql pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp cpdsadapter driveradaptercpds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp cpdsadapter driveradaptercpds vulnerabilityurl | 0 |
42,216 | 10,887,311,233 | IssuesEvent | 2019-11-18 14:19:30 | idaholab/raven | https://api.github.com/repos/idaholab/raven | opened | [DEFECT] Input reader doesn't strip Out whitespace when reading array | defect priority_normal | --------
Defect Description
--------
**Describe the defect**
I was utilizing an array of RELAP5 input card numbers for a sampler, so the same sampling was performed for multiple locations in my RELAP5
[Spaces_In_Sampler_Array.zip](https://github.com/idaholab/raven/files/3858978/Spaces_In_Sampler_Array.zip)
input and when i put a space in the input array it caused RAVEN to abort.
<variable name="7290102:1, 7290103:1">
##### What did you see instead?
RAVEN aborted with a non-descriptive error message:
Traceback (most recent call last):
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 201, in addRealization
rlz = dict((var,rlz[var]) for var in self.getVars()+self.indexes)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 201, in <genexpr>
rlz = dict((var,rlz[var]) for var in self.getVars()+self.indexes)
KeyError: 'ProbabilityWeight- 7290103:1'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Driver.py", line 285, in <module>
raven()
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Driver.py", line 238, in raven
simulation.run()
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Simulation.py", line 780, in run
stepInstance.takeAstep(stepInputDict)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 315, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 640, in _localTakeAstepRun
myLambda([finishedJob,outputs[outIndex]])
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 588, in <lambda>
self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) )
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Models/Code.py", line 747, in collectOutput
output.addRealization(evaluation)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 204, in addRealization
self.raiseAnError(KeyError,'Provided realization does not have all requisite values for object "{}": "{}"'.format(self.name,e.args[0]))
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/MessageHandler.py", line 99, in raiseAnError
self.messageHandler.error(self,etype,msg,str(tag),verbosity,color)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/MessageHandler.py", line 313, in error
raise etype(message)
KeyError: 'Provided realization does not have all requisite values for object "RELAPoutput": "ProbabilityWeight- 7290103:1"'
/home/cgosdin/RAVEN_FPS/RAVEN_FPS/raven_framework: line 56: 49370 Segmentation fault (core dumped) $PYTHON_COMMAND -O $RAVEN_DIR/framework/Driver.py "${ARGS[@]}"
##### Do you have a suggested fix for the development team?
Input reader should strip whitespace before reading.
**Describe how to Reproduce**
Steps to reproduce the behavior:
1. Unzip file.
2
[raven.log](https://github.com/idaholab/raven/files/3858985/raven.log)
. Rerun RAVEN.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: LSB Version: n/a
Distributor ID: SUSE
Description: SUSE Linux Enterprise Server 12 SP4
Release: 12.4
Codename: n/a
- Dependencies Installation: CONDA
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| 1.0 | [DEFECT] Input reader doesn't strip Out whitespace when reading array - --------
Defect Description
--------
**Describe the defect**
I was utilizing an array of RELAP5 input card numbers for a sampler, so the same sampling was performed for multiple locations in my RELAP5
[Spaces_In_Sampler_Array.zip](https://github.com/idaholab/raven/files/3858978/Spaces_In_Sampler_Array.zip)
input and when i put a space in the input array it caused RAVEN to abort.
<variable name="7290102:1, 7290103:1">
##### What did you see instead?
RAVEN aborted with a non-descriptive error message:
Traceback (most recent call last):
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 201, in addRealization
rlz = dict((var,rlz[var]) for var in self.getVars()+self.indexes)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 201, in <genexpr>
rlz = dict((var,rlz[var]) for var in self.getVars()+self.indexes)
KeyError: 'ProbabilityWeight- 7290103:1'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Driver.py", line 285, in <module>
raven()
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Driver.py", line 238, in raven
simulation.run()
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Simulation.py", line 780, in run
stepInstance.takeAstep(stepInputDict)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 315, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 640, in _localTakeAstepRun
myLambda([finishedJob,outputs[outIndex]])
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Steps.py", line 588, in <lambda>
self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) )
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/Models/Code.py", line 747, in collectOutput
output.addRealization(evaluation)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/DataObjects/DataSet.py", line 204, in addRealization
self.raiseAnError(KeyError,'Provided realization does not have all requisite values for object "{}": "{}"'.format(self.name,e.args[0]))
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/MessageHandler.py", line 99, in raiseAnError
self.messageHandler.error(self,etype,msg,str(tag),verbosity,color)
File "/home/cgosdin/RAVEN_FPS/RAVEN_FPS/framework/MessageHandler.py", line 313, in error
raise etype(message)
KeyError: 'Provided realization does not have all requisite values for object "RELAPoutput": "ProbabilityWeight- 7290103:1"'
/home/cgosdin/RAVEN_FPS/RAVEN_FPS/raven_framework: line 56: 49370 Segmentation fault (core dumped) $PYTHON_COMMAND -O $RAVEN_DIR/framework/Driver.py "${ARGS[@]}"
##### Do you have a suggested fix for the development team?
Input reader should strip whitespace before reading.
**Describe how to Reproduce**
Steps to reproduce the behavior:
1. Unzip file.
2
[raven.log](https://github.com/idaholab/raven/files/3858985/raven.log)
. Rerun RAVEN.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: LSB Version: n/a
Distributor ID: SUSE
Description: SUSE Linux Enterprise Server 12 SP4
Release: 12.4
Codename: n/a
- Dependencies Installation: CONDA
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| defect | input reader doesn t strip out whitespace when reading array defect description describe the defect i was utilizing an array of input card numbers for a sampler so the same sampling was performed for multiple locations in my input and when i put a space in the input array it caused raven to abort what did you see instead raven aborted with a non descriptive error message traceback most recent call last file home cgosdin raven fps raven fps framework dataobjects dataset py line in addrealization rlz dict var rlz for var in self getvars self indexes file home cgosdin raven fps raven fps framework dataobjects dataset py line in rlz dict var rlz for var in self getvars self indexes keyerror probabilityweight during handling of the above exception another exception occurred traceback most recent call last file home cgosdin raven fps raven fps framework driver py line in raven file home cgosdin raven fps raven fps framework driver py line in raven simulation run file home cgosdin raven fps raven fps framework simulation py line in run stepinstance takeastep stepinputdict file home cgosdin raven fps raven fps framework steps py line in takeastep self localtakeasteprun indictionary file home cgosdin raven fps raven fps framework steps py line in localtakeasteprun mylambda file home cgosdin raven fps raven fps framework steps py line in self outputcollectionlambda append lambda x indictionary collectoutput x x outindex file home cgosdin raven fps raven fps framework models code py line in collectoutput output addrealization evaluation file home cgosdin raven fps raven fps framework dataobjects dataset py line in addrealization self raiseanerror keyerror provided realization does not have all requisite values for object format self name e args file home cgosdin raven fps raven fps framework messagehandler py line in raiseanerror self messagehandler error self etype msg str tag verbosity color file home cgosdin raven fps raven fps framework messagehandler py line in error raise etype message keyerror provided realization does not have all requisite values for object relapoutput probabilityweight home cgosdin raven fps raven fps raven framework line segmentation fault core dumped python command o raven dir framework driver py args do you have a suggested fix for the development team input reader should strip whitespace before reading describe how to reproduce steps to reproduce the behavior unzip file rerun raven screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os lsb version n a distributor id suse description suse linux enterprise server release codename n a dependencies installation conda for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided | 1 |
21,479 | 3,512,018,511 | IssuesEvent | 2016-01-10 18:34:41 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | A small amount of CPU time is wasted on Operator.name().toLowerCase() | C: Functionality P: Medium R: Fixed T: Defect | In order to transform `AND` and `OR` keywords to lower case, the `Operator.name()` is converted to lower case every time these keywords are rendered. This is rather wasteful and can be omitted by providing constants.
A JMC profiling session on a benchmark with 1M times the same query being executed shows that `toLowerCase()` accounts for 1% | 1.0 | A small amount of CPU time is wasted on Operator.name().toLowerCase() - In order to transform `AND` and `OR` keywords to lower case, the `Operator.name()` is converted to lower case every time these keywords are rendered. This is rather wasteful and can be omitted by providing constants.
A JMC profiling session on a benchmark with 1M times the same query being executed shows that `toLowerCase()` accounts for 1% | defect | a small amount of cpu time is wasted on operator name tolowercase in order to transform and and or keywords to lower case the operator name is converted to lower case every time these keywords are rendered this is rather wasteful and can be omitted by providing constants a jmc profiling session on a benchmark with times the same query being executed shows that tolowercase accounts for | 1 |
33,253 | 12,196,707,623 | IssuesEvent | 2020-04-29 19:28:22 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | opened | May get an NPE in URLEncoder.encode when OAuth provder gets bad clientId TS003459997 | bug release bug team:Security SSO | **Describe the bug**
In oauth, it appears that, when making an authorization request, when a clientId does not exist, you may get an NEP from URLEncode. Here is the stack trace:
2020-03-19T12:27:41.904+0000] 00000af8 id=d9aa0b80 .ibm.ws.webcontainer.security.WebAppSecurityCollaboratorImpl < resetSyncToOSThread Exit
[2020-03-19T12:27:41.896+0000] 00000ad4 id=00000000 com.ibm.ws.webcontainer.util.ApplicationErrorUtils E SRVE0777E: Exception thrown by application class 'java.net.URLEncoder.encode:225'
java.lang.NullPointerException
at java.net.URLEncoder.encode(URLEncoder.java:225)
at com.ibm.ws.security.oauth20.error.impl.OAuth20AuthorizeRequestExceptionHandler.handleResultException(OAuth20AuthorizeRequestExceptionHandler.java:251)
at com.ibm.ws.security.oauth20.web.UserAuthentication.renderErrorPage(UserAuthentication.java:487)
at com.ibm.ws.security.oauth20.web.OAuth20EndpointServices.handleEndpointRequest(OAuth20EndpointServices.java:252)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServices.handleOidcRequest(OidcEndpointServices.java:274)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServlet.doPost(OidcEndpointServlet.java:55)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServlet.doGet(OidcEndpointServlet.java:49)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:686)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:791)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.service(ServletWrapper.java:1230)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:729)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:426)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.invokeTarget(WebAppFilterChain.java:182)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:93)
at com.ibm.ws.security.openidconnect.web.OidcRequestFilter.setEndpointRequest(OidcRequestFilter.java:41)
at com.ibm.ws.security.oauth20.web.OAuth20RequestFilter.doFilter(OAuth20RequestFilter.java:93)
at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:201)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:90)
at com.ibm.ws.app.manager.wab.internal.OsgiDirectoryProtectionFilter.doFilter(OsgiDirectoryProtectionFilter.java:90)
at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:201)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:90)
at com.ibm.ws.webcontainer.filter.WebAppFilterManager.doFilter(WebAppFilterManager.java:1001)
...
com.ibm.ws.security.oauth20.error.impl.OAuth20AuthorizeRequestExceptionHandler.handleResultException(**OAuth20AuthorizeRequestExceptionHandler.java:251)**
is this in the line in the try block
```
String errorDesc = e2.formatSelf(req.getLocale(), encoding);
try {
errorDesc = URLEncoder.encode(errorDesc, "utf-8");
} catch (UnsupportedEncodingException e1) {
if (tc.isDebugEnabled())
Tr.debug(tc, "Internal error encoding error description", new Object[] { e1 });
}
```
I suggest two things here:
1. Don't run the URLEncoder.encode if errorDesc is null.
2. Change UnsupportedEncodingException to Excepton.
**Expected behavior**
Don't get an NPE in URLEncode
**Diagnostic information:**
- OpenLiberty Version: [e.g. 19.0.0.12]
- Java Version: [i.e. full output of `java -version`]
Java(TM) SE Runtime Environment (8.0.6.5 - pxa6480sr6fp5ifix-20200304_01(SR6 FP5+IJ23018
| True | May get an NPE in URLEncoder.encode when OAuth provder gets bad clientId TS003459997 - **Describe the bug**
In oauth, it appears that, when making an authorization request, when a clientId does not exist, you may get an NEP from URLEncode. Here is the stack trace:
2020-03-19T12:27:41.904+0000] 00000af8 id=d9aa0b80 .ibm.ws.webcontainer.security.WebAppSecurityCollaboratorImpl < resetSyncToOSThread Exit
[2020-03-19T12:27:41.896+0000] 00000ad4 id=00000000 com.ibm.ws.webcontainer.util.ApplicationErrorUtils E SRVE0777E: Exception thrown by application class 'java.net.URLEncoder.encode:225'
java.lang.NullPointerException
at java.net.URLEncoder.encode(URLEncoder.java:225)
at com.ibm.ws.security.oauth20.error.impl.OAuth20AuthorizeRequestExceptionHandler.handleResultException(OAuth20AuthorizeRequestExceptionHandler.java:251)
at com.ibm.ws.security.oauth20.web.UserAuthentication.renderErrorPage(UserAuthentication.java:487)
at com.ibm.ws.security.oauth20.web.OAuth20EndpointServices.handleEndpointRequest(OAuth20EndpointServices.java:252)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServices.handleOidcRequest(OidcEndpointServices.java:274)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServlet.doPost(OidcEndpointServlet.java:55)
at com.ibm.ws.security.openidconnect.web.OidcEndpointServlet.doGet(OidcEndpointServlet.java:49)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:686)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:791)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.service(ServletWrapper.java:1230)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:729)
at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:426)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.invokeTarget(WebAppFilterChain.java:182)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:93)
at com.ibm.ws.security.openidconnect.web.OidcRequestFilter.setEndpointRequest(OidcRequestFilter.java:41)
at com.ibm.ws.security.oauth20.web.OAuth20RequestFilter.doFilter(OAuth20RequestFilter.java:93)
at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:201)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:90)
at com.ibm.ws.app.manager.wab.internal.OsgiDirectoryProtectionFilter.doFilter(OsgiDirectoryProtectionFilter.java:90)
at com.ibm.ws.webcontainer.filter.FilterInstanceWrapper.doFilter(FilterInstanceWrapper.java:201)
at com.ibm.ws.webcontainer.filter.WebAppFilterChain.doFilter(WebAppFilterChain.java:90)
at com.ibm.ws.webcontainer.filter.WebAppFilterManager.doFilter(WebAppFilterManager.java:1001)
...
com.ibm.ws.security.oauth20.error.impl.OAuth20AuthorizeRequestExceptionHandler.handleResultException(**OAuth20AuthorizeRequestExceptionHandler.java:251)**
is this in the line in the try block
```
String errorDesc = e2.formatSelf(req.getLocale(), encoding);
try {
errorDesc = URLEncoder.encode(errorDesc, "utf-8");
} catch (UnsupportedEncodingException e1) {
if (tc.isDebugEnabled())
Tr.debug(tc, "Internal error encoding error description", new Object[] { e1 });
}
```
I suggest two things here:
1. Don't run the URLEncoder.encode if errorDesc is null.
2. Change UnsupportedEncodingException to Excepton.
**Expected behavior**
Don't get an NPE in URLEncode
**Diagnostic information:**
- OpenLiberty Version: [e.g. 19.0.0.12]
- Java Version: [i.e. full output of `java -version`]
Java(TM) SE Runtime Environment (8.0.6.5 - pxa6480sr6fp5ifix-20200304_01(SR6 FP5+IJ23018
| non_defect | may get an npe in urlencoder encode when oauth provder gets bad clientid describe the bug in oauth it appears that when making an authorization request when a clientid does not exist you may get an nep from urlencode here is the stack trace id ibm ws webcontainer security webappsecuritycollaboratorimpl resetsynctoosthread exit id com ibm ws webcontainer util applicationerrorutils e exception thrown by application class java net urlencoder encode java lang nullpointerexception at java net urlencoder encode urlencoder java at com ibm ws security error impl handleresultexception java at com ibm ws security web userauthentication rendererrorpage userauthentication java at com ibm ws security web handleendpointrequest java at com ibm ws security openidconnect web oidcendpointservices handleoidcrequest oidcendpointservices java at com ibm ws security openidconnect web oidcendpointservlet dopost oidcendpointservlet java at com ibm ws security openidconnect web oidcendpointservlet doget oidcendpointservlet java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at com ibm ws webcontainer servlet servletwrapper service servletwrapper java at com ibm ws webcontainer servlet servletwrapper handlerequest servletwrapper java at com ibm ws webcontainer servlet servletwrapper handlerequest servletwrapper java at com ibm ws webcontainer filter webappfilterchain invoketarget webappfilterchain java at com ibm ws webcontainer filter webappfilterchain dofilter webappfilterchain java at com ibm ws security openidconnect web oidcrequestfilter setendpointrequest oidcrequestfilter java at com ibm ws security web dofilter java at com ibm ws webcontainer filter filterinstancewrapper dofilter filterinstancewrapper java at com ibm ws webcontainer filter webappfilterchain dofilter webappfilterchain java at com ibm ws app manager wab internal osgidirectoryprotectionfilter dofilter osgidirectoryprotectionfilter java at com ibm ws webcontainer filter filterinstancewrapper dofilter filterinstancewrapper java at com ibm ws webcontainer filter webappfilterchain dofilter webappfilterchain java at com ibm ws webcontainer filter webappfiltermanager dofilter webappfiltermanager java com ibm ws security error impl handleresultexception java is this in the line in the try block string errordesc formatself req getlocale encoding try errordesc urlencoder encode errordesc utf catch unsupportedencodingexception if tc isdebugenabled tr debug tc internal error encoding error description new object i suggest two things here don t run the urlencoder encode if errordesc is null change unsupportedencodingexception to excepton expected behavior don t get an npe in urlencode diagnostic information openliberty version java version java tm se runtime environment | 0 |
18,588 | 3,073,881,660 | IssuesEvent | 2015-08-20 01:25:17 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | opened | VM service: Class.error documentation is incorrect | Area-Observatory Type-Defect | In the [VM service documentation](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#class), the `Class.error` field is documented to have type `@Instance`. However, in practice this field returns a value of type `@Error`. `@Error` is probably correct, since it includes extra metadata like the error kind and stack trace. | 1.0 | VM service: Class.error documentation is incorrect - In the [VM service documentation](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#class), the `Class.error` field is documented to have type `@Instance`. However, in practice this field returns a value of type `@Error`. `@Error` is probably correct, since it includes extra metadata like the error kind and stack trace. | defect | vm service class error documentation is incorrect in the the class error field is documented to have type instance however in practice this field returns a value of type error error is probably correct since it includes extra metadata like the error kind and stack trace | 1 |
57,156 | 15,712,047,146 | IssuesEvent | 2021-03-27 10:19:02 | google/mozc | https://api.github.com/repos/google/mozc | closed | hand writing canvas size | Priority-Medium Type-Defect auto-migrated obsolete | ```
What steps will reproduce the problem?
1. Start hand writing mode: $ /usr/lib/mozc/mozc_tool --mode=hand_writing
What is the expected output?
Being able to resize the drawing canvas manually (handles) or via settings
What do you see instead?
A very small canvas, not resizable, that makes it very difficult to use with a
computer touch screen (drawing kanji with a finger)
What version of the product are you using? On what operating system?
Arch linux, mozc-im (ibus-mozc, mozc, uim-mozc) v1.15.1857.102-1
Please provide any additional information below.
I would like to be able to resize the drawing canvas as I wish (or have it
resize automatically depending on the window size), so I can replace
tegaki-recognize with mozc_tool hand_writing
```
Original issue reported on code.google.com by `lemaire.adrien` on 12 Aug 2014 at 11:20
| 1.0 | hand writing canvas size - ```
What steps will reproduce the problem?
1. Start hand writing mode: $ /usr/lib/mozc/mozc_tool --mode=hand_writing
What is the expected output?
Being able to resize the drawing canvas manually (handles) or via settings
What do you see instead?
A very small canvas, not resizable, that makes it very difficult to use with a
computer touch screen (drawing kanji with a finger)
What version of the product are you using? On what operating system?
Arch linux, mozc-im (ibus-mozc, mozc, uim-mozc) v1.15.1857.102-1
Please provide any additional information below.
I would like to be able to resize the drawing canvas as I wish (or have it
resize automatically depending on the window size), so I can replace
tegaki-recognize with mozc_tool hand_writing
```
Original issue reported on code.google.com by `lemaire.adrien` on 12 Aug 2014 at 11:20
| defect | hand writing canvas size what steps will reproduce the problem start hand writing mode usr lib mozc mozc tool mode hand writing what is the expected output being able to resize the drawing canvas manually handles or via settings what do you see instead a very small canvas not resizable that makes it very difficult to use with a computer touch screen drawing kanji with a finger what version of the product are you using on what operating system arch linux mozc im ibus mozc mozc uim mozc please provide any additional information below i would like to be able to resize the drawing canvas as i wish or have it resize automatically depending on the window size so i can replace tegaki recognize with mozc tool hand writing original issue reported on code google com by lemaire adrien on aug at | 1 |
101,135 | 4,107,967,020 | IssuesEvent | 2016-06-06 14:44:41 | DDMAL/Rodan | https://api.github.com/repos/DDMAL/Rodan | opened | Creation of empty ResourceList w/o ResourceType should have default ResourceType | Priority: HIGH Type: BUG | When creating an empty ResourceList and a ResourceType is not provided, Rodan should set the default ResourceType to 'application/octet-stream'.
| 1.0 | Creation of empty ResourceList w/o ResourceType should have default ResourceType - When creating an empty ResourceList and a ResourceType is not provided, Rodan should set the default ResourceType to 'application/octet-stream'.
| non_defect | creation of empty resourcelist w o resourcetype should have default resourcetype when creating an empty resourcelist and a resourcetype is not provided rodan should set the default resourcetype to application octet stream | 0 |
80,986 | 30,646,815,430 | IssuesEvent | 2023-07-25 05:45:08 | line/armeria | https://api.github.com/repos/line/armeria | closed | Error using abstract namespace socket path with Unix Domain Socket setup. | defect | I'm trying to get the Unix Domain Socket Support support from 1.24.0 working. I'm having trouble specifying a socket that's in the abstract namespace (e.g. "@foo.sock"). My previous uds proxy server used null in the socket path ("\u0000foo.sock") and this worked fine. Now if I do it, I get something like:
```
java.nio.file.InvalidPathException: Nul character not allowed: \u0000foo.sock
at java.base/sun.nio.fs.UnixPath.checkNotNul(UnixPath.java:90)
at java.base/sun.nio.fs.UnixPath.normalizeAndCheck(UnixPath.java:80)
at java.base/sun.nio.fs.UnixPath.<init>(UnixPath.java:68)
at java.base/sun.nio.fs.UnixFileSystem.getPath(UnixFileSystem.java:279)
at java.base/java.nio.file.Path.of(Path.java:147)
at java.base/java.nio.file.Paths.get(Paths.java:69)
at com.linecorp.armeria.common.util.DomainSocketAddress.of(DomainSocketAddress.java:59)
```
The call I used to generate the above:
`serverBuilder.http(DomainSocketAddress.of(io.netty.channel.unix.DomainSocketAddress(socketPath) )` | 1.0 | Error using abstract namespace socket path with Unix Domain Socket setup. - I'm trying to get the Unix Domain Socket Support support from 1.24.0 working. I'm having trouble specifying a socket that's in the abstract namespace (e.g. "@foo.sock"). My previous uds proxy server used null in the socket path ("\u0000foo.sock") and this worked fine. Now if I do it, I get something like:
```
java.nio.file.InvalidPathException: Nul character not allowed: \u0000foo.sock
at java.base/sun.nio.fs.UnixPath.checkNotNul(UnixPath.java:90)
at java.base/sun.nio.fs.UnixPath.normalizeAndCheck(UnixPath.java:80)
at java.base/sun.nio.fs.UnixPath.<init>(UnixPath.java:68)
at java.base/sun.nio.fs.UnixFileSystem.getPath(UnixFileSystem.java:279)
at java.base/java.nio.file.Path.of(Path.java:147)
at java.base/java.nio.file.Paths.get(Paths.java:69)
at com.linecorp.armeria.common.util.DomainSocketAddress.of(DomainSocketAddress.java:59)
```
The call I used to generate the above:
`serverBuilder.http(DomainSocketAddress.of(io.netty.channel.unix.DomainSocketAddress(socketPath) )` | defect | error using abstract namespace socket path with unix domain socket setup i m trying to get the unix domain socket support support from working i m having trouble specifying a socket that s in the abstract namespace e g foo sock my previous uds proxy server used null in the socket path sock and this worked fine now if i do it i get something like java nio file invalidpathexception nul character not allowed sock at java base sun nio fs unixpath checknotnul unixpath java at java base sun nio fs unixpath normalizeandcheck unixpath java at java base sun nio fs unixpath unixpath java at java base sun nio fs unixfilesystem getpath unixfilesystem java at java base java nio file path of path java at java base java nio file paths get paths java at com linecorp armeria common util domainsocketaddress of domainsocketaddress java the call i used to generate the above serverbuilder http domainsocketaddress of io netty channel unix domainsocketaddress socketpath | 1 |
3,506 | 2,610,063,924 | IssuesEvent | 2015-02-26 18:18:48 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | ้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑ | auto-migrated Priority-Medium Type-Defect | ```
้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ24ๅฐๆถๅฅ
ๅบทๅจ่ฏข็ญ็บฟ:0576-88066933-(ๆฃๆฃ800080609)-(ๅพฎไฟกๅทtzwzszyy)ๅป้ขๅฐๅ:
ๅฐๅทๅธๆคๆฑๅบๆซๅ่ทฏ229ๅท๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ:ไนๅ104๏ฟฝ
๏ฟฝ๏ฟฝ108ใ118ใ198ๅๆคๆฑไธ้ๆธ
ๅ
ฌไบค่ฝฆ็ด่พพๆซๅๅฐๅบ๏ผไนๅ107ใ105
ใ109ใ112ใ901ใ 902ๅ
ฌไบค่ฝฆๅฐๆๆๅนฟๅบไธ่ฝฆ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ
่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ
๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ
ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ
๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ
๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ
่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ
็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:59 | 1.0 | ้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑ - ```
้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ24ๅฐๆถๅฅ
ๅบทๅจ่ฏข็ญ็บฟ:0576-88066933-(ๆฃๆฃ800080609)-(ๅพฎไฟกๅทtzwzszyy)ๅป้ขๅฐๅ:
ๅฐๅทๅธๆคๆฑๅบๆซๅ่ทฏ229ๅท๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ:ไนๅ104๏ฟฝ
๏ฟฝ๏ฟฝ108ใ118ใ198ๅๆคๆฑไธ้ๆธ
ๅ
ฌไบค่ฝฆ็ด่พพๆซๅๅฐๅบ๏ผไนๅ107ใ105
ใ109ใ112ใ901ใ 902ๅ
ฌไบค่ฝฆๅฐๆๆๅนฟๅบไธ่ฝฆ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ
่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ
๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ
ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ
๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ
๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ
่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ
็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:59 | defect | ้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑ ้ปๅฒฉๆฃๆฅไธๅญไธ่ฒ้่ฆๅคๅฐ้ฑใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ ๅบทๅจ่ฏข็ญ็บฟ ๅพฎไฟกๅทtzwzszyy ๅป้ขๅฐๅ ๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ ๏ฟฝ ๏ฟฝ๏ฟฝ ใ ใ ๏ผ ใ ใ ใ ใ ใ ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ ่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ ๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ ๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ ๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ ่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ ็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ original issue reported on code google com by poweragr gmail com on may at | 1 |
3,330 | 2,610,060,730 | IssuesEvent | 2015-02-26 18:17:53 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | ่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝ | auto-migrated Priority-Medium Type-Defect | ```
่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ24ๅฐๆถๅฅๅบทๅจ่ฏข
็ญ็บฟ:0576-88066933-(ๆฃๆฃ800080609)-(ๅพฎไฟกๅทtzwzszyy)ๅป้ขๅฐๅ:ๅฐๅทๅธ
ๆคๆฑๅบๆซๅ่ทฏ229ๅท๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ:ไนๅ104ใ108ใ1
18ใ198ๅๆคๆฑไธ้ๆธ
ๅ
ฌไบค่ฝฆ็ด่พพๆซๅๅฐๅบ๏ผไนๅ107ใ105ใ109ใ
112ใ901ใ 902ๅ
ฌไบค่ฝฆๅฐๆๆๅนฟๅบไธ่ฝฆ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ
่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ
๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ
ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ
๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ
๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ
่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ
็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:14 | 1.0 | ่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝ - ```
่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ24ๅฐๆถๅฅๅบทๅจ่ฏข
็ญ็บฟ:0576-88066933-(ๆฃๆฃ800080609)-(ๅพฎไฟกๅทtzwzszyy)ๅป้ขๅฐๅ:ๅฐๅทๅธ
ๆคๆฑๅบๆซๅ่ทฏ229ๅท๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ:ไนๅ104ใ108ใ1
18ใ198ๅๆคๆฑไธ้ๆธ
ๅ
ฌไบค่ฝฆ็ด่พพๆซๅๅฐๅบ๏ผไนๅ107ใ105ใ109ใ
112ใ901ใ 902ๅ
ฌไบค่ฝฆๅฐๆๆๅนฟๅบไธ่ฝฆ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ
่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ
๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ
ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ
๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ
๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ
่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ
็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:14 | defect | ่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝ ่ทฏๆกฅๆฒปไธ่ฒๅช้ๆๆๅฅฝใๅฐๅทไบๆดฒ็ๆฎๅป้ขใ ็ญ็บฟ ๅพฎไฟกๅทtzwzszyy ๅป้ขๅฐๅ ๅฐๅทๅธ ๏ผๆซๅๅคง่ฝฌ็ๆ๏ผไน่ฝฆ็บฟ่ทฏ ใ ใ ใ ๏ผ ใ ใ ใ ใ ใ ๏ผๆญฅ่กๅณๅฏๅฐ้ขใ ่ฏ็้กน็ฎ๏ผ้ณ็ฟ๏ผๆฉๆณ๏ผๅๅ่
บ็๏ผๅๅ่
บๅข็๏ผ้พๅคด็๏ผ๏ฟฝ๏ฟฝ ๏ฟฝ็ฒพ๏ผๆ ็ฒพใๅ
็ฎๅ
่๏ผ็ฒพ็ดข้่ๆฒๅผ ๏ผๆท็
็ญใ ๅฐๅทไบๆดฒ็ๆฎๅป้ขๆฏๅฐๅทๆๅคง็็ท็งๅป้ข๏ผๆๅจไธๅฎถๅจ็บฟๅ
๏ฟฝ๏ฟฝ ๏ฟฝๅจ่ฏข๏ผๆฅๆไธไธๅฎๅ็็ท็งๆฃๆฅๆฒป็่ฎพๅค๏ผไธฅๆ ผๆ็
งๅฝๅฎถๆ ๏ฟฝ ๏ฟฝ๏ฟฝๆถ่ดนใๅฐ็ซฏๅป็่ฎพๅค๏ผไธไธ็ๅๆญฅใๆๅจไธๅฎถ๏ผๆๅฐฑไธไธๅ
ธ ่ใไบบๆงๅๆๅก๏ผไธๅไปฅๆฃ่
ไธบไธญๅฟใ ็็ท็งๅฐฑ้ๅฐๅทไบๆดฒ็ๆฎๅป้ข๏ผไธไธ็ท็งไธบ็ทไบบใ original issue reported on code google com by poweragr gmail com on may at | 1 |
230,671 | 18,682,620,595 | IssuesEvent | 2021-11-01 08:18:07 | BHoM/SAP_Toolkit | https://api.github.com/repos/BHoM/SAP_Toolkit | opened | SAP_Toolkit: Set up example set of dwellings for 2D and 3D workflows | type:test-script | <!-- PLEASE ENSURE YOU REVIEW THE CONTENT OF EACH ISSUE CAREFULLY, INCLUDING SUBSEQUENT COMMENTS BY YOURSELF OR OTHERS. -->
<!-- IN PARTICULAR PLEASE ENSURE THAT SENSITIVE OR INAPPROPRIATE INFORMATION IS NOT UPLOADED -->
#### Definition of the test :
<!-- Please describe the test, including the proposed scope and range of input parameters -->
| 1.0 | SAP_Toolkit: Set up example set of dwellings for 2D and 3D workflows - <!-- PLEASE ENSURE YOU REVIEW THE CONTENT OF EACH ISSUE CAREFULLY, INCLUDING SUBSEQUENT COMMENTS BY YOURSELF OR OTHERS. -->
<!-- IN PARTICULAR PLEASE ENSURE THAT SENSITIVE OR INAPPROPRIATE INFORMATION IS NOT UPLOADED -->
#### Definition of the test :
<!-- Please describe the test, including the proposed scope and range of input parameters -->
| non_defect | sap toolkit set up example set of dwellings for and workflows definition of the test | 0 |
32,782 | 15,615,544,662 | IssuesEvent | 2021-03-19 19:22:15 | microsoft/azuredatastudio | https://api.github.com/repos/microsoft/azuredatastudio | closed | Declarative table with multiple columns of components causes UI to freeze. | Area - Extensibility Area - Performance Bug Triage: Done | When trying to fill in the data of a declarative table that contains two columns of valueType component, the UI freezes. The table needs to be able to have about 300 rows. When trying to populate this it causes the UI to automatically freeze. The page takes some time to load and although it will show the populated rows eventually, the page not responsive.
When decreasing the table to 20 rows, the UI is able to perform some functions before becoming nonresponsive. This can be seen when trying to update the table by filtering out rows. (This also shows another issue that the table removes the components from columns when updating.)
Example:
Initial load of 20 rows
<img width="400" alt="initial" src="https://user-images.githubusercontent.com/69922333/103569777-c8ff0700-4e7c-11eb-89cb-c6c8826874cf.PNG">
After filtering the number of rows
<img width="400" alt="filter" src="https://user-images.githubusercontent.com/69922333/103569846-e0d68b00-4e7c-11eb-9435-a7b11efb4701.PNG">
Steps to Reproduce:
1. For a a minimal repro, checkout branch [nasc/declarativeTableFreeze](https://github.com/microsoft/azuredatastudio/tree/nasc/declarativeTableFreeze/extensions/arcRemovedDeclarativeTableBug). It adds an extension that opens a window with this behavior.
2. The code to alter number of parameters used looks like this in extension.ts
<img width="400" alt="redocode" src="https://user-images.githubusercontent.com/69922333/103700464-57e15180-4f59-11eb-94a1-44c184d01728.PNG">
4. When using engineSettings.getSmallData() the tab will open with no issue. Test the search box a few times and observe when it freezes.
5. When using engineSettings.getLargeData() it can take up to 15 seconds for the table to populate but the UI will not be responsive. | True | Declarative table with multiple columns of components causes UI to freeze. - When trying to fill in the data of a declarative table that contains two columns of valueType component, the UI freezes. The table needs to be able to have about 300 rows. When trying to populate this it causes the UI to automatically freeze. The page takes some time to load and although it will show the populated rows eventually, the page not responsive.
When decreasing the table to 20 rows, the UI is able to perform some functions before becoming nonresponsive. This can be seen when trying to update the table by filtering out rows. (This also shows another issue that the table removes the components from columns when updating.)
Example:
Initial load of 20 rows
<img width="400" alt="initial" src="https://user-images.githubusercontent.com/69922333/103569777-c8ff0700-4e7c-11eb-89cb-c6c8826874cf.PNG">
After filtering the number of rows
<img width="400" alt="filter" src="https://user-images.githubusercontent.com/69922333/103569846-e0d68b00-4e7c-11eb-9435-a7b11efb4701.PNG">
Steps to Reproduce:
1. For a a minimal repro, checkout branch [nasc/declarativeTableFreeze](https://github.com/microsoft/azuredatastudio/tree/nasc/declarativeTableFreeze/extensions/arcRemovedDeclarativeTableBug). It adds an extension that opens a window with this behavior.
2. The code to alter number of parameters used looks like this in extension.ts
<img width="400" alt="redocode" src="https://user-images.githubusercontent.com/69922333/103700464-57e15180-4f59-11eb-94a1-44c184d01728.PNG">
4. When using engineSettings.getSmallData() the tab will open with no issue. Test the search box a few times and observe when it freezes.
5. When using engineSettings.getLargeData() it can take up to 15 seconds for the table to populate but the UI will not be responsive. | non_defect | declarative table with multiple columns of components causes ui to freeze when trying to fill in the data of a declarative table that contains two columns of valuetype component the ui freezes the table needs to be able to have about rows when trying to populate this it causes the ui to automatically freeze the page takes some time to load and although it will show the populated rows eventually the page not responsive when decreasing the table to rows the ui is able to perform some functions before becoming nonresponsive this can be seen when trying to update the table by filtering out rows this also shows another issue that the table removes the components from columns when updating example initial load of rows img width alt initial src after filtering the number of rows img width alt filter src steps to reproduce for a a minimal repro checkout branch it adds an extension that opens a window with this behavior the code to alter number of parameters used looks like this in extension ts img width alt redocode src when using enginesettings getsmalldata the tab will open with no issue test the search box a few times and observe when it freezes when using enginesettings getlargedata it can take up to seconds for the table to populate but the ui will not be responsive | 0 |
3,123 | 2,607,984,839 | IssuesEvent | 2015-02-26 00:51:12 | chrsmithdemos/zen-coding | https://api.github.com/repos/chrsmithdemos/zen-coding | closed | css3 ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ะฑะตะท. | auto-migrated Priority-Medium Type-Defect | ```
ะะฐ ะฟัะฐะบัะธะบะต ะฒัะตะณะดะฐ ะฟัะธั
ะพะดะธััั ะฟะธัะฐัั css3
ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ัะฐะผะพ ัะฒะพะนััะฒะฐ ะฒ
ะบะพะฝัะต. ัะพ-ะฑะธัั:
-webkit-box-shadow: 0 0 0 #000;
-moz-box-shadow: 0 0 0 #000;
box-shadow: 0 0 0 #000;
C ะฟะพะผะพััั zendcoding:
bxsh:w+
bxsh:m+
bxsh+
ะะพะถะตั ัะพะบัะฐัะตะฝะฝัั ะทะฐะฟะธัั ัะดะตะปะฐัั?
ะะฐะฟัะธะผะตั:
bxsh++
ะธะปะธ
bxsh:wms+ / bxsh:wms
w:-webkit-,m:-moz-, s: self - ะฑะตะท ะฟัะตัะธะบัะฐ
ะธะปะธ...
```
-----
Original issue reported on code.google.com by `kratk...@gmail.com` on 24 Nov 2011 at 12:17 | 1.0 | css3 ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ะฑะตะท. - ```
ะะฐ ะฟัะฐะบัะธะบะต ะฒัะตะณะดะฐ ะฟัะธั
ะพะดะธััั ะฟะธัะฐัั css3
ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ัะฐะผะพ ัะฒะพะนััะฒะฐ ะฒ
ะบะพะฝัะต. ัะพ-ะฑะธัั:
-webkit-box-shadow: 0 0 0 #000;
-moz-box-shadow: 0 0 0 #000;
box-shadow: 0 0 0 #000;
C ะฟะพะผะพััั zendcoding:
bxsh:w+
bxsh:m+
bxsh+
ะะพะถะตั ัะพะบัะฐัะตะฝะฝัั ะทะฐะฟะธัั ัะดะตะปะฐัั?
ะะฐะฟัะธะผะตั:
bxsh++
ะธะปะธ
bxsh:wms+ / bxsh:wms
w:-webkit-,m:-moz-, s: self - ะฑะตะท ะฟัะตัะธะบัะฐ
ะธะปะธ...
```
-----
Original issue reported on code.google.com by `kratk...@gmail.com` on 24 Nov 2011 at 12:17 | defect | ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ะฑะตะท ะฝะฐ ะฟัะฐะบัะธะบะต ะฒัะตะณะดะฐ ะฟัะธั
ะพะดะธััั ะฟะธัะฐัั ัะฒะพะนััะฒะฐ ั ะฟัะตัะธะบัะฐะผะธ ะธ ัะฐะผะพ ัะฒะพะนััะฒะฐ ะฒ ะบะพะฝัะต ัะพ ะฑะธัั webkit box shadow moz box shadow box shadow c ะฟะพะผะพััั zendcoding bxsh w bxsh m bxsh ะผะพะถะตั ัะพะบัะฐัะตะฝะฝัั ะทะฐะฟะธัั ัะดะตะปะฐัั ะฝะฐะฟัะธะผะตั bxsh ะธะปะธ bxsh wms bxsh wms w webkit m moz s self ะฑะตะท ะฟัะตัะธะบัะฐ ะธะปะธ original issue reported on code google com by kratk gmail com on nov at | 1 |
24,204 | 3,924,419,134 | IssuesEvent | 2016-04-22 15:09:05 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Checkpointing with parallel mesh isn't quite right | C: MOOSE P: minor T: defect | ### Description of the enhancement or error report
When using checkpoint=true while running with parallel mesh, the mesh checkpoint files are not deleted. The issue is that the checkpoint filename is not correct.
The correct output files get created:
e.g. "0000_mesh.cpr-\<n\>" where the last part "-\<n\>" represents the processor id.
The Checkpoint Outputter doesn't delete the right filename. It only tries to delete this pattern:
"0000_mesh.cpr"
### Rationale for the enhancement or information for reproducing the error
You'll get a lot of checkpoint files piling up if you were to use this combination in a real simulation. This could case disk space issues.
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
This bug makes this combination expensive to use (in terms of disk space). | 1.0 | Checkpointing with parallel mesh isn't quite right - ### Description of the enhancement or error report
When using checkpoint=true while running with parallel mesh, the mesh checkpoint files are not deleted. The issue is that the checkpoint filename is not correct.
The correct output files get created:
e.g. "0000_mesh.cpr-\<n\>" where the last part "-\<n\>" represents the processor id.
The Checkpoint Outputter doesn't delete the right filename. It only tries to delete this pattern:
"0000_mesh.cpr"
### Rationale for the enhancement or information for reproducing the error
You'll get a lot of checkpoint files piling up if you were to use this combination in a real simulation. This could case disk space issues.
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
This bug makes this combination expensive to use (in terms of disk space). | defect | checkpointing with parallel mesh isn t quite right description of the enhancement or error report when using checkpoint true while running with parallel mesh the mesh checkpoint files are not deleted the issue is that the checkpoint filename is not correct the correct output files get created e g mesh cpr where the last part represents the processor id the checkpoint outputter doesn t delete the right filename it only tries to delete this pattern mesh cpr rationale for the enhancement or information for reproducing the error you ll get a lot of checkpoint files piling up if you were to use this combination in a real simulation this could case disk space issues identified impact i e internal object changes limited interface changes public api change or a list of specific applications impacted this bug makes this combination expensive to use in terms of disk space | 1 |
248,517 | 21,037,236,126 | IssuesEvent | 2022-03-31 08:59:12 | consta-design-system/uikit | https://api.github.com/repos/consta-design-system/uikit | reopened | ContextMenu: ย ะดะพะฑะฐะฒะธัั ัะตััั | test ๐ฅ๐ฅ๐ฅ priority | > ะะพะปะฝัะต [ะฟัะฐะฒะธะปะฐ ะพัะพัะผะปะตะฝะธั issue](https://consta-uikit.vercel.app/?path=/docs/common-develop-issues--page)
**ะะฟะธัะฐะฝะธะต ัะธัะธ ะธะปะธ ัะปัััะตะฝะธั**
ะั ะฑัะดะตะผ ัะฐะดั:
- ะฟะพะดัะพะฑะฝะพะผั ะพะฟะธัะฐะฝะธั
- ัะบัะธะฝัะพัะฐะผ ะธ ะปัะฑัะผ ะดััะณะธะผ ะธะทะพะฑัะฐะถะตะฝะธัะผ
ะัะปะธ ะธะทะพะฑัะฐะถะตะฝะธั ะฝะต ะฟัะธะบัะตะฟะปััััั ะบ ะทะฐะดะฐัะต, ะฟัะธะบัะตะฟะธัะต ะฐัั
ะธะฒ ะธะปะธ ะฟะพััะฐะฒััะต ัััะปะบั ะฝะฐ ะพะฑะปะฐะบะพ. ะะฑััะผ ะปัะฑะพะน, ะณะปะฐะฒะฝะพะต โ ะฟะพะฝััะฝะพ ะธ ะฟะพ-ัะตะปะพะฒะตัะตัะบะธ. ะฅะพัะตััั, ััะพะฑั ะฒ ะปัะฑะพะน ะผะพะผะตะฝั ะธ ะฐะฒัะพัั, ะธ ะธัะฟะพะปะฝะธัะตะปั ะฑัะปะพ ะฟะพะฝััะฝะพ, ััะพ ะดะตะปะฐัั.
**ะะพัะตะผั ัะปัััะตะฝะธะต ะธะปะธ ัะธัะฐ ะฝะตะพะฑั
ะพะดะธะผะฐ?**
ะะฟะธัะธัะต ะฐัะณัะผะตะฝัั ะฒ ะฟะพะปัะทั ัะตัะตะฝะธั
**ะ ะตะฐะปะธะทะฐัะธั ะฟัะตะดะปะพะถะตะฝะธั**
ะะฟะธัะธัะต ัะฟะพัะพะฑั ัะตะฐะปะธะทะฐัะธะธ ัะตัะตะฝะธั
**ะะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั**
ะัะต, ััะพ ััะธัะฐะตัะต ะฝัะถะฝัะผ
| 1.0 | ContextMenu: ย ะดะพะฑะฐะฒะธัั ัะตััั - > ะะพะปะฝัะต [ะฟัะฐะฒะธะปะฐ ะพัะพัะผะปะตะฝะธั issue](https://consta-uikit.vercel.app/?path=/docs/common-develop-issues--page)
**ะะฟะธัะฐะฝะธะต ัะธัะธ ะธะปะธ ัะปัััะตะฝะธั**
ะั ะฑัะดะตะผ ัะฐะดั:
- ะฟะพะดัะพะฑะฝะพะผั ะพะฟะธัะฐะฝะธั
- ัะบัะธะฝัะพัะฐะผ ะธ ะปัะฑัะผ ะดััะณะธะผ ะธะทะพะฑัะฐะถะตะฝะธัะผ
ะัะปะธ ะธะทะพะฑัะฐะถะตะฝะธั ะฝะต ะฟัะธะบัะตะฟะปััััั ะบ ะทะฐะดะฐัะต, ะฟัะธะบัะตะฟะธัะต ะฐัั
ะธะฒ ะธะปะธ ะฟะพััะฐะฒััะต ัััะปะบั ะฝะฐ ะพะฑะปะฐะบะพ. ะะฑััะผ ะปัะฑะพะน, ะณะปะฐะฒะฝะพะต โ ะฟะพะฝััะฝะพ ะธ ะฟะพ-ัะตะปะพะฒะตัะตัะบะธ. ะฅะพัะตััั, ััะพะฑั ะฒ ะปัะฑะพะน ะผะพะผะตะฝั ะธ ะฐะฒัะพัั, ะธ ะธัะฟะพะปะฝะธัะตะปั ะฑัะปะพ ะฟะพะฝััะฝะพ, ััะพ ะดะตะปะฐัั.
**ะะพัะตะผั ัะปัััะตะฝะธะต ะธะปะธ ัะธัะฐ ะฝะตะพะฑั
ะพะดะธะผะฐ?**
ะะฟะธัะธัะต ะฐัะณัะผะตะฝัั ะฒ ะฟะพะปัะทั ัะตัะตะฝะธั
**ะ ะตะฐะปะธะทะฐัะธั ะฟัะตะดะปะพะถะตะฝะธั**
ะะฟะธัะธัะต ัะฟะพัะพะฑั ัะตะฐะปะธะทะฐัะธะธ ัะตัะตะฝะธั
**ะะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั**
ะัะต, ััะพ ััะธัะฐะตัะต ะฝัะถะฝัะผ
| non_defect | contextmenu ย ะดะพะฑะฐะฒะธัั ัะตััั ะฟะพะปะฝัะต ะพะฟะธัะฐะฝะธะต ัะธัะธ ะธะปะธ ัะปัััะตะฝะธั ะผั ะฑัะดะตะผ ัะฐะดั ะฟะพะดัะพะฑะฝะพะผั ะพะฟะธัะฐะฝะธั ัะบัะธะฝัะพัะฐะผ ะธ ะปัะฑัะผ ะดััะณะธะผ ะธะทะพะฑัะฐะถะตะฝะธัะผ ะตัะปะธ ะธะทะพะฑัะฐะถะตะฝะธั ะฝะต ะฟัะธะบัะตะฟะปััััั ะบ ะทะฐะดะฐัะต ะฟัะธะบัะตะฟะธัะต ะฐัั
ะธะฒ ะธะปะธ ะฟะพััะฐะฒััะต ัััะปะบั ะฝะฐ ะพะฑะปะฐะบะพ ะพะฑััะผ ะปัะฑะพะน ะณะปะฐะฒะฝะพะต โ ะฟะพะฝััะฝะพ ะธ ะฟะพ ัะตะปะพะฒะตัะตัะบะธ ั
ะพัะตััั ััะพะฑั ะฒ ะปัะฑะพะน ะผะพะผะตะฝั ะธ ะฐะฒัะพัั ะธ ะธัะฟะพะปะฝะธัะตะปั ะฑัะปะพ ะฟะพะฝััะฝะพ ััะพ ะดะตะปะฐัั ะฟะพัะตะผั ัะปัััะตะฝะธะต ะธะปะธ ัะธัะฐ ะฝะตะพะฑั
ะพะดะธะผะฐ ะพะฟะธัะธัะต ะฐัะณัะผะตะฝัั ะฒ ะฟะพะปัะทั ัะตัะตะฝะธั ัะตะฐะปะธะทะฐัะธั ะฟัะตะดะปะพะถะตะฝะธั ะพะฟะธัะธัะต ัะฟะพัะพะฑั ัะตะฐะปะธะทะฐัะธะธ ัะตัะตะฝะธั ะดะพะฟะพะปะฝะธัะตะปัะฝะฐั ะธะฝัะพัะผะฐัะธั ะฒัะต ััะพ ััะธัะฐะตัะต ะฝัะถะฝัะผ | 0 |
22,215 | 3,618,875,522 | IssuesEvent | 2016-02-08 13:51:39 | ApertureLabsLtd/mphidflash | https://api.github.com/repos/ApertureLabsLtd/mphidflash | closed | PIC32MX675F512H | auto-migrated Priority-Medium Type-Defect | ```
Hi,
Recently, I use PIC32MX675F512H.
I wrote USB HID bootloader and succeeded to erase, program and verify
with PIC32UBL.exe on Windows7. Of course, the application firmware is
no trouble.
However, the latest mphidflash(svn trunk ver.) cannot be used on this
bootloader.
I tried on Mac OS X 10.6.8 and the massage "USB HID device found" is shown
but the process is stopped.
Are there any solution?
*PIC32 bootloader's url
http://www.microchip.com/stellent/idcplg?IdcService=SS_GET_PAGE&nodeId=1824&appn
ote=en554836
Thank,
```
Original issue reported on code.google.com by `yamamo2s...@gmail.com` on 1 Sep 2011 at 1:52 | 1.0 | PIC32MX675F512H - ```
Hi,
Recently, I use PIC32MX675F512H.
I wrote USB HID bootloader and succeeded to erase, program and verify
with PIC32UBL.exe on Windows7. Of course, the application firmware is
no trouble.
However, the latest mphidflash(svn trunk ver.) cannot be used on this
bootloader.
I tried on Mac OS X 10.6.8 and the massage "USB HID device found" is shown
but the process is stopped.
Are there any solution?
*PIC32 bootloader's url
http://www.microchip.com/stellent/idcplg?IdcService=SS_GET_PAGE&nodeId=1824&appn
ote=en554836
Thank,
```
Original issue reported on code.google.com by `yamamo2s...@gmail.com` on 1 Sep 2011 at 1:52 | defect | hi recently i use i wrote usb hid bootloader and succeeded to erase program and verify with exe on of course the application firmware is no trouble however the latest mphidflash svn trunk ver cannot be used on this bootloader i tried on mac os x and the massage usb hid device found is shown but the process is stopped are there any solution bootloader s url ote thank original issue reported on code google com by gmail com on sep at | 1 |
14,102 | 2,789,898,988 | IssuesEvent | 2015-05-08 22:16:11 | google/google-visualization-api-issues | https://api.github.com/repos/google/google-visualization-api-issues | opened | Bug: Table width setting is skipped | Priority-Medium Type-Defect | Original [issue 472](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=472) created by orwant on 2010-12-17T13:57:38.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Create a new google.visualization.Table object inside a DIV with "overflow-y:auto".
2. Set the width-property to n-pixels: width: '800px'
3. Draw the table
4. Observe that the table's size is no wider than that required by the columns.
What component is this issue related to:
Issue is related to the google.visualization.Table object.
Are you using the test environment (version 1.1)? Yes
<b>What operating system and browser are you using?</b>
Windows 7
Firefox 3.6.13
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | Bug: Table width setting is skipped - Original [issue 472](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=472) created by orwant on 2010-12-17T13:57:38.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Create a new google.visualization.Table object inside a DIV with "overflow-y:auto".
2. Set the width-property to n-pixels: width: '800px'
3. Draw the table
4. Observe that the table's size is no wider than that required by the columns.
What component is this issue related to:
Issue is related to the google.visualization.Table object.
Are you using the test environment (version 1.1)? Yes
<b>What operating system and browser are you using?</b>
Windows 7
Firefox 3.6.13
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| defect | bug table width setting is skipped original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code create a new google visualization table object inside a div with quot overflow y auto quot set the width property to n pixels width draw the table observe that the table s size is no wider than that required by the columns what component is this issue related to issue is related to the google visualization table object are you using the test environment version yes what operating system and browser are you using windows firefox for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 1 |
338,395 | 30,295,722,904 | IssuesEvent | 2023-07-09 20:42:00 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | reopened | Frequent test failures of `TestStartStop/group/newest-cni/serial/VerifyKubernetesImages` | priority/backlog lifecycle/rotten kind/failing-test | This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[QEMU_macOS](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=QEMU_macOS&test=TestStartStop/group/newest-cni/serial/VerifyKubernetesImages&period=last90)|100.00| | 1.0 | Frequent test failures of `TestStartStop/group/newest-cni/serial/VerifyKubernetesImages` - This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[QEMU_macOS](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=QEMU_macOS&test=TestStartStop/group/newest-cni/serial/VerifyKubernetesImages&period=last90)|100.00| | non_defect | frequent test failures of teststartstop group newest cni serial verifykubernetesimages this test has high flake rates for the following environments environment flake rate | 0 |
674,285 | 23,045,039,429 | IssuesEvent | 2022-07-23 19:16:36 | pystardust/ani-cli | https://api.github.com/repos/pystardust/ani-cli | closed | "Episodes not released yet!" Error For All Anime | type: bug priority 2: medium | Version: Latest
OS: Manjaro Linux for x64
Shell: Bash
Attempting to view any anime results in the error described in the title.
I can search for anime just fine, but once I select one and the program searches for episodes,
it errors out.
The program worked completely fine yesterday, so I believe that this is an issue with the providers that we scrape. I have already tried reinstalling.
**Steps To Reproduce**
1. Run the program.
2. Search for anime.
3. Select one.
**Expected behavior**
A list of episodes should be shown
| 1.0 | "Episodes not released yet!" Error For All Anime - Version: Latest
OS: Manjaro Linux for x64
Shell: Bash
Attempting to view any anime results in the error described in the title.
I can search for anime just fine, but once I select one and the program searches for episodes,
it errors out.
The program worked completely fine yesterday, so I believe that this is an issue with the providers that we scrape. I have already tried reinstalling.
**Steps To Reproduce**
1. Run the program.
2. Search for anime.
3. Select one.
**Expected behavior**
A list of episodes should be shown
| non_defect | episodes not released yet error for all anime version latest os manjaro linux for shell bash attempting to view any anime results in the error described in the title i can search for anime just fine but once i select one and the program searches for episodes it errors out the program worked completely fine yesterday so i believe that this is an issue with the providers that we scrape i have already tried reinstalling steps to reproduce run the program search for anime select one expected behavior a list of episodes should be shown | 0 |
469,731 | 13,524,525,440 | IssuesEvent | 2020-09-15 11:41:21 | Poobslag/turbofat | https://api.github.com/repos/Poobslag/turbofat | opened | Overworld scene should be cached for instant loading | priority-4 | The overworld scene should be cached for instant loading.
There is a branch at https://github.com/Poobslag/turbofat/tree/instant-puzzle-launch-time where we attempted this for the puzzle scene, but it introduced some bugs and is incomplete.
This will also necessitate a way to populate the overworld with the default moving creatures. These creatures should follow arbitrary paths instead of having multiple scripts. For example, 'Run toward (150, 235) until you get there... then return to (0, 446)... then repeat' | 1.0 | Overworld scene should be cached for instant loading - The overworld scene should be cached for instant loading.
There is a branch at https://github.com/Poobslag/turbofat/tree/instant-puzzle-launch-time where we attempted this for the puzzle scene, but it introduced some bugs and is incomplete.
This will also necessitate a way to populate the overworld with the default moving creatures. These creatures should follow arbitrary paths instead of having multiple scripts. For example, 'Run toward (150, 235) until you get there... then return to (0, 446)... then repeat' | non_defect | overworld scene should be cached for instant loading the overworld scene should be cached for instant loading there is a branch at where we attempted this for the puzzle scene but it introduced some bugs and is incomplete this will also necessitate a way to populate the overworld with the default moving creatures these creatures should follow arbitrary paths instead of having multiple scripts for example run toward until you get there then return to then repeat | 0 |
11,577 | 2,658,737,997 | IssuesEvent | 2015-03-18 17:07:10 | ceres-solver/ceres-solver | https://api.github.com/repos/ceres-solver/ceres-solver | closed | ceres-solver 1.9.0 from MacPorts segfaults on a particular problem | Priority-Medium Type-Defect | Original [issue 150](https://code.google.com/p/ceres-solver/issues/detail?id=150) created by sandwichmaker on 2014-08-13T16:15:11.000Z:
<b>What steps will reproduce the problem?</b>
1. Install ceres-solver with MacPorts (either from binary with "port install ceres-solver" or source with "port -s install ceres-solver").
2. Compile the program at http://julianpanetta.com/macports_bugs/test_ceres.cc with /usr/bin/clang++:
/usr/bin/clang++ -g -O0 -std=c++11 -I/opt/local/include
-I/opt/local/include/eigen3 test_ceres.cc -L/opt/local/lib -lceres -lglog
-lgflags -lcholmod -lcxsparse -framework accelerate -o broken
3) ./broken
<b>What is the expected output? What do you see instead?</b>
It should print 0..19, but it intermittently segfaults partway through or throws a std::length error because Ceres has scribbled over the arrays v1s and v2s.
<b>What version of the product are you using? On what operating system?</b>
I've tried with 1.9.0 installed from MacPorts on both OS 10.9.3 and 10.9.4.
<b>Please provide any additional information below.</b>
I've filed a ticket with MacPorts because this only seems to happen with a MacPorts-installed (statically linked) Ceres. The test program runs fine when statically linked against a Ceres version manually built with cmake (both the git head, and the stable 1.9.0 release). However, the MacPorts configuration is so minimal it's unclear where the problem could be. More details are at:
https://trac.macports.org/ticket/44627 | 1.0 | ceres-solver 1.9.0 from MacPorts segfaults on a particular problem - Original [issue 150](https://code.google.com/p/ceres-solver/issues/detail?id=150) created by sandwichmaker on 2014-08-13T16:15:11.000Z:
<b>What steps will reproduce the problem?</b>
1. Install ceres-solver with MacPorts (either from binary with "port install ceres-solver" or source with "port -s install ceres-solver").
2. Compile the program at http://julianpanetta.com/macports_bugs/test_ceres.cc with /usr/bin/clang++:
/usr/bin/clang++ -g -O0 -std=c++11 -I/opt/local/include
-I/opt/local/include/eigen3 test_ceres.cc -L/opt/local/lib -lceres -lglog
-lgflags -lcholmod -lcxsparse -framework accelerate -o broken
3) ./broken
<b>What is the expected output? What do you see instead?</b>
It should print 0..19, but it intermittently segfaults partway through or throws a std::length error because Ceres has scribbled over the arrays v1s and v2s.
<b>What version of the product are you using? On what operating system?</b>
I've tried with 1.9.0 installed from MacPorts on both OS 10.9.3 and 10.9.4.
<b>Please provide any additional information below.</b>
I've filed a ticket with MacPorts because this only seems to happen with a MacPorts-installed (statically linked) Ceres. The test program runs fine when statically linked against a Ceres version manually built with cmake (both the git head, and the stable 1.9.0 release). However, the MacPorts configuration is so minimal it's unclear where the problem could be. More details are at:
https://trac.macports.org/ticket/44627 | defect | ceres solver from macports segfaults on a particular problem original created by sandwichmaker on what steps will reproduce the problem install ceres solver with macports either from binary with quot port install ceres solver quot or source with quot port s install ceres solver quot compile the program at with usr bin clang usr bin clang g std c i opt local include i opt local include test ceres cc l opt local lib lceres lglog lgflags lcholmod lcxsparse framework accelerate o broken broken what is the expected output what do you see instead it should print but it intermittently segfaults partway through or throws a std length error because ceres has scribbled over the arrays and what version of the product are you using on what operating system i ve tried with installed from macports on both os and please provide any additional information below i ve filed a ticket with macports because this only seems to happen with a macports installed statically linked ceres the test program runs fine when statically linked against a ceres version manually built with cmake both the git head and the stable release however the macports configuration is so minimal it s unclear where the problem could be more details are at | 1 |
100,168 | 16,484,053,055 | IssuesEvent | 2021-05-24 15:25:27 | gdcorp-action-public-forks/keep-a-changelog-new-release | https://api.github.com/repos/gdcorp-action-public-forks/keep-a-changelog-new-release | closed | WS-2020-0042 (High) detected in acorn-6.4.0.tgz - autoclosed | security vulnerability | ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-6.4.0.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.4.0.tgz">https://registry.npmjs.org/acorn/-/acorn-6.4.0.tgz</a></p>
<p>Path to dependency file: keep-a-changelog-new-release/package.json</p>
<p>Path to vulnerable library: keep-a-changelog-new-release/node_modules/acorn</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- espree-5.0.1.tgz
- :x: **acorn-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gdcorp-action-public-forks/keep-a-changelog-new-release/commit/0e201c270deac5c59f9e4ea7726d6ee5225ab235">0e201c270deac5c59f9e4ea7726d6ee5225ab235</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-08</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"acorn","packageVersion":"6.4.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;espree:5.0.1;acorn:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0042","vulnerabilityDetails":"acorn is vulnerable to REGEX DoS. A regex of the form /[x-\\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.","vulnerabilityUrl":"https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2020-0042 (High) detected in acorn-6.4.0.tgz - autoclosed - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-6.4.0.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.4.0.tgz">https://registry.npmjs.org/acorn/-/acorn-6.4.0.tgz</a></p>
<p>Path to dependency file: keep-a-changelog-new-release/package.json</p>
<p>Path to vulnerable library: keep-a-changelog-new-release/node_modules/acorn</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- espree-5.0.1.tgz
- :x: **acorn-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gdcorp-action-public-forks/keep-a-changelog-new-release/commit/0e201c270deac5c59f9e4ea7726d6ee5225ab235">0e201c270deac5c59f9e4ea7726d6ee5225ab235</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-08</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"acorn","packageVersion":"6.4.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;espree:5.0.1;acorn:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0042","vulnerabilityDetails":"acorn is vulnerable to REGEX DoS. A regex of the form /[x-\\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.","vulnerabilityUrl":"https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | ws high detected in acorn tgz autoclosed ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file keep a changelog new release package json path to vulnerable library keep a changelog new release node modules acorn dependency hierarchy eslint tgz root library espree tgz x acorn tgz vulnerable library found in head commit a href found in base branch master vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree eslint espree acorn isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser vulnerabilityurl | 0 |
28,329 | 5,241,282,198 | IssuesEvent | 2017-01-31 15:21:47 | google/gtm-oauth2 | https://api.github.com/repos/google/gtm-oauth2 | closed | Extension Grants | auto-migrated Priority-Medium Type-Defect | ```
Resource Owner Password Credentials Grant is not supported yet.
and in order to comply with the document "rfc6749", especially "4.5. Extension
Grants", provider-extended grant should be supported.
```
Original issue reported on code.google.com by `lee.juyo...@gmail.com` on 13 Feb 2014 at 11:47
| 1.0 | Extension Grants - ```
Resource Owner Password Credentials Grant is not supported yet.
and in order to comply with the document "rfc6749", especially "4.5. Extension
Grants", provider-extended grant should be supported.
```
Original issue reported on code.google.com by `lee.juyo...@gmail.com` on 13 Feb 2014 at 11:47
| defect | extension grants resource owner password credentials grant is not supported yet and in order to comply with the document especially extension grants provider extended grant should be supported original issue reported on code google com by lee juyo gmail com on feb at | 1 |
89,689 | 25,885,347,266 | IssuesEvent | 2022-12-14 14:12:05 | KosmX/minecraftPlayerAnimator | https://api.github.com/repos/KosmX/minecraftPlayerAnimator | closed | Could not get unknown property 'player_anim' for root project 'forge' of type org.gradle.api.Project. | invalid build | I am getting this error:
` Could not get unknown property 'player_anim' for root project 'forge' of type org.gradle.api.Project.`
While running the RunClient task. | 1.0 | Could not get unknown property 'player_anim' for root project 'forge' of type org.gradle.api.Project. - I am getting this error:
` Could not get unknown property 'player_anim' for root project 'forge' of type org.gradle.api.Project.`
While running the RunClient task. | non_defect | could not get unknown property player anim for root project forge of type org gradle api project i am getting this error could not get unknown property player anim for root project forge of type org gradle api project while running the runclient task | 0 |
142,816 | 13,041,518,371 | IssuesEvent | 2020-07-28 20:32:36 | PyTorchLightning/pytorch-lightning | https://api.github.com/repos/PyTorchLightning/pytorch-lightning | closed | How to remove `v_num` from the progress bar ? | documentation good first issue help wanted question | Version: 0.7.3
The `v_num` is automatically added to progress_bar when some logger is used
It is not much a problem for tensorboard when v_num is just a simple number
But v_num for mlfow takes a lot of space
Traning step
```
def training_step(self, batch, batch_nb):
....
log = { "trn_loss": 0.1, "lr": 0.001 }
return {"loss": loss, "log": log, "progress_bar": log}
```
Progress bar when using with mlflow logger
```
[00:33<00:46, 1.62s/it, loss=0.740, lr=8e-6, trn_loss=0.659, v_num=18_28bc973b1f0e42e8b4d664d1ef7812f6]
```
Also, `loss` is automatically added to `progress_bar`
| 1.0 | How to remove `v_num` from the progress bar ? - Version: 0.7.3
The `v_num` is automatically added to progress_bar when some logger is used
It is not much a problem for tensorboard when v_num is just a simple number
But v_num for mlfow takes a lot of space
Traning step
```
def training_step(self, batch, batch_nb):
....
log = { "trn_loss": 0.1, "lr": 0.001 }
return {"loss": loss, "log": log, "progress_bar": log}
```
Progress bar when using with mlflow logger
```
[00:33<00:46, 1.62s/it, loss=0.740, lr=8e-6, trn_loss=0.659, v_num=18_28bc973b1f0e42e8b4d664d1ef7812f6]
```
Also, `loss` is automatically added to `progress_bar`
| non_defect | how to remove v num from the progress bar version the v num is automatically added to progress bar when some logger is used it is not much a problem for tensorboard when v num is just a simple number but v num for mlfow takes a lot of space traning step def training step self batch batch nb log trn loss lr return loss loss log log progress bar log progress bar when using with mlflow logger also loss is automatically added to progress bar | 0 |
55,778 | 3,074,322,619 | IssuesEvent | 2015-08-20 06:15:05 | bevanhunt/meteor-leaflet | https://api.github.com/repos/bevanhunt/meteor-leaflet | closed | upgrade leaflet to 1.0.0-beta1 | blocked feature priority | Huge performance improvements - contains breaking changes - so should up this package to 1.0.0 - must test locally with all other included libraries. | 1.0 | upgrade leaflet to 1.0.0-beta1 - Huge performance improvements - contains breaking changes - so should up this package to 1.0.0 - must test locally with all other included libraries. | non_defect | upgrade leaflet to huge performance improvements contains breaking changes so should up this package to must test locally with all other included libraries | 0 |
274,680 | 30,113,518,907 | IssuesEvent | 2023-06-30 09:37:03 | core-coin/web-coreblockchain | https://api.github.com/repos/core-coin/web-coreblockchain | opened | Dependency check of Core Blockchain (web-coreblockchain-master ) Report-1 | bug security | this is the first report of the dependency check of different libraries which is used in Blockchain (Eweb-coreblockchain-master package) via OWASP dependency Checker and SonarQube application. According to test via these tools, there are some dependencies which are needed to update as soon as possible (Although there are some false positives). in the report, all the dependencies and their related CVE vulnerabilities with additional information are provided. Please see the attachment report and if you have any question feel free to ask me.
[dependency-check-report.html](/ping/exchange/uploads/a51d1b76f19da0f568ef67632bcd361a/dependency-check-report.html)this is the first report of the dependency check of different libraries which is used in Blockchain (Eweb-coreblockchain-master package) via OWASP dependency Checker and SonarQube application. According to test via these tools, there are some dependencies which are needed to update as soon as possible (Although there are some false positives). in the report, all the dependencies and their related CVE vulnerabilities with additional information are provided. Please see the attachment report and if you have any question feel free to ask me.
[dependency-check-report.html](https://github.com/ping/exchange/uploads/a51d1b76f19da0f568ef67632bcd361a/dependency-check-report.html)
[dependency-check-report.zip](https://github.com/core-coin/web-coreblockchain/files/11915110/dependency-check-report.zip)
this is an old issue related to the CoreBlockchain which was reported in exchange Gitlab repository by mistake one year ago!
@rastislavcore | True | Dependency check of Core Blockchain (web-coreblockchain-master ) Report-1 - this is the first report of the dependency check of different libraries which is used in Blockchain (Eweb-coreblockchain-master package) via OWASP dependency Checker and SonarQube application. According to test via these tools, there are some dependencies which are needed to update as soon as possible (Although there are some false positives). in the report, all the dependencies and their related CVE vulnerabilities with additional information are provided. Please see the attachment report and if you have any question feel free to ask me.
[dependency-check-report.html](/ping/exchange/uploads/a51d1b76f19da0f568ef67632bcd361a/dependency-check-report.html)this is the first report of the dependency check of different libraries which is used in Blockchain (Eweb-coreblockchain-master package) via OWASP dependency Checker and SonarQube application. According to test via these tools, there are some dependencies which are needed to update as soon as possible (Although there are some false positives). in the report, all the dependencies and their related CVE vulnerabilities with additional information are provided. Please see the attachment report and if you have any question feel free to ask me.
[dependency-check-report.html](https://github.com/ping/exchange/uploads/a51d1b76f19da0f568ef67632bcd361a/dependency-check-report.html)
[dependency-check-report.zip](https://github.com/core-coin/web-coreblockchain/files/11915110/dependency-check-report.zip)
this is an old issue related to the CoreBlockchain which was reported in exchange Gitlab repository by mistake one year ago!
@rastislavcore | non_defect | dependency check of core blockchain web coreblockchain master report this is the first report of the dependency check of different libraries which is used in blockchain eweb coreblockchain master package via owasp dependency checker and sonarqube application according to test via these tools there are some dependencies which are needed to update as soon as possible although there are some false positives in the report all the dependencies and their related cve vulnerabilities with additional information are provided please see the attachment report and if you have any question feel free to ask me ping exchange uploads dependency check report html this is the first report of the dependency check of different libraries which is used in blockchain eweb coreblockchain master package via owasp dependency checker and sonarqube application according to test via these tools there are some dependencies which are needed to update as soon as possible although there are some false positives in the report all the dependencies and their related cve vulnerabilities with additional information are provided please see the attachment report and if you have any question feel free to ask me this is an old issue related to the coreblockchain which was reported in exchange gitlab repository by mistake one year ago rastislavcore | 0 |
26,106 | 4,593,599,671 | IssuesEvent | 2016-09-21 01:57:20 | afisher1/GridLAB-D | https://api.github.com/repos/afisher1/GridLAB-D | closed | #18 Scalar in unit definition are ignored,
| defect | This is unfortunate because it makes it necessary to define each scale type (kW, MW) explicitly. Anything not defined explicity, but reasonably expected isn't going to work.
,
| 1.0 | #18 Scalar in unit definition are ignored,
- This is unfortunate because it makes it necessary to define each scale type (kW, MW) explicitly. Anything not defined explicity, but reasonably expected isn't going to work.
,
| defect | scalar in unit definition are ignored this is unfortunate because it makes it necessary to define each scale type kw mw explicitly anything not defined explicity but reasonably expected isn t going to work | 1 |
219,933 | 16,856,780,874 | IssuesEvent | 2021-06-21 07:53:15 | microsoftgraph/microsoft-graph-toolkit | https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit | opened | [Docs] Explain how to setup event handlers for events emitted by MGT components | Area: Documentation | ## Description
We should extend the docs with an explanation how to setup event handlers for events emitted by MGT components. Right now, we only mention events in the context of adding event handlers to events inside templates. For discoverability, we should also consider linking to this new page from the events section in all components so that developers new to MGT can quickly navigate to instructions how to setup event handlers.
I'll pick this up. | 1.0 | [Docs] Explain how to setup event handlers for events emitted by MGT components - ## Description
We should extend the docs with an explanation how to setup event handlers for events emitted by MGT components. Right now, we only mention events in the context of adding event handlers to events inside templates. For discoverability, we should also consider linking to this new page from the events section in all components so that developers new to MGT can quickly navigate to instructions how to setup event handlers.
I'll pick this up. | non_defect | explain how to setup event handlers for events emitted by mgt components description we should extend the docs with an explanation how to setup event handlers for events emitted by mgt components right now we only mention events in the context of adding event handlers to events inside templates for discoverability we should also consider linking to this new page from the events section in all components so that developers new to mgt can quickly navigate to instructions how to setup event handlers i ll pick this up | 0 |
268,451 | 8,407,040,358 | IssuesEvent | 2018-10-11 19:43:49 | python/mypy | https://api.github.com/repos/python/mypy | opened | Allow setting final instance attributes in __new__ | feature priority-1-normal | These patterns should be allowed:
```python
class C:
def __new__(cls, x: int) -> C:
self = object.__new__(cls)
self.x: Final = x
return self
```
and
```python
class C:
x: Final[int]
def __new__(cls) -> C:
self = object.__new__(cls)
self.x: Final = 1
return self
```
Currently this is blocked by https://github.com/python/mypy/issues/1021 | 1.0 | Allow setting final instance attributes in __new__ - These patterns should be allowed:
```python
class C:
def __new__(cls, x: int) -> C:
self = object.__new__(cls)
self.x: Final = x
return self
```
and
```python
class C:
x: Final[int]
def __new__(cls) -> C:
self = object.__new__(cls)
self.x: Final = 1
return self
```
Currently this is blocked by https://github.com/python/mypy/issues/1021 | non_defect | allow setting final instance attributes in new these patterns should be allowed python class c def new cls x int c self object new cls self x final x return self and python class c x final def new cls c self object new cls self x final return self currently this is blocked by | 0 |
48,184 | 13,067,500,363 | IssuesEvent | 2020-07-31 00:39:38 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | steamshovel - python bits still explicitly import PyQt4 (Trac #1919) | Migrated from Trac combo core defect | fix that
Migrated from https://code.icecube.wisc.edu/ticket/1919
```json
{
"status": "closed",
"changetime": "2017-10-03T13:07:30",
"description": "fix that",
"reporter": "nega",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1507036050170209",
"component": "combo core",
"summary": "steamshovel - python bits still explicitly import PyQt4",
"priority": "normal",
"keywords": "python pyqt qt5",
"time": "2016-12-05T21:58:34",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | steamshovel - python bits still explicitly import PyQt4 (Trac #1919) - fix that
Migrated from https://code.icecube.wisc.edu/ticket/1919
```json
{
"status": "closed",
"changetime": "2017-10-03T13:07:30",
"description": "fix that",
"reporter": "nega",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1507036050170209",
"component": "combo core",
"summary": "steamshovel - python bits still explicitly import PyQt4",
"priority": "normal",
"keywords": "python pyqt qt5",
"time": "2016-12-05T21:58:34",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
| defect | steamshovel python bits still explicitly import trac fix that migrated from json status closed changetime description fix that reporter nega cc david schultz resolution fixed ts component combo core summary steamshovel python bits still explicitly import priority normal keywords python pyqt time milestone long term future owner nega type defect | 1 |
113,158 | 4,543,934,278 | IssuesEvent | 2016-09-10 11:50:06 | thommoboy/There-are-no-brakes | https://api.github.com/repos/thommoboy/There-are-no-brakes | closed | textboxes have no priority, just goes by whoever entered an area last | bug Priority Medium Tutorial | particularly annoying when, for example, the players need to see info on the pyramid door but are instead seeing info on showing the icons | 1.0 | textboxes have no priority, just goes by whoever entered an area last - particularly annoying when, for example, the players need to see info on the pyramid door but are instead seeing info on showing the icons | non_defect | textboxes have no priority just goes by whoever entered an area last particularly annoying when for example the players need to see info on the pyramid door but are instead seeing info on showing the icons | 0 |
388,285 | 11,485,855,886 | IssuesEvent | 2020-02-11 08:44:12 | DigitalCampus/moodle-block_oppia_mobile_export | https://api.github.com/repos/DigitalCampus/moodle-block_oppia_mobile_export | closed | Create 'reference' course(s) | enhancement high priority | Specifically courses that show all the different functionality that we can use for testing purposes | 1.0 | Create 'reference' course(s) - Specifically courses that show all the different functionality that we can use for testing purposes | non_defect | create reference course s specifically courses that show all the different functionality that we can use for testing purposes | 0 |
23,182 | 3,774,640,780 | IssuesEvent | 2016-03-17 10:09:12 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | opened | No cast for elements when creating Dictionary | defect | ### Expected
Type of an item added to Dictionary is `Bridge.Decimal`
```
var decimalDict = Bridge.merge(new Bridge.Dictionary$2(Bridge.Int,Bridge.Decimal)(), [
[0, new Bridge.Decimal(5)]
] );
```
### Actual
Type of an item added to Dictionary is `Number`
```
var decimalDict = Bridge.merge(new Bridge.Dictionary$2(Bridge.Int,Bridge.Decimal)(), [
[0, 5]
] );
```
### Steps To Reproduce
http://live.bridge.net/#94bf6277e5445dd1aa2a
```csharp
public class App
{
[Ready]
public static void Main()
{
var decimalDict = new Dictionary<int, decimal> { { 0, 5 } };
Console.Log(decimalDict[0].GetClassName());
}
}
``` | 1.0 | No cast for elements when creating Dictionary - ### Expected
Type of an item added to Dictionary is `Bridge.Decimal`
```
var decimalDict = Bridge.merge(new Bridge.Dictionary$2(Bridge.Int,Bridge.Decimal)(), [
[0, new Bridge.Decimal(5)]
] );
```
### Actual
Type of an item added to Dictionary is `Number`
```
var decimalDict = Bridge.merge(new Bridge.Dictionary$2(Bridge.Int,Bridge.Decimal)(), [
[0, 5]
] );
```
### Steps To Reproduce
http://live.bridge.net/#94bf6277e5445dd1aa2a
```csharp
public class App
{
[Ready]
public static void Main()
{
var decimalDict = new Dictionary<int, decimal> { { 0, 5 } };
Console.Log(decimalDict[0].GetClassName());
}
}
``` | defect | no cast for elements when creating dictionary expected type of an item added to dictionary is bridge decimal var decimaldict bridge merge new bridge dictionary bridge int bridge decimal actual type of an item added to dictionary is number var decimaldict bridge merge new bridge dictionary bridge int bridge decimal steps to reproduce csharp public class app public static void main var decimaldict new dictionary console log decimaldict getclassname | 1 |
7,139 | 2,610,328,914 | IssuesEvent | 2015-02-26 19:45:59 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Balancing | auto-migrated Priority-Medium Type-Defect | ```
Will there be any big fixes in balancing of the units?
- a lot of ships have way too much hardpoints (i.e. Deadnaught)
- some ships are way too weak (CIS/Republic Corvettes)
- Fighters and Bombers have a lot of health
```
-----
Original issue reported on code.google.com by `tds...@spamavert.com` on 12 Jun 2011 at 4:25 | 1.0 | Balancing - ```
Will there be any big fixes in balancing of the units?
- a lot of ships have way too much hardpoints (i.e. Deadnaught)
- some ships are way too weak (CIS/Republic Corvettes)
- Fighters and Bombers have a lot of health
```
-----
Original issue reported on code.google.com by `tds...@spamavert.com` on 12 Jun 2011 at 4:25 | defect | balancing will there be any big fixes in balancing of the units a lot of ships have way too much hardpoints i e deadnaught some ships are way too weak cis republic corvettes fighters and bombers have a lot of health original issue reported on code google com by tds spamavert com on jun at | 1 |
26,446 | 4,711,125,780 | IssuesEvent | 2016-10-14 12:51:20 | kronometrix/recording | https://api.github.com/repos/kronometrix/recording | opened | wsrec does not have a release version | defect-high | wsrec does not honour the release version.
```
krmx@raspberrypi:~/workspace/kronometrix/recording/bin $ ./wsrec -V
Unknown option: V
USAGE: wsrec [-dp] | [interval [count]]
eg, wsrec # print human readable output
wsrec 5 # print every 5 seconds
wsrec 1 5 # print 5 times, every 1 second
wsrec -p 60 # print every 60 seconds parseable output
FIELDS:
#01 timestamp : seconds since Epoch, time
#02 ih : indoor humidity, number
#03 oh : outdoor humidity,, number
#04 ita : indoor air temperature Celsius, number
#05 ota : outdoor air temperature Celsius, number
#06 dpa : outdoor dew point air temperature Celsius , number
#07 wca : outdoor wind chill temperature, Celsius, number
#08 wind : wind speed m/s, rate
#09 wgust : wind gust m/s, rate
#10 windir : wind Direction
#11 rain1 : rain 1 hours, number
#12 rain24 : rain 24 hours, number
#13 totalrain : total rain, number
#14 pres : absolute air pressure, number
NOTES: Retrieves and logs the weather stats from the following weather
stations: WH-1080/1081
```
| 1.0 | wsrec does not have a release version - wsrec does not honour the release version.
```
krmx@raspberrypi:~/workspace/kronometrix/recording/bin $ ./wsrec -V
Unknown option: V
USAGE: wsrec [-dp] | [interval [count]]
eg, wsrec # print human readable output
wsrec 5 # print every 5 seconds
wsrec 1 5 # print 5 times, every 1 second
wsrec -p 60 # print every 60 seconds parseable output
FIELDS:
#01 timestamp : seconds since Epoch, time
#02 ih : indoor humidity, number
#03 oh : outdoor humidity,, number
#04 ita : indoor air temperature Celsius, number
#05 ota : outdoor air temperature Celsius, number
#06 dpa : outdoor dew point air temperature Celsius , number
#07 wca : outdoor wind chill temperature, Celsius, number
#08 wind : wind speed m/s, rate
#09 wgust : wind gust m/s, rate
#10 windir : wind Direction
#11 rain1 : rain 1 hours, number
#12 rain24 : rain 24 hours, number
#13 totalrain : total rain, number
#14 pres : absolute air pressure, number
NOTES: Retrieves and logs the weather stats from the following weather
stations: WH-1080/1081
```
| defect | wsrec does not have a release version wsrec does not honour the release version krmx raspberrypi workspace kronometrix recording bin wsrec v unknown option v usage wsrec eg wsrec print human readable output wsrec print every seconds wsrec print times every second wsrec p print every seconds parseable output fields timestamp seconds since epoch time ih indoor humidity number oh outdoor humidity number ita indoor air temperature celsius number ota outdoor air temperature celsius number dpa outdoor dew point air temperature celsius number wca outdoor wind chill temperature celsius number wind wind speed m s rate wgust wind gust m s rate windir wind direction rain hours number rain hours number totalrain total rain number pres absolute air pressure number notes retrieves and logs the weather stats from the following weather stations wh | 1 |
9,867 | 2,616,004,736 | IssuesEvent | 2015-03-02 00:49:20 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | Terran Vulture Crash | auto-migrated Maintainability Priority-Critical Type-Defect Usability | ```
Game crashes when a selected Terran Vulture containing spider mines has
been killed.
```
Original issue reported on code.google.com by `AHeinerm` on 21 Sep 2008 at 2:54 | 1.0 | Terran Vulture Crash - ```
Game crashes when a selected Terran Vulture containing spider mines has
been killed.
```
Original issue reported on code.google.com by `AHeinerm` on 21 Sep 2008 at 2:54 | defect | terran vulture crash game crashes when a selected terran vulture containing spider mines has been killed original issue reported on code google com by aheinerm on sep at | 1 |
68,501 | 21,665,146,075 | IssuesEvent | 2022-05-07 03:48:48 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [๐ Bug]: drag_and_drop can't act will with canvas | I-defect needs-triaging | ### What happened?
for example with http://literallycanvas.com/:
i want to draw a line in it ,i tried `ActionChains(driver).drag_and_drop_by_offset(canvas,50,50).perform()`,but only draw a point
then i tried `ActionChains(driver).click_and_hold(canvas).move_by_offset(50, 0).release().perform()`,also draw a point
finally i draw a line with `ActionChains(driver).click_and_hold(canvas).move_by_offset(50, 0).move_by_offset(0, 0).release().perform()`
but it's not reasonable
### How can we reproduce the issue?
```shell
it seems like a holdingmouse_move action require something like 'Intermediate state',which drag_and_drop didn't provide
```
### Relevant log output
```shell
None
```
### Operating System
Window10
### Selenium version
python4.1.3
### What are the browser(s) and version(s) where you see this issue?
chrome 101
### What are the browser driver(s) and version(s) where you see this issue?
chrome 101
### Are you using Selenium Grid?
no | 1.0 | [๐ Bug]: drag_and_drop can't act will with canvas - ### What happened?
for example with http://literallycanvas.com/:
i want to draw a line in it ,i tried `ActionChains(driver).drag_and_drop_by_offset(canvas,50,50).perform()`,but only draw a point
then i tried `ActionChains(driver).click_and_hold(canvas).move_by_offset(50, 0).release().perform()`,also draw a point
finally i draw a line with `ActionChains(driver).click_and_hold(canvas).move_by_offset(50, 0).move_by_offset(0, 0).release().perform()`
but it's not reasonable
### How can we reproduce the issue?
```shell
it seems like a holdingmouse_move action require something like 'Intermediate state',which drag_and_drop didn't provide
```
### Relevant log output
```shell
None
```
### Operating System
Window10
### Selenium version
python4.1.3
### What are the browser(s) and version(s) where you see this issue?
chrome 101
### What are the browser driver(s) and version(s) where you see this issue?
chrome 101
### Are you using Selenium Grid?
no | defect | drag and drop can t act will with canvas what happened for example with i want to draw a line in it i tried actionchains driver drag and drop by offset canvas perform but only draw a point then i tried actionchains driver click and hold canvas move by offset release perform also draw a point finally i draw a line with actionchains driver click and hold canvas move by offset move by offset release perform but it s not reasonable how can we reproduce the issue shell it seems like a holdingmouse move action require something like intermediate state which drag and drop didn t provide relevant log output shell none operating system selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chrome are you using selenium grid no | 1 |
18,424 | 3,059,715,805 | IssuesEvent | 2015-08-14 16:27:40 | PyTables/PyTables | https://api.github.com/repos/PyTables/PyTables | closed | Some tests require UTF-8 locale | defect | When we build PyTables for Ubuntu, the build machines by definition use a C/POSIX locale. This breaks the following tests, all of which error with UnicodeEncodeErrors:
ERROR: test01 (tables.tests.test_basics.UnicodeFilename)
Checking creating a filename with Unicode chars.
ERROR: test02 (tables.tests.test_basics.UnicodeFilename)
Checking tables.is_hdf5_file with a Unicode filename.
ERROR: test03 (tables.tests.test_basics.UnicodeFilename)
Checking is_pytables_file with a Unicode filename.
ERROR: test_store_and_load_with_non_ascii_attributes (tables.tests.test_basics.UnicodeFilename)
Forcing the test suite to run under say en_US.UTF-8 allows the tests to pass. | 1.0 | Some tests require UTF-8 locale - When we build PyTables for Ubuntu, the build machines by definition use a C/POSIX locale. This breaks the following tests, all of which error with UnicodeEncodeErrors:
ERROR: test01 (tables.tests.test_basics.UnicodeFilename)
Checking creating a filename with Unicode chars.
ERROR: test02 (tables.tests.test_basics.UnicodeFilename)
Checking tables.is_hdf5_file with a Unicode filename.
ERROR: test03 (tables.tests.test_basics.UnicodeFilename)
Checking is_pytables_file with a Unicode filename.
ERROR: test_store_and_load_with_non_ascii_attributes (tables.tests.test_basics.UnicodeFilename)
Forcing the test suite to run under say en_US.UTF-8 allows the tests to pass. | defect | some tests require utf locale when we build pytables for ubuntu the build machines by definition use a c posix locale this breaks the following tests all of which error with unicodeencodeerrors error tables tests test basics unicodefilename checking creating a filename with unicode chars error tables tests test basics unicodefilename checking tables is file with a unicode filename error tables tests test basics unicodefilename checking is pytables file with a unicode filename error test store and load with non ascii attributes tables tests test basics unicodefilename forcing the test suite to run under say en us utf allows the tests to pass | 1 |
349,335 | 24,942,595,868 | IssuesEvent | 2022-10-31 20:19:10 | bounswe/bounswe2022group9 | https://api.github.com/repos/bounswe/bounswe2022group9 | closed | Revising Requirements of the Project | Documentation Priority: Medium Completed | We have discussed requirements set for Cmpe352 in our last weekly meeting and decided to update some of them, add new ones and delete some of them which we think out of our project's scope. We should update Requirements in wiki page accordingly. | 1.0 | Revising Requirements of the Project - We have discussed requirements set for Cmpe352 in our last weekly meeting and decided to update some of them, add new ones and delete some of them which we think out of our project's scope. We should update Requirements in wiki page accordingly. | non_defect | revising requirements of the project we have discussed requirements set for in our last weekly meeting and decided to update some of them add new ones and delete some of them which we think out of our project s scope we should update requirements in wiki page accordingly | 0 |
229,435 | 7,574,695,129 | IssuesEvent | 2018-04-23 21:52:05 | rzwitserloot/lombok | https://api.github.com/repos/rzwitserloot/lombok | closed | onX support for constructor argument annotations | low-priority | Feature request to be able to include annotations on generated constructors for ```@RequiredArgsConstructor``` and/or ```@AllAgsConstructor```
Right now it is possible to add an annotation to the constructor itself via:
```java
@AllArgsConstructor(onConstructor=@__(@Inject))
```
but to add annotations to the arguments does not seem to be possible. This is supported via ```@Setter``` annotation, at least I think, though this related issue hinting maybe the documentation is not quite accurate
https://github.com/rzwitserloot/lombok/issues/1178
I think this could work either by adding an annotation to the fields themselves or some new syntax added to the AllArgsConstructor / RequiredArgsConstructor annotation.
The use case is being able to generate constructors for DI-managed beans that need the arguments to contain additional qualifier annotations to inject the proper bean. | 1.0 | onX support for constructor argument annotations - Feature request to be able to include annotations on generated constructors for ```@RequiredArgsConstructor``` and/or ```@AllAgsConstructor```
Right now it is possible to add an annotation to the constructor itself via:
```java
@AllArgsConstructor(onConstructor=@__(@Inject))
```
but to add annotations to the arguments does not seem to be possible. This is supported via ```@Setter``` annotation, at least I think, though this related issue hinting maybe the documentation is not quite accurate
https://github.com/rzwitserloot/lombok/issues/1178
I think this could work either by adding an annotation to the fields themselves or some new syntax added to the AllArgsConstructor / RequiredArgsConstructor annotation.
The use case is being able to generate constructors for DI-managed beans that need the arguments to contain additional qualifier annotations to inject the proper bean. | non_defect | onx support for constructor argument annotations feature request to be able to include annotations on generated constructors for requiredargsconstructor and or allagsconstructor right now it is possible to add an annotation to the constructor itself via java allargsconstructor onconstructor inject but to add annotations to the arguments does not seem to be possible this is supported via setter annotation at least i think though this related issue hinting maybe the documentation is not quite accurate i think this could work either by adding an annotation to the fields themselves or some new syntax added to the allargsconstructor requiredargsconstructor annotation the use case is being able to generate constructors for di managed beans that need the arguments to contain additional qualifier annotations to inject the proper bean | 0 |
425,422 | 29,478,059,755 | IssuesEvent | 2023-06-02 01:15:39 | drjdlarson/gncpy | https://api.github.com/repos/drjdlarson/gncpy | opened | Version number on published documentation | documentation enhancement | The version number on the published documentation uses the guessed value from setuptools_scm instead of matching the tag directly (i.e. it shows 1.0.2.dev0 instead of showing 1.0.1). | 1.0 | Version number on published documentation - The version number on the published documentation uses the guessed value from setuptools_scm instead of matching the tag directly (i.e. it shows 1.0.2.dev0 instead of showing 1.0.1). | non_defect | version number on published documentation the version number on the published documentation uses the guessed value from setuptools scm instead of matching the tag directly i e it shows instead of showing | 0 |
51,088 | 13,188,096,422 | IssuesEvent | 2020-08-13 05:32:44 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | [filterscripts] Exception if no best Fit was found (Trac #1950) | Migrated from Trac combo reconstruction defect | Hello,
I am encountering this:
```text
ERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())
ERROR (I3Module): OnlineL2_findBestSucceededFit: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/data/user/flauber/scripts/executables/SimulationFiltering.py", line 412, in <module>
main(opts)
File "/data/user/flauber/scripts/executables/SimulationFiltering.py", line 371, in main
tray.Execute()
File "/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py", line 234, in Execute
super(I3Tray, self).Execute()
File "/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/onlinel2filter.py", line 570, in findBestSucceededFit
frame.Put(name+"_BestFitName", "None")
Boost.Python.ArgumentError: Python argument types in
I3Frame.Put(I3Frame, str, str)
did not match C++ signature:
Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>, I3Frame::Stream)
Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>)
```
while running SimulationFiltering.py from filterscripts/resources/scripts
I guess, the "None" needs to be wrapped to become an dataclasses.I3String or similar.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1950">https://code.icecube.wisc.edu/ticket/1950</a>, reported by flauber and owned by thomas.kintscher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-02-20T09:08:11",
"description": "Hello,\n\nI am encountering this:\n{{{\nERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())\nERROR (I3Module): OnlineL2_findBestSucceededFit: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/flauber/scripts/executables/SimulationFiltering.py\", line 412, in <module>\n main(opts)\n File \"/data/user/flauber/scripts/executables/SimulationFiltering.py\", line 371, in main\n tray.Execute()\n File \"/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/onlinel2filter.py\", line 570, in findBestSucceededFit\n frame.Put(name+\"_BestFitName\", \"None\")\nBoost.Python.ArgumentError: Python argument types in\n I3Frame.Put(I3Frame, str, str)\ndid not match C++ signature:\n Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>, I3Frame::Stream)\n Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>)\n}}}\n\nwhile running SimulationFiltering.py from filterscripts/resources/scripts\n\nI guess, the \"None\" needs to be wrapped to become an dataclasses.I3String or similar.\n",
"reporter": "flauber",
"cc": "",
"resolution": "fixed",
"_ts": "1487581691387840",
"component": "combo reconstruction",
"summary": "[filterscripts] Exception if no best Fit was found",
"priority": "normal",
"keywords": "filterscripts",
"time": "2017-02-20T08:51:22",
"milestone": "",
"owner": "thomas.kintscher",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [filterscripts] Exception if no best Fit was found (Trac #1950) - Hello,
I am encountering this:
```text
ERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())
ERROR (I3Module): OnlineL2_findBestSucceededFit: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/data/user/flauber/scripts/executables/SimulationFiltering.py", line 412, in <module>
main(opts)
File "/data/user/flauber/scripts/executables/SimulationFiltering.py", line 371, in main
tray.Execute()
File "/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py", line 234, in Execute
super(I3Tray, self).Execute()
File "/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/onlinel2filter.py", line 570, in findBestSucceededFit
frame.Put(name+"_BestFitName", "None")
Boost.Python.ArgumentError: Python argument types in
I3Frame.Put(I3Frame, str, str)
did not match C++ signature:
Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>, I3Frame::Stream)
Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>)
```
while running SimulationFiltering.py from filterscripts/resources/scripts
I guess, the "None" needs to be wrapped to become an dataclasses.I3String or similar.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1950">https://code.icecube.wisc.edu/ticket/1950</a>, reported by flauber and owned by thomas.kintscher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-02-20T09:08:11",
"description": "Hello,\n\nI am encountering this:\n{{{\nERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())\nERROR (I3Module): OnlineL2_findBestSucceededFit: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/flauber/scripts/executables/SimulationFiltering.py\", line 412, in <module>\n main(opts)\n File \"/data/user/flauber/scripts/executables/SimulationFiltering.py\", line 371, in main\n tray.Execute()\n File \"/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/mnt/lfs3/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/onlinel2filter.py\", line 570, in findBestSucceededFit\n frame.Put(name+\"_BestFitName\", \"None\")\nBoost.Python.ArgumentError: Python argument types in\n I3Frame.Put(I3Frame, str, str)\ndid not match C++ signature:\n Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>, I3Frame::Stream)\n Put(I3Frame {lvalue}, std::string, boost::shared_ptr<I3FrameObject const>)\n}}}\n\nwhile running SimulationFiltering.py from filterscripts/resources/scripts\n\nI guess, the \"None\" needs to be wrapped to become an dataclasses.I3String or similar.\n",
"reporter": "flauber",
"cc": "",
"resolution": "fixed",
"_ts": "1487581691387840",
"component": "combo reconstruction",
"summary": "[filterscripts] Exception if no best Fit was found",
"priority": "normal",
"keywords": "filterscripts",
"time": "2017-02-20T08:51:22",
"milestone": "",
"owner": "thomas.kintscher",
"type": "defect"
}
```
</p>
</details>
| defect | exception if no best fit was found trac hello i am encountering this text error pythonfunction error running python function as module pythonfunction cxx in virtual void pythonfunction process error findbestsucceededfit exception thrown cxx in void do void traceback most recent call last file data user flauber scripts executables simulationfiltering py line in main opts file data user flauber scripts executables simulationfiltering py line in main tray execute file mnt user flauber software icerec build lib py line in execute super self execute file mnt user flauber software icerec build lib icecube filterscripts py line in findbestsucceededfit frame put name bestfitname none boost python argumenterror python argument types in put str str did not match c signature put lvalue std string boost shared ptr stream put lvalue std string boost shared ptr while running simulationfiltering py from filterscripts resources scripts i guess the none needs to be wrapped to become an dataclasses or similar migrated from json status closed changetime description hello n ni am encountering this n nerror pythonfunction error running python function as module pythonfunction cxx in virtual void pythonfunction process nerror findbestsucceededfit exception thrown cxx in void do void ntraceback most recent call last n file data user flauber scripts executables simulationfiltering py line in n main opts n file data user flauber scripts executables simulationfiltering py line in main n tray execute n file mnt user flauber software icerec build lib py line in execute n super self execute n file mnt user flauber software icerec build lib icecube filterscripts py line in findbestsucceededfit n frame put name bestfitname none nboost python argumenterror python argument types in n put str str ndid not match c signature n put lvalue std string boost shared ptr stream n put lvalue std string boost shared ptr n n nwhile running simulationfiltering py from filterscripts resources scripts n ni guess the none needs to be wrapped to become an dataclasses or similar n reporter flauber cc resolution fixed ts component combo reconstruction summary exception if no best fit was found priority normal keywords filterscripts time milestone owner thomas kintscher type defect | 1 |
47,837 | 13,066,266,246 | IssuesEvent | 2020-07-30 21:20:11 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | DomTools documentation (Trac #1244) | Migrated from Trac combo reconstruction defect | no documentation other than release notes, no maintainer listed
Migrated from https://code.icecube.wisc.edu/ticket/1244
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "no documentation other than release notes, no maintainer listed",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "DomTools documentation",
"priority": "blocker",
"keywords": "",
"time": "2015-08-20T08:49:36",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | DomTools documentation (Trac #1244) - no documentation other than release notes, no maintainer listed
Migrated from https://code.icecube.wisc.edu/ticket/1244
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "no documentation other than release notes, no maintainer listed",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "DomTools documentation",
"priority": "blocker",
"keywords": "",
"time": "2015-08-20T08:49:36",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| defect | domtools documentation trac no documentation other than release notes no maintainer listed migrated from json status closed changetime description no documentation other than release notes no maintainer listed reporter kjmeagher cc resolution fixed ts component combo reconstruction summary domtools documentation priority blocker keywords time milestone owner olivas type defect | 1 |
68,850 | 21,927,270,495 | IssuesEvent | 2022-05-23 06:18:30 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | False is not Empty check in whereInList() | defect ORM | ### Description
Input:
```php
$this->Model->whereInList ( 'id', [], ['allowEmpty' => true])
```
Expected Output:
```php
[]
```
Output currently:
`Error: Impossible to generate conditions with empty list of values for field (id)`
Expected Solution:
This:
```php
if ($options['allowEmpty'] && !$values)
```
https://github.com/cakephp/cakephp/blob/4.x/src/Database/Query.php#L1085
Could be:
```php
if ($options['allowEmpty'] && empty($values))
```
### CakePHP Version
4.3
### PHP Version
8.1 | 1.0 | False is not Empty check in whereInList() - ### Description
Input:
```php
$this->Model->whereInList ( 'id', [], ['allowEmpty' => true])
```
Expected Output:
```php
[]
```
Output currently:
`Error: Impossible to generate conditions with empty list of values for field (id)`
Expected Solution:
This:
```php
if ($options['allowEmpty'] && !$values)
```
https://github.com/cakephp/cakephp/blob/4.x/src/Database/Query.php#L1085
Could be:
```php
if ($options['allowEmpty'] && empty($values))
```
### CakePHP Version
4.3
### PHP Version
8.1 | defect | false is not empty check in whereinlist description input php this model whereinlist id expected output php output currently error impossible to generate conditions with empty list of values for field id expected solution this php if options values could be php if options empty values cakephp version php version | 1 |
746,913 | 26,050,677,813 | IssuesEvent | 2022-12-22 18:19:28 | kubernetes/ingress-nginx | https://api.github.com/repos/kubernetes/ingress-nginx | closed | avoid builds and tests for non-code changes | kind/bug priority/important-longterm triage/accepted | - Problem to be solved is that builds & tests are fired, even if a PR changes only documentation or examples that do not change the controller
- Look at closed PRs that only changed documentation under /docs or /deploy. You will find that the entire build & test CI was executed
- The task is look for GHActions feature/syntax that excludes project content like /docs or /deploy from the build & test CI pipeline
/triage accepted
/priority important-longterm
cc @SaumyaBhushan | 1.0 | avoid builds and tests for non-code changes - - Problem to be solved is that builds & tests are fired, even if a PR changes only documentation or examples that do not change the controller
- Look at closed PRs that only changed documentation under /docs or /deploy. You will find that the entire build & test CI was executed
- The task is look for GHActions feature/syntax that excludes project content like /docs or /deploy from the build & test CI pipeline
/triage accepted
/priority important-longterm
cc @SaumyaBhushan | non_defect | avoid builds and tests for non code changes problem to be solved is that builds tests are fired even if a pr changes only documentation or examples that do not change the controller look at closed prs that only changed documentation under docs or deploy you will find that the entire build test ci was executed the task is look for ghactions feature syntax that excludes project content like docs or deploy from the build test ci pipeline triage accepted priority important longterm cc saumyabhushan | 0 |
42,237 | 10,914,478,991 | IssuesEvent | 2019-11-21 09:16:48 | colour-science/colour | https://api.github.com/repos/colour-science/colour | closed | Fix incorrect "Krystek (1985)" implementation. | API Defect Minor | Nasty bug induced by not looking carefully enough at the reference: https://github.com/colour-science/colour/pull/298#issuecomment-555548579. Sadly, even with incorrect values, it produces meaningful results...
| 1.0 | Fix incorrect "Krystek (1985)" implementation. - Nasty bug induced by not looking carefully enough at the reference: https://github.com/colour-science/colour/pull/298#issuecomment-555548579. Sadly, even with incorrect values, it produces meaningful results...
| defect | fix incorrect krystek implementation nasty bug induced by not looking carefully enough at the reference sadly even with incorrect values it produces meaningful results | 1 |
18,254 | 4,241,262,135 | IssuesEvent | 2016-07-06 15:48:43 | coala-analyzer/coala | https://api.github.com/repos/coala-analyzer/coala | closed | docs: Add FAQ page | area/documentation difficulty/low | I've noticed there's a lot of newcomers that ask similarily the same questions, as in: What are those things failing/not failing in my PR, what issue do I choose etc, because probably they are too lazy to read the full docs (I cannot blame them, it's a lot!).
To fix this, I was thinking we could have a FAQ page where we'd put a few questions that we could simply redirect people to and there they'd read pretty much all they need to know. | 1.0 | docs: Add FAQ page - I've noticed there's a lot of newcomers that ask similarily the same questions, as in: What are those things failing/not failing in my PR, what issue do I choose etc, because probably they are too lazy to read the full docs (I cannot blame them, it's a lot!).
To fix this, I was thinking we could have a FAQ page where we'd put a few questions that we could simply redirect people to and there they'd read pretty much all they need to know. | non_defect | docs add faq page i ve noticed there s a lot of newcomers that ask similarily the same questions as in what are those things failing not failing in my pr what issue do i choose etc because probably they are too lazy to read the full docs i cannot blame them it s a lot to fix this i was thinking we could have a faq page where we d put a few questions that we could simply redirect people to and there they d read pretty much all they need to know | 0 |
45,463 | 12,813,518,142 | IssuesEvent | 2020-07-04 13:28:39 | ontop/ontop | https://api.github.com/repos/ontop/ontop | closed | error with integer type in SPARQL | status: outdated type: defect w: datatype | Reported by @martinrezk
Hi guys,
Quick feedback.
When you do a mapping in protรฉgรฉ (with onto 1.16.1) like
```
:{x} :myProp 14
Selectโฆ
```
Protรฉgรฉ automatically adds ^^xsd:integer, which is ok.
Now, if you query
```
?x :myProp โ14โ^^xsd:integer
```
it gives empty results, more precisely it doesnโt generate any sql. If you change the mapping and the query to xsd:string it works fine.
Cheers
| 1.0 | error with integer type in SPARQL - Reported by @martinrezk
Hi guys,
Quick feedback.
When you do a mapping in protรฉgรฉ (with onto 1.16.1) like
```
:{x} :myProp 14
Selectโฆ
```
Protรฉgรฉ automatically adds ^^xsd:integer, which is ok.
Now, if you query
```
?x :myProp โ14โ^^xsd:integer
```
it gives empty results, more precisely it doesnโt generate any sql. If you change the mapping and the query to xsd:string it works fine.
Cheers
| defect | error with integer type in sparql reported by martinrezk hi guys quick feedback when you do a mapping in protรฉgรฉ with onto like x myprop selectโฆ protรฉgรฉ automatically adds xsd integer which is ok now if you query x myprop โ โ xsd integer it gives empty results more precisely it doesnโt generate any sql if you change the mapping and the query to xsd string it works fine cheers | 1 |
29,141 | 8,300,095,466 | IssuesEvent | 2018-09-21 06:49:06 | michael-a-hansen/spitfire | https://api.github.com/repos/michael-a-hansen/spitfire | opened | Terminate the installation as soon as failure occurs | build/install/ci | Whenever cmake fails or griffon fails to compile, the installation just keeps on trucking and inevitably fails, leaving a useless error message. The installation script needs to have better error checking so that it terminates as soon as a fatal error is encountered, and ideally even gives a helpful error message. | 1.0 | Terminate the installation as soon as failure occurs - Whenever cmake fails or griffon fails to compile, the installation just keeps on trucking and inevitably fails, leaving a useless error message. The installation script needs to have better error checking so that it terminates as soon as a fatal error is encountered, and ideally even gives a helpful error message. | non_defect | terminate the installation as soon as failure occurs whenever cmake fails or griffon fails to compile the installation just keeps on trucking and inevitably fails leaving a useless error message the installation script needs to have better error checking so that it terminates as soon as a fatal error is encountered and ideally even gives a helpful error message | 0 |
845 | 2,594,204,190 | IssuesEvent | 2015-02-20 00:41:32 | BALL-Project/ball | https://api.github.com/repos/BALL-Project/ball | closed | Fix/reimplement AddHydrogenProcessor | C: BALL Core P: major R: fixed T: defect | **Reported by Marcel on 20 Feb 38923081 04:26 UTC**
AddHydrogenProcessor should either be fixed or (perhaps better?) reimplemented.
Currently assigned angles often differ drastically from the correct ones (especially in case of sp3-hybr. carbons). | 1.0 | Fix/reimplement AddHydrogenProcessor - **Reported by Marcel on 20 Feb 38923081 04:26 UTC**
AddHydrogenProcessor should either be fixed or (perhaps better?) reimplemented.
Currently assigned angles often differ drastically from the correct ones (especially in case of sp3-hybr. carbons). | defect | fix reimplement addhydrogenprocessor reported by marcel on feb utc addhydrogenprocessor should either be fixed or perhaps better reimplemented currently assigned angles often differ drastically from the correct ones especially in case of hybr carbons | 1 |
85,552 | 10,618,369,922 | IssuesEvent | 2019-10-13 03:48:51 | carbon-design-system/ibm-dotcom-library | https://api.github.com/repos/carbon-design-system/ibm-dotcom-library | closed | QA the ibm.com Masthead | Sprint Must Have design dotcom migrate | _hahnrob created the following on Sep 10:_
<!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
DDS Designer
> I need to:
test the ibm.com Masthead
> so that I can:
be sure the developer's code correctly works as designed
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- [Create the first iteration of the new Masthead https://github.com/carbon-design-system/ibm-dotcom-library/issues/419](https://zenhub.ibm.com/app/workspaces/digital-design-system-5b3a815a8bf6a932efa6fd03/issues/webstandards/digital-design/727)
#### Acceptance criteria
- [ ] Approval that the code delivered works as designed
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria โ when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1662_ | 1.0 | QA the ibm.com Masthead - _hahnrob created the following on Sep 10:_
<!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
DDS Designer
> I need to:
test the ibm.com Masthead
> so that I can:
be sure the developer's code correctly works as designed
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- [Create the first iteration of the new Masthead https://github.com/carbon-design-system/ibm-dotcom-library/issues/419](https://zenhub.ibm.com/app/workspaces/digital-design-system-5b3a815a8bf6a932efa6fd03/issues/webstandards/digital-design/727)
#### Acceptance criteria
- [ ] Approval that the code delivered works as designed
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria โ when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1662_ | non_defect | qa the ibm com masthead hahnrob created the following on sep user story as a dds designer i need to test the ibm com masthead so that i can be sure the developer s code correctly works as designed additional information acceptance criteria approval that the code delivered works as designed original issue | 0 |
44,055 | 11,360,731,975 | IssuesEvent | 2020-01-26 09:36:47 | tensorflow/addons | https://api.github.com/repos/tensorflow/addons | closed | Move tests and releases to GitHub Actions | build help wanted test-cases | GitHub Actions are now available and provide better integration for releasing / testing. We'll need to keep kokoro for GPU tests but in time we should transition what we can.
- [x] Test Cases
- [ ] Releases | 1.0 | Move tests and releases to GitHub Actions - GitHub Actions are now available and provide better integration for releasing / testing. We'll need to keep kokoro for GPU tests but in time we should transition what we can.
- [x] Test Cases
- [ ] Releases | non_defect | move tests and releases to github actions github actions are now available and provide better integration for releasing testing we ll need to keep kokoro for gpu tests but in time we should transition what we can test cases releases | 0 |
496,257 | 14,343,808,267 | IssuesEvent | 2020-11-28 11:24:10 | azerothcore/azerothcore-wotlk | https://api.github.com/repos/azerothcore/azerothcore-wotlk | closed | Useless item_template.h enum | CORE Priority - Trivial | In item_template.h, we have this and it's not used at all, it's confusing and useless
```cpp
enum ItemFlags2 : uint32
{
ITEM_FLAG2_FACTION_HORDE = 0x00000001,
ITEM_FLAG2_FACTION_ALLIANCE = 0x00000002,
ITEM_FLAG2_DONT_IGNORE_BUY_PRICE = 0x00000004, // when item uses extended cost, gold is also required
ITEM_FLAG2_CLASSIFY_AS_CASTER = 0x00000008,
ITEM_FLAG2_CLASSIFY_AS_PHYSICAL = 0x00000010,
ITEM_FLAG2_EVERYONE_CAN_ROLL_NEED = 0x00000020,
ITEM_FLAG2_NO_TRADE_BIND_ON_ACQUIRE = 0x00000040,
ITEM_FLAG2_CAN_TRADE_BIND_ON_ACQUIRE = 0x00000080,
ITEM_FLAG2_CAN_ONLY_ROLL_GREED = 0x00000100,
ITEM_FLAG2_CASTER_WEAPON = 0x00000200,
ITEM_FLAG2_DELETE_ON_LOGIN = 0x00000400,
ITEM_FLAG2_INTERNAL_ITEM = 0x00000800,
ITEM_FLAG2_NO_VENDOR_VALUE = 0x00001000,
ITEM_FLAG2_SHOW_BEFORE_DISCOVERED = 0x00002000,
ITEM_FLAG2_OVERRIDE_GOLD_COST = 0x00004000,
ITEM_FLAG2_IGNORE_DEFAULT_RATED_BG_RESTRICTIONS = 0x00008000,
ITEM_FLAG2_NOT_USABLE_IN_RATED_BG = 0x00010000,
ITEM_FLAG2_BNET_ACCOUNT_TRADE_OK = 0x00020000,
ITEM_FLAG2_CONFIRM_BEFORE_USE = 0x00040000,
ITEM_FLAG2_REEVALUATE_BONDING_ON_TRANSFORM = 0x00080000,
ITEM_FLAG2_NO_TRANSFORM_ON_CHARGE_DEPLETION = 0x00100000,
ITEM_FLAG2_NO_ALTER_ITEM_VISUAL = 0x00200000,
ITEM_FLAG2_NO_SOURCE_FOR_ITEM_VISUAL = 0x00400000,
ITEM_FLAG2_IGNORE_QUALITY_FOR_ITEM_VISUAL_SOURCE = 0x00800000,
ITEM_FLAG2_NO_DURABILITY = 0x01000000,
ITEM_FLAG2_ROLE_TANK = 0x02000000,
ITEM_FLAG2_ROLE_HEALER = 0x04000000,
ITEM_FLAG2_ROLE_DAMAGE = 0x08000000,
ITEM_FLAG2_CAN_DROP_IN_CHALLENGE_MODE = 0x10000000,
ITEM_FLAG2_NEVER_STACK_IN_LOOT_UI = 0x20000000,
ITEM_FLAG2_DISENCHANT_TO_LOOT_TABLE = 0x40000000,
ITEM_FLAG2_USED_IN_A_TRADESKILL = 0x80000000
};
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/91518473-useless-item_template-h-enum?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Useless item_template.h enum - In item_template.h, we have this and it's not used at all, it's confusing and useless
```cpp
enum ItemFlags2 : uint32
{
ITEM_FLAG2_FACTION_HORDE = 0x00000001,
ITEM_FLAG2_FACTION_ALLIANCE = 0x00000002,
ITEM_FLAG2_DONT_IGNORE_BUY_PRICE = 0x00000004, // when item uses extended cost, gold is also required
ITEM_FLAG2_CLASSIFY_AS_CASTER = 0x00000008,
ITEM_FLAG2_CLASSIFY_AS_PHYSICAL = 0x00000010,
ITEM_FLAG2_EVERYONE_CAN_ROLL_NEED = 0x00000020,
ITEM_FLAG2_NO_TRADE_BIND_ON_ACQUIRE = 0x00000040,
ITEM_FLAG2_CAN_TRADE_BIND_ON_ACQUIRE = 0x00000080,
ITEM_FLAG2_CAN_ONLY_ROLL_GREED = 0x00000100,
ITEM_FLAG2_CASTER_WEAPON = 0x00000200,
ITEM_FLAG2_DELETE_ON_LOGIN = 0x00000400,
ITEM_FLAG2_INTERNAL_ITEM = 0x00000800,
ITEM_FLAG2_NO_VENDOR_VALUE = 0x00001000,
ITEM_FLAG2_SHOW_BEFORE_DISCOVERED = 0x00002000,
ITEM_FLAG2_OVERRIDE_GOLD_COST = 0x00004000,
ITEM_FLAG2_IGNORE_DEFAULT_RATED_BG_RESTRICTIONS = 0x00008000,
ITEM_FLAG2_NOT_USABLE_IN_RATED_BG = 0x00010000,
ITEM_FLAG2_BNET_ACCOUNT_TRADE_OK = 0x00020000,
ITEM_FLAG2_CONFIRM_BEFORE_USE = 0x00040000,
ITEM_FLAG2_REEVALUATE_BONDING_ON_TRANSFORM = 0x00080000,
ITEM_FLAG2_NO_TRANSFORM_ON_CHARGE_DEPLETION = 0x00100000,
ITEM_FLAG2_NO_ALTER_ITEM_VISUAL = 0x00200000,
ITEM_FLAG2_NO_SOURCE_FOR_ITEM_VISUAL = 0x00400000,
ITEM_FLAG2_IGNORE_QUALITY_FOR_ITEM_VISUAL_SOURCE = 0x00800000,
ITEM_FLAG2_NO_DURABILITY = 0x01000000,
ITEM_FLAG2_ROLE_TANK = 0x02000000,
ITEM_FLAG2_ROLE_HEALER = 0x04000000,
ITEM_FLAG2_ROLE_DAMAGE = 0x08000000,
ITEM_FLAG2_CAN_DROP_IN_CHALLENGE_MODE = 0x10000000,
ITEM_FLAG2_NEVER_STACK_IN_LOOT_UI = 0x20000000,
ITEM_FLAG2_DISENCHANT_TO_LOOT_TABLE = 0x40000000,
ITEM_FLAG2_USED_IN_A_TRADESKILL = 0x80000000
};
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/91518473-useless-item_template-h-enum?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | useless item template h enum in item template h we have this and it s not used at all it s confusing and useless cpp enum item faction horde item faction alliance item dont ignore buy price when item uses extended cost gold is also required item classify as caster item classify as physical item everyone can roll need item no trade bind on acquire item can trade bind on acquire item can only roll greed item caster weapon item delete on login item internal item item no vendor value item show before discovered item override gold cost item ignore default rated bg restrictions item not usable in rated bg item bnet account trade ok item confirm before use item reevaluate bonding on transform item no transform on charge depletion item no alter item visual item no source for item visual item ignore quality for item visual source item no durability item role tank item role healer item role damage item can drop in challenge mode item never stack in loot ui item disenchant to loot table item used in a tradeskill want to back this issue we accept bounties via | 0 |
86,200 | 8,028,679,362 | IssuesEvent | 2018-07-27 13:43:13 | italia/spid | https://api.github.com/repos/italia/spid | closed | Richiesta controllo metadata - Comune di Senago | metadata nuovo md test | Buongiorno,
si richiede la validazione dei metadati per conto del comune:
https://sportellounico.comune.senago.mi.it/simplesaml/module.php/saml/sp/metadata.php/senago | 1.0 | Richiesta controllo metadata - Comune di Senago - Buongiorno,
si richiede la validazione dei metadati per conto del comune:
https://sportellounico.comune.senago.mi.it/simplesaml/module.php/saml/sp/metadata.php/senago | non_defect | richiesta controllo metadata comune di senago buongiorno si richiede la validazione dei metadati per conto del comune | 0 |
66,520 | 20,256,213,461 | IssuesEvent | 2022-02-14 23:38:30 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | `Clear room list filter field` does not apply anymore with new spotlight | T-Defect S-Minor O-Occasional | ### Steps to reproduce
you can't type there anymore. escape will not clear but instead close the spotlight search popup.
### Outcome
remove shortcut from list when using new spotlight
### Operating system
arch
### Application version
Element Nightly version: 2022021401 Olm version: 3.2.8
### How did you install the app?
aur
### Homeserver
1.52
### Will you send logs?
No | 1.0 | `Clear room list filter field` does not apply anymore with new spotlight - ### Steps to reproduce
you can't type there anymore. escape will not clear but instead close the spotlight search popup.
### Outcome
remove shortcut from list when using new spotlight
### Operating system
arch
### Application version
Element Nightly version: 2022021401 Olm version: 3.2.8
### How did you install the app?
aur
### Homeserver
1.52
### Will you send logs?
No | defect | clear room list filter field does not apply anymore with new spotlight steps to reproduce you can t type there anymore escape will not clear but instead close the spotlight search popup outcome remove shortcut from list when using new spotlight operating system arch application version element nightly version olm version how did you install the app aur homeserver will you send logs no | 1 |
200,594 | 22,820,320,419 | IssuesEvent | 2022-07-12 01:07:56 | ebubeaso/IT-coding-work | https://api.github.com/repos/ebubeaso/IT-coding-work | opened | CVE-2022-31129 (High) detected in moment-2.29.1.tgz | security vulnerability | ## CVE-2022-31129 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.29.1.tgz</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.29.1.tgz">https://registry.npmjs.org/moment/-/moment-2.29.1.tgz</a></p>
<p>Path to dependency file: /nodejs-work/Usage/backend/package.json</p>
<p>Path to vulnerable library: /nodejs-work/Usage/backend/node_modules/moment/package.json</p>
<p>
Dependency Hierarchy:
- mariadb-2.5.3.tgz (Root Library)
- moment-timezone-0.5.33.tgz
- :x: **moment-2.29.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ebubeaso/IT-coding-work/commit/6d107a6688bc22c52eeb62e12abbb00206f7105f">6d107a6688bc22c52eeb62e12abbb00206f7105f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
moment is a JavaScript date library for parsing, validating, manipulating, and formatting dates. Affected versions of moment were found to use an inefficient parsing algorithm. Specifically using string-to-date parsing in moment (more specifically rfc2822 parsing, which is tried by default) has quadratic (N^2) complexity on specific inputs. Users may notice a noticeable slowdown is observed with inputs above 10k characters. Users who pass user-provided strings without sanity length checks to moment constructor are vulnerable to (Re)DoS attacks. The problem is patched in 2.29.4, the patch can be applied to all affected versions with minimal tweaking. Users are advised to upgrade. Users unable to upgrade should consider limiting date lengths accepted from user input.
<p>Publish Date: 2022-07-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31129>CVE-2022-31129</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-wc69-rhjr-hc9g">https://github.com/moment/moment/security/advisories/GHSA-wc69-rhjr-hc9g</a></p>
<p>Release Date: 2022-07-06</p>
<p>Fix Resolution: moment - 2.29.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-31129 (High) detected in moment-2.29.1.tgz - ## CVE-2022-31129 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.29.1.tgz</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.29.1.tgz">https://registry.npmjs.org/moment/-/moment-2.29.1.tgz</a></p>
<p>Path to dependency file: /nodejs-work/Usage/backend/package.json</p>
<p>Path to vulnerable library: /nodejs-work/Usage/backend/node_modules/moment/package.json</p>
<p>
Dependency Hierarchy:
- mariadb-2.5.3.tgz (Root Library)
- moment-timezone-0.5.33.tgz
- :x: **moment-2.29.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ebubeaso/IT-coding-work/commit/6d107a6688bc22c52eeb62e12abbb00206f7105f">6d107a6688bc22c52eeb62e12abbb00206f7105f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
moment is a JavaScript date library for parsing, validating, manipulating, and formatting dates. Affected versions of moment were found to use an inefficient parsing algorithm. Specifically using string-to-date parsing in moment (more specifically rfc2822 parsing, which is tried by default) has quadratic (N^2) complexity on specific inputs. Users may notice a noticeable slowdown is observed with inputs above 10k characters. Users who pass user-provided strings without sanity length checks to moment constructor are vulnerable to (Re)DoS attacks. The problem is patched in 2.29.4, the patch can be applied to all affected versions with minimal tweaking. Users are advised to upgrade. Users unable to upgrade should consider limiting date lengths accepted from user input.
<p>Publish Date: 2022-07-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31129>CVE-2022-31129</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-wc69-rhjr-hc9g">https://github.com/moment/moment/security/advisories/GHSA-wc69-rhjr-hc9g</a></p>
<p>Release Date: 2022-07-06</p>
<p>Fix Resolution: moment - 2.29.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in moment tgz cve high severity vulnerability vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file nodejs work usage backend package json path to vulnerable library nodejs work usage backend node modules moment package json dependency hierarchy mariadb tgz root library moment timezone tgz x moment tgz vulnerable library found in head commit a href found in base branch master vulnerability details moment is a javascript date library for parsing validating manipulating and formatting dates affected versions of moment were found to use an inefficient parsing algorithm specifically using string to date parsing in moment more specifically parsing which is tried by default has quadratic n complexity on specific inputs users may notice a noticeable slowdown is observed with inputs above characters users who pass user provided strings without sanity length checks to moment constructor are vulnerable to re dos attacks the problem is patched in the patch can be applied to all affected versions with minimal tweaking users are advised to upgrade users unable to upgrade should consider limiting date lengths accepted from user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution moment step up your open source security game with mend | 0 |
84,047 | 10,348,342,632 | IssuesEvent | 2019-09-04 19:34:24 | kubernetes-sigs/cluster-api | https://api.github.com/repos/kubernetes-sigs/cluster-api | closed | Changes to Gitbook business model that may affect this project | kind/documentation lifecycle/active priority/important-longterm | This project seems to rely on Gitbook to generate and serve documentation. Recently, a new version of [Gitbook (v2) dropped support for serving the resources locally](https://docs.gitbook.com/v2-changes/important-differences) which I think will at least affect how developers interact with the tool and the resources managed by it. Did anyone here look into this already?
cc @timothysc | 1.0 | Changes to Gitbook business model that may affect this project - This project seems to rely on Gitbook to generate and serve documentation. Recently, a new version of [Gitbook (v2) dropped support for serving the resources locally](https://docs.gitbook.com/v2-changes/important-differences) which I think will at least affect how developers interact with the tool and the resources managed by it. Did anyone here look into this already?
cc @timothysc | non_defect | changes to gitbook business model that may affect this project this project seems to rely on gitbook to generate and serve documentation recently a new version of which i think will at least affect how developers interact with the tool and the resources managed by it did anyone here look into this already cc timothysc | 0 |
58,553 | 16,597,122,446 | IssuesEvent | 2021-06-01 14:40:08 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | DataTable: scroll headers do not align with body cells in Chromium based browsers | defect elite theme | **Describe the defect**
In scroll mode, headers do not align with body cells. See:
https://www.primefaces.org/showcase/ui/data/datatable/scroll.xhtml
<img width="1111" alt="Screenshot 2021-05-31 at 14 37 07" src="https://user-images.githubusercontent.com/7500178/120290979-216af100-c2c3-11eb-8ab7-e0511cdade25.png">
**Environment:**
- PF Version: _10.0.0_, _11.0.0-SNAPSHOT_
- Affected browsers: _Chrome Version 91.0.4472.77, Safari Version 14.0.3 (16610.4.3.1.7)_ (not with Firefox 88.0.1)
I haven't noticed this on previous Chrome versions | 1.0 | DataTable: scroll headers do not align with body cells in Chromium based browsers - **Describe the defect**
In scroll mode, headers do not align with body cells. See:
https://www.primefaces.org/showcase/ui/data/datatable/scroll.xhtml
<img width="1111" alt="Screenshot 2021-05-31 at 14 37 07" src="https://user-images.githubusercontent.com/7500178/120290979-216af100-c2c3-11eb-8ab7-e0511cdade25.png">
**Environment:**
- PF Version: _10.0.0_, _11.0.0-SNAPSHOT_
- Affected browsers: _Chrome Version 91.0.4472.77, Safari Version 14.0.3 (16610.4.3.1.7)_ (not with Firefox 88.0.1)
I haven't noticed this on previous Chrome versions | defect | datatable scroll headers do not align with body cells in chromium based browsers describe the defect in scroll mode headers do not align with body cells see img width alt screenshot at src environment pf version snapshot affected browsers chrome version safari version not with firefox i haven t noticed this on previous chrome versions | 1 |
734,780 | 25,363,285,557 | IssuesEvent | 2022-11-21 02:35:17 | wso2/api-manager | https://api.github.com/repos/wso2/api-manager | closed | Receiving 404 message error after we deploy a new API revision. | Type/Bug Priority/Normal U2/WUM Affected/APIM-4.0.0 | ### Description
Hi Team,
In the APIM 4.0.0 we are facing the following error when invoking the API after performing API revision
{"code":"404","type":"Status report","message":"Not Found","description":"The requested resource is not available."}
We have observed these 404 error messages in 2 scenarios as follows:
Case 1: Deploy a new Revision; After that, Make 10K requests with 20 threads/seconds
For this case we have tested on a lower environment where there was no extra traffic beside the one created by us. We tested the below scenario:
Run 10.000 API calls
Wait for 5 minutes (no traffic)
Deploy a new API revision
Wait for 3 minutes (no traffic)
Run 10.000 API calls --> we received 404 on 50% of the calls
Case 2: Makes 10,000 API calls with 20 threads/seconds and changes the revision (deploy and un-deploy) at the same time
If the number of Initializing and Destroying Api logs are the same we will not encounter any issue during 10.000 API calls.
If the number of Initializing logs is less the Destroying API logs --> we encounter 404 error message
Could you please provide your inputs on these above observations of the customer and your suggestions on how we can resolve these issues.
### Steps to Reproduce
Case 1: Deploy a new Revision; After that, Make 10K requests with 20 threads/seconds
For this case we have tested on a lower environment where there was no extra traffic beside the one created by us. We tested the below scenario:
Run 10.000 API calls
Wait for 5 minutes (no traffic)
Deploy a new API revision
Wait for 3 minutes (no traffic)
Run 10.000 API calls --> we received 404 on 50% of the calls
### Affected Component
APIM
### Version
4.0.0
### Environment Details (with versions)
_No response_
### Relevant Log Output
_No response_
### Related Issues
_No response_
### Suggested Labels
_No response_ | 1.0 | Receiving 404 message error after we deploy a new API revision. - ### Description
Hi Team,
In the APIM 4.0.0 we are facing the following error when invoking the API after performing API revision
{"code":"404","type":"Status report","message":"Not Found","description":"The requested resource is not available."}
We have observed these 404 error messages in 2 scenarios as follows:
Case 1: Deploy a new Revision; After that, Make 10K requests with 20 threads/seconds
For this case we have tested on a lower environment where there was no extra traffic beside the one created by us. We tested the below scenario:
Run 10.000 API calls
Wait for 5 minutes (no traffic)
Deploy a new API revision
Wait for 3 minutes (no traffic)
Run 10.000 API calls --> we received 404 on 50% of the calls
Case 2: Makes 10,000 API calls with 20 threads/seconds and changes the revision (deploy and un-deploy) at the same time
If the number of Initializing and Destroying Api logs are the same we will not encounter any issue during 10.000 API calls.
If the number of Initializing logs is less the Destroying API logs --> we encounter 404 error message
Could you please provide your inputs on these above observations of the customer and your suggestions on how we can resolve these issues.
### Steps to Reproduce
Case 1: Deploy a new Revision; After that, Make 10K requests with 20 threads/seconds
For this case we have tested on a lower environment where there was no extra traffic beside the one created by us. We tested the below scenario:
Run 10.000 API calls
Wait for 5 minutes (no traffic)
Deploy a new API revision
Wait for 3 minutes (no traffic)
Run 10.000 API calls --> we received 404 on 50% of the calls
### Affected Component
APIM
### Version
4.0.0
### Environment Details (with versions)
_No response_
### Relevant Log Output
_No response_
### Related Issues
_No response_
### Suggested Labels
_No response_ | non_defect | receiving message error after we deploy a new api revision description hi team in the apim we are facing the following error when invoking the api after performing api revision code type status report message not found description the requested resource is not available we have observed these error messages in scenarios as follows case deploy a new revision after that make requests with threads seconds for this case we have tested on a lower environment where there was no extra traffic beside the one created by us we tested the below scenario run api calls wait for minutes no traffic deploy a new api revision wait for minutes no traffic run api calls we received on of the calls case makes api calls with threads seconds and changes the revision deploy and un deploy at the same time if the number of initializing and destroying api logs are the same we will not encounter any issue during api calls if the number of initializing logs is less the destroying api logs we encounter error message could you please provide your inputs on these above observations of the customer and your suggestions on how we can resolve these issues steps to reproduce case deploy a new revision after that make requests with threads seconds for this case we have tested on a lower environment where there was no extra traffic beside the one created by us we tested the below scenario run api calls wait for minutes no traffic deploy a new api revision wait for minutes no traffic run api calls we received on of the calls affected component apim version environment details with versions no response relevant log output no response related issues no response suggested labels no response | 0 |
70,465 | 23,179,744,230 | IssuesEvent | 2022-07-31 23:34:15 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | opened | Error in session write | defect | ### Description
Hello, I recently tried to write value of a Session in my Component with:
$this->getController()->getRequest()->getSession()->write('visit_check', true);
Then the website throw this error: Cake\Http\Session::_overwrite(): Argument #1 ($old) must be of type array, null given, called in C:\laragon\www\Project\vendor\cakephp\cakephp\src\Http\Session.php on line 508
### CakePHP Version
4.4.3
### PHP Version
8.1.8 | 1.0 | Error in session write - ### Description
Hello, I recently tried to write value of a Session in my Component with:
$this->getController()->getRequest()->getSession()->write('visit_check', true);
Then the website throw this error: Cake\Http\Session::_overwrite(): Argument #1 ($old) must be of type array, null given, called in C:\laragon\www\Project\vendor\cakephp\cakephp\src\Http\Session.php on line 508
### CakePHP Version
4.4.3
### PHP Version
8.1.8 | defect | error in session write description hello i recently tried to write value of a session in my component with this getcontroller getrequest getsession write visit check true then the website throw this error cake http session overwrite argument old must be of type array null given called in c laragon www project vendor cakephp cakephp src http session php on line cakephp version php version | 1 |
135,017 | 18,537,639,354 | IssuesEvent | 2021-10-21 13:12:13 | elastic/integrations | https://api.github.com/repos/elastic/integrations | closed | Update title and description based on guidelines for Elastic Agent based Security integrations | Team:Security-External Integrations 7.16 candidate | In 7.16, we are planning to introduce the unified integrations view. The primary goal of this view is to improve the discoverability for all data ingest options at Elastic. As part of this initiative, we'd like to improve the titles and descriptions of each integration card and do so in a uniform way.
## Guidelines
These are the general guidelines we plan to use for the integration cards

## Elastic Agent example
Here's an example for how agent can leverage these guidelines, but please follow the guidelines as you see fit.
| Current Title | Current Description | New Title | New Description |
|---------------|----------------------------------------------------------------------------------|-------------|-----------------------------------------------------------------------|
| ActiveMQ | ActiveMQ Integration | ActiveMQ | Collect logs and metrics from ActiveMQ servers with Elastic Agent. |
| Apache | This Elastic integration collects logs and metrics from Apache servers | Apache | Collect logs and metrics from Apache servers with Elastic Agent. |
| Auditd | This Elastic integration collects and parses logs from the Audit daemon (auditd) | Auditd | Collect logs from the Linux audit daemon with Elastic Agent. |
| AWS | This integration collects logs and metrics from Amazon Web Services (AWS) | AWS | Collect logs and metrics from Amazon Web Services with Elastic Agent. |
| AWS Billing | | AWS Billing | Collect logs and metrics from AWS Billing with Elastic Agent. |
| Azure Logs | This Elastic integration collects logs from Azure | Azure | Collect logs from Microsoft Azure with Elastic Agent. |
| True | Update title and description based on guidelines for Elastic Agent based Security integrations - In 7.16, we are planning to introduce the unified integrations view. The primary goal of this view is to improve the discoverability for all data ingest options at Elastic. As part of this initiative, we'd like to improve the titles and descriptions of each integration card and do so in a uniform way.
## Guidelines
These are the general guidelines we plan to use for the integration cards

## Elastic Agent example
Here's an example for how agent can leverage these guidelines, but please follow the guidelines as you see fit.
| Current Title | Current Description | New Title | New Description |
|---------------|----------------------------------------------------------------------------------|-------------|-----------------------------------------------------------------------|
| ActiveMQ | ActiveMQ Integration | ActiveMQ | Collect logs and metrics from ActiveMQ servers with Elastic Agent. |
| Apache | This Elastic integration collects logs and metrics from Apache servers | Apache | Collect logs and metrics from Apache servers with Elastic Agent. |
| Auditd | This Elastic integration collects and parses logs from the Audit daemon (auditd) | Auditd | Collect logs from the Linux audit daemon with Elastic Agent. |
| AWS | This integration collects logs and metrics from Amazon Web Services (AWS) | AWS | Collect logs and metrics from Amazon Web Services with Elastic Agent. |
| AWS Billing | | AWS Billing | Collect logs and metrics from AWS Billing with Elastic Agent. |
| Azure Logs | This Elastic integration collects logs from Azure | Azure | Collect logs from Microsoft Azure with Elastic Agent. |
| non_defect | update title and description based on guidelines for elastic agent based security integrations in we are planning to introduce the unified integrations view the primary goal of this view is to improve the discoverability for all data ingest options at elastic as part of this initiative we d like to improve the titles and descriptions of each integration card and do so in a uniform way guidelines these are the general guidelines we plan to use for the integration cards elastic agent example here s an example for how agent can leverage these guidelines but please follow the guidelines as you see fit current title current description new title new description activemq activemq integration activemq collect logs and metrics from activemq servers with elastic agent apache this elastic integration collects logs and metrics from apache servers apache collect logs and metrics from apache servers with elastic agent auditd this elastic integration collects and parses logs from the audit daemon auditd auditd collect logs from the linux audit daemon with elastic agent aws this integration collects logs and metrics from amazon web services aws aws collect logs and metrics from amazon web services with elastic agent aws billing aws billing collect logs and metrics from aws billing with elastic agent azure logs this elastic integration collects logs from azure azure collect logs from microsoft azure with elastic agent | 0 |
400,735 | 11,779,917,504 | IssuesEvent | 2020-03-16 18:57:46 | alexdevteam/munchkin-online | https://api.github.com/repos/alexdevteam/munchkin-online | closed | Support text wrapping around (Text Renderer) | enhancement high priority | Let the text renderer support wrapping around when given a rect to be placed into. | 1.0 | Support text wrapping around (Text Renderer) - Let the text renderer support wrapping around when given a rect to be placed into. | non_defect | support text wrapping around text renderer let the text renderer support wrapping around when given a rect to be placed into | 0 |
51,262 | 13,207,403,673 | IssuesEvent | 2020-08-14 22:58:28 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | trac attach file to ticket fails. (Trac #108) | Incomplete Migration Migrated from Trac defect infrastructure | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/108">https://code.icecube.wisc.edu/projects/icecube/ticket/108</a>, reported by blaufussand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-12-03T15:32:37",
"_ts": "1196695957000000",
"description": "Trying to attach a .py script to a ticket from OS X on my MacBook.\n\nIt barfs.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2007-08-31T18:54:13",
"component": "infrastructure",
"summary": "trac attach file to ticket fails.",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
| 1.0 | trac attach file to ticket fails. (Trac #108) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/108">https://code.icecube.wisc.edu/projects/icecube/ticket/108</a>, reported by blaufussand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-12-03T15:32:37",
"_ts": "1196695957000000",
"description": "Trying to attach a .py script to a ticket from OS X on my MacBook.\n\nIt barfs.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2007-08-31T18:54:13",
"component": "infrastructure",
"summary": "trac attach file to ticket fails.",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
| defect | trac attach file to ticket fails trac migrated from json status closed changetime ts description trying to attach a py script to a ticket from os x on my macbook n nit barfs reporter blaufuss cc resolution fixed time component infrastructure summary trac attach file to ticket fails priority normal keywords milestone owner cgils type defect | 1 |
101,781 | 4,135,708,563 | IssuesEvent | 2016-06-13 01:00:34 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | opened | RigidBodyTree::findLinkId() is questionable | priority: medium team: software core type: cleanup | This method returns the body index of a link. Shouldn't this be an internal implementation detail that's hidden from the end user?
For context, see: https://github.com/RobotLocomotion/drake/pull/2426#issuecomment-224096790 | 1.0 | RigidBodyTree::findLinkId() is questionable - This method returns the body index of a link. Shouldn't this be an internal implementation detail that's hidden from the end user?
For context, see: https://github.com/RobotLocomotion/drake/pull/2426#issuecomment-224096790 | non_defect | rigidbodytree findlinkid is questionable this method returns the body index of a link shouldn t this be an internal implementation detail that s hidden from the end user for context see | 0 |
351,477 | 25,028,179,807 | IssuesEvent | 2022-11-04 09:58:35 | andrasfuchs/BioBalanceDetector | https://api.github.com/repos/andrasfuchs/BioBalanceDetector | closed | Update wiki with the setup steps | documentation | Add the setup section for Windows, Ubuntu, Raspberry Pi and Android to [the wiki page](https://github.com/andrasfuchs/BioBalanceDetector/wiki/Proto-%238-Body-Monitor).
Update the sampling rate and FFT size limitations for each platform. | 1.0 | Update wiki with the setup steps - Add the setup section for Windows, Ubuntu, Raspberry Pi and Android to [the wiki page](https://github.com/andrasfuchs/BioBalanceDetector/wiki/Proto-%238-Body-Monitor).
Update the sampling rate and FFT size limitations for each platform. | non_defect | update wiki with the setup steps add the setup section for windows ubuntu raspberry pi and android to update the sampling rate and fft size limitations for each platform | 0 |
729,629 | 25,136,180,475 | IssuesEvent | 2022-11-09 18:47:53 | WordPress/openverse-frontend | https://api.github.com/repos/WordPress/openverse-frontend | closed | The `old_header` buttons have incorrect height | ๐ง priority: high ๐ goal: fix ๐น aspect: interface | ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
The (old) header buttons on desktop widths stretch to the height of the highest element. Currently, the tallest element is the `VLogoButtonOld`, which is 50px high instead of 48px as in the mockups. This causes the other buttons to stretch to 50px.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. <!-- Step 1 ... -->Go to search-staging.openverse.engineering/search/?q=cat
2. <!-- Step 2 ... -->Check the buttons in the header.
3. See error: the buttons are 50px high instead of 48px.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
<img width="871" alt="Screen Shot 2022-10-29 at 4 26 42 PM" src="https://user-images.githubusercontent.com/15233243/198834394-fb891645-fd06-48b6-b323-3a274be1cf82.png">
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
This change wasn't caught by the Visual regression tests because I've previously set the `maxDiffPixelRatio` to 0.06 (default is 0).
The reason I increased this value is the fact that the media reuse tests often scroll a couple of pixels up or down (I think because of the main image loading at various speeds) and cause the failure of visual regression tests even though there is no problem in reality. This can probably be fixed by moving the media reuse tests to Storybook VR tests. I'll open the issue for that, too.
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] ๐ I would be interested in resolving this bug.
| 1.0 | The `old_header` buttons have incorrect height - ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
The (old) header buttons on desktop widths stretch to the height of the highest element. Currently, the tallest element is the `VLogoButtonOld`, which is 50px high instead of 48px as in the mockups. This causes the other buttons to stretch to 50px.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. <!-- Step 1 ... -->Go to search-staging.openverse.engineering/search/?q=cat
2. <!-- Step 2 ... -->Check the buttons in the header.
3. See error: the buttons are 50px high instead of 48px.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
<img width="871" alt="Screen Shot 2022-10-29 at 4 26 42 PM" src="https://user-images.githubusercontent.com/15233243/198834394-fb891645-fd06-48b6-b323-3a274be1cf82.png">
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
This change wasn't caught by the Visual regression tests because I've previously set the `maxDiffPixelRatio` to 0.06 (default is 0).
The reason I increased this value is the fact that the media reuse tests often scroll a couple of pixels up or down (I think because of the main image loading at various speeds) and cause the failure of visual regression tests even though there is no problem in reality. This can probably be fixed by moving the media reuse tests to Storybook VR tests. I'll open the issue for that, too.
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] ๐ I would be interested in resolving this bug.
| non_defect | the old header buttons have incorrect height description the old header buttons on desktop widths stretch to the height of the highest element currently the tallest element is the vlogobuttonold which is high instead of as in the mockups this causes the other buttons to stretch to reproduction go to search staging openverse engineering search q cat check the buttons in the header see error the buttons are high instead of screenshots img width alt screen shot at pm src additional context this change wasn t caught by the visual regression tests because i ve previously set the maxdiffpixelratio to default is the reason i increased this value is the fact that the media reuse tests often scroll a couple of pixels up or down i think because of the main image loading at various speeds and cause the failure of visual regression tests even though there is no problem in reality this can probably be fixed by moving the media reuse tests to storybook vr tests i ll open the issue for that too resolution ๐ i would be interested in resolving this bug | 0 |
54,823 | 23,341,414,536 | IssuesEvent | 2022-08-09 14:18:41 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Create a dedicated API for increasing popularity of field | loe:hours Feature:Data Views Team:AppServicesSv impact:low | https://github.com/elastic/kibana/pull/31652 adds more granular feature control, such as Index Pattern Management. Setting Index Pattern Management to `read` or `none` will break our popular fields increment in Discover when adding fields to columns in the doc table. We could fix this by adding a dedicated API for incrementing field popularity. Then all we would have to do is add a tag to the route and give the Discover feature access to it in API privilege. | 1.0 | Create a dedicated API for increasing popularity of field - https://github.com/elastic/kibana/pull/31652 adds more granular feature control, such as Index Pattern Management. Setting Index Pattern Management to `read` or `none` will break our popular fields increment in Discover when adding fields to columns in the doc table. We could fix this by adding a dedicated API for incrementing field popularity. Then all we would have to do is add a tag to the route and give the Discover feature access to it in API privilege. | non_defect | create a dedicated api for increasing popularity of field adds more granular feature control such as index pattern management setting index pattern management to read or none will break our popular fields increment in discover when adding fields to columns in the doc table we could fix this by adding a dedicated api for incrementing field popularity then all we would have to do is add a tag to the route and give the discover feature access to it in api privilege | 0 |
60,195 | 17,023,366,109 | IssuesEvent | 2021-07-03 01:38:57 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [Openlayers] Tiles are scattered in Opera | Component: website Priority: major Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 8.45am, Sunday, 15th February 2009]**
When moving and scaling the map in Opera, the tiles are not on it's place, but scattered around. It will take a while, until the tiles are loaded, and the map is readable again. | 1.0 | [Openlayers] Tiles are scattered in Opera - **[Submitted to the original trac issue database at 8.45am, Sunday, 15th February 2009]**
When moving and scaling the map in Opera, the tiles are not on it's place, but scattered around. It will take a while, until the tiles are loaded, and the map is readable again. | defect | tiles are scattered in opera when moving and scaling the map in opera the tiles are not on it s place but scattered around it will take a while until the tiles are loaded and the map is readable again | 1 |
26,041 | 4,559,089,098 | IssuesEvent | 2016-09-14 00:12:42 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | NullReferenceException. Object reference not set to an instance of an object. | defect | ```
System.NullReferenceException: Object reference not set to an instance of an object.
at Bridge.Translator.SharpSixRewriter.VisitArgument(ArgumentSyntax node)
```
### Steps To Reproduce
http://deck.net/c1948b2d7f56af0c72f8f2f97a578e1e
```csharp
public class Array
{
public void Add(params object[] obj)
{
}
}
public class Program
{
public void Main()
{
Array callbacks = new Array();
callbacks.Add(((System.Action<object>)(instance_callback)));
}
//remove static and its will compile
static public void instance_callback(object a = null)
{
}
}
``` | 1.0 | NullReferenceException. Object reference not set to an instance of an object. - ```
System.NullReferenceException: Object reference not set to an instance of an object.
at Bridge.Translator.SharpSixRewriter.VisitArgument(ArgumentSyntax node)
```
### Steps To Reproduce
http://deck.net/c1948b2d7f56af0c72f8f2f97a578e1e
```csharp
public class Array
{
public void Add(params object[] obj)
{
}
}
public class Program
{
public void Main()
{
Array callbacks = new Array();
callbacks.Add(((System.Action<object>)(instance_callback)));
}
//remove static and its will compile
static public void instance_callback(object a = null)
{
}
}
``` | defect | nullreferenceexception object reference not set to an instance of an object system nullreferenceexception object reference not set to an instance of an object at bridge translator sharpsixrewriter visitargument argumentsyntax node steps to reproduce csharp public class array public void add params object obj public class program public void main array callbacks new array callbacks add system action instance callback remove static and its will compile static public void instance callback object a null | 1 |
131,789 | 10,710,360,708 | IssuesEvent | 2019-10-25 01:54:04 | apache/skywalking | https://api.github.com/repos/apache/skywalking | closed | Create install folder in main repo | test | This task is a pre-requirement to make istio/envoy ci test happen in CI.
We should create an `install` folder in the root folder then store all of the relevant items in it.
@innerpeacez Can you help on to copy the newest helm chart into `install/kubernetes/helm`? and change all oap images tags to `latest`, thanks. | 1.0 | Create install folder in main repo - This task is a pre-requirement to make istio/envoy ci test happen in CI.
We should create an `install` folder in the root folder then store all of the relevant items in it.
@innerpeacez Can you help on to copy the newest helm chart into `install/kubernetes/helm`? and change all oap images tags to `latest`, thanks. | non_defect | create install folder in main repo this task is a pre requirement to make istio envoy ci test happen in ci we should create an install folder in the root folder then store all of the relevant items in it innerpeacez can you help on to copy the newest helm chart into install kubernetes helm and change all oap images tags to latest thanks | 0 |
48,297 | 13,068,427,393 | IssuesEvent | 2020-07-31 03:32:39 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | L1/L2 Keep modules appear to delete TrayInfo (Trac #2146) | Migrated from Trac combo reconstruction defect | Recently when investigating possible bugs in custom simulation I was examining the recorded I3TrayInfo objects, but hit a wall at L1/L2, as the `I` frames from previous levels existed but were empty. This looks like the work of the `Keep` module, so I propose that "I3TrayInfo" should be added to the whitelists used in the processing scripts to prevent this kind of information loss.
Migrated from https://code.icecube.wisc.edu/ticket/2146
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"description": "Recently when investigating possible bugs in custom simulation I was examining the recorded I3TrayInfo objects, but hit a wall at L1/L2, as the `I` frames from previous levels existed but were empty. This looks like the work of the `Keep` module, so I propose that \"I3TrayInfo\" should be added to the whitelists used in the processing scripts to prevent this kind of information loss. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1550067323910946",
"component": "combo reconstruction",
"summary": "L1/L2 Keep modules appear to delete TrayInfo",
"priority": "minor",
"keywords": "",
"time": "2018-04-19T15:35:17",
"milestone": "",
"owner": "mjl5147",
"type": "defect"
}
```
| 1.0 | L1/L2 Keep modules appear to delete TrayInfo (Trac #2146) - Recently when investigating possible bugs in custom simulation I was examining the recorded I3TrayInfo objects, but hit a wall at L1/L2, as the `I` frames from previous levels existed but were empty. This looks like the work of the `Keep` module, so I propose that "I3TrayInfo" should be added to the whitelists used in the processing scripts to prevent this kind of information loss.
Migrated from https://code.icecube.wisc.edu/ticket/2146
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"description": "Recently when investigating possible bugs in custom simulation I was examining the recorded I3TrayInfo objects, but hit a wall at L1/L2, as the `I` frames from previous levels existed but were empty. This looks like the work of the `Keep` module, so I propose that \"I3TrayInfo\" should be added to the whitelists used in the processing scripts to prevent this kind of information loss. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1550067323910946",
"component": "combo reconstruction",
"summary": "L1/L2 Keep modules appear to delete TrayInfo",
"priority": "minor",
"keywords": "",
"time": "2018-04-19T15:35:17",
"milestone": "",
"owner": "mjl5147",
"type": "defect"
}
```
| defect | keep modules appear to delete trayinfo trac recently when investigating possible bugs in custom simulation i was examining the recorded objects but hit a wall at as the i frames from previous levels existed but were empty this looks like the work of the keep module so i propose that should be added to the whitelists used in the processing scripts to prevent this kind of information loss migrated from json status closed changetime description recently when investigating possible bugs in custom simulation i was examining the recorded objects but hit a wall at as the i frames from previous levels existed but were empty this looks like the work of the keep module so i propose that should be added to the whitelists used in the processing scripts to prevent this kind of information loss reporter cweaver cc resolution fixed ts component combo reconstruction summary keep modules appear to delete trayinfo priority minor keywords time milestone owner type defect | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.