Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
853
| labels
stringlengths 4
898
| body
stringlengths 2
262k
| index
stringclasses 13
values | text_combine
stringlengths 96
262k
| label
stringclasses 2
values | text
stringlengths 96
250k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
769,701
| 27,016,818,788
|
IssuesEvent
|
2023-02-10 20:14:16
|
tallyhowallet/extension
|
https://api.github.com/repos/tallyhowallet/extension
|
closed
|
Unable to connect to yearn
|
Type: Bug Status: Pending Priority: Medium
|
### Discord Discussion Link
_No response_
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
Connect Tally to yearn!
### What did not work?
When Tally is enabled as default wallet... I'm not able to connect with Tally; rather I get a notice re: mm
<img width="1331" alt="Screen Shot 2022-02-23 at 4 31 02 PM" src="https://user-images.githubusercontent.com/7005061/155411994-e1717b7f-07c5-4ca7-95e4-4ac3f4fd1650.png">
https://user-images.githubusercontent.com/7005061/155412004-12fdd36f-b116-4665-95a3-5d6496ec9214.mov
### Version
_No response_
### Relevant log output
_No response_
|
1.0
|
Unable to connect to yearn - ### Discord Discussion Link
_No response_
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
Connect Tally to yearn!
### What did not work?
When Tally is enabled as default wallet... I'm not able to connect with Tally; rather I get a notice re: mm
<img width="1331" alt="Screen Shot 2022-02-23 at 4 31 02 PM" src="https://user-images.githubusercontent.com/7005061/155411994-e1717b7f-07c5-4ca7-95e4-4ac3f4fd1650.png">
https://user-images.githubusercontent.com/7005061/155412004-12fdd36f-b116-4665-95a3-5d6496ec9214.mov
### Version
_No response_
### Relevant log output
_No response_
|
non_build
|
unable to connect to yearn discord discussion link no response what browsers are you seeing the problem on chrome what were you trying to do connect tally to yearn what did not work when tally is enabled as default wallet i m not able to connect with tally rather i get a notice re mm img width alt screen shot at pm src version no response relevant log output no response
| 0
|
15,859
| 6,047,695,660
|
IssuesEvent
|
2017-06-12 14:57:38
|
meteor/meteor
|
https://api.github.com/repos/meteor/meteor
|
closed
|
Define concatenated css files with html import?
|
feature Project:Isobuild:Minifiers
|
http://stackoverflow.com/questions/31077273/how-can-an-external-stylesheet-be-used-to-style-polymer-1-0-elements
As asked in this stackoverflow, currently when using polymer you are able to style paper-elements (or custom elements in general cross browser in the future) with custom properties or whatever they're called sorry. However, you need to add is="custom-style" to your style tag.
Meteor takes all the css files and concatenates them automatically and uses the link tag to import them, so you don't even end up with a style tag. A solution for the problem is doing an html import and having the styles included in that file inline. But as mentioned in my comment on stackoverflow this means you need to add at least one additional http request. I know I'm super nitpicky here but developing for the Philippines (slowest internet speed in Asia, 3.6MBPS average) and Asia just means you want to optimize even the slighest detail for speed.
Would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don't need to generate two seperate css files?
|
1.0
|
Define concatenated css files with html import? - http://stackoverflow.com/questions/31077273/how-can-an-external-stylesheet-be-used-to-style-polymer-1-0-elements
As asked in this stackoverflow, currently when using polymer you are able to style paper-elements (or custom elements in general cross browser in the future) with custom properties or whatever they're called sorry. However, you need to add is="custom-style" to your style tag.
Meteor takes all the css files and concatenates them automatically and uses the link tag to import them, so you don't even end up with a style tag. A solution for the problem is doing an html import and having the styles included in that file inline. But as mentioned in my comment on stackoverflow this means you need to add at least one additional http request. I know I'm super nitpicky here but developing for the Philippines (slowest internet speed in Asia, 3.6MBPS average) and Asia just means you want to optimize even the slighest detail for speed.
Would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don't need to generate two seperate css files?
|
build
|
define concatenated css files with html import as asked in this stackoverflow currently when using polymer you are able to style paper elements or custom elements in general cross browser in the future with custom properties or whatever they re called sorry however you need to add is custom style to your style tag meteor takes all the css files and concatenates them automatically and uses the link tag to import them so you don t even end up with a style tag a solution for the problem is doing an html import and having the styles included in that file inline but as mentioned in my comment on stackoverflow this means you need to add at least one additional http request i know i m super nitpicky here but developing for the philippines slowest internet speed in asia average and asia just means you want to optimize even the slighest detail for speed would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don t need to generate two seperate css files
| 1
|
46,165
| 11,795,098,320
|
IssuesEvent
|
2020-03-18 08:15:07
|
xamarin/xamarin-android
|
https://api.github.com/repos/xamarin/xamarin-android
|
opened
|
R8 throws an error when code shrinker is set to none
|
Area: App+Library Build
|
### Steps to Reproduce
0. Create empty project
1. Install package com.onesginal,
2. Set DexCompiler to d8
3. Set Code Shrinker to none
4. Enable multidex
### Expected Behavior
There are no errors from R8 (because code shrinker is set to none)
### Actual Behavior
R8 Fires error
```
2>R8 : warning : Missing class: com.google.android.gms.location.LocationListener
2>R8 : error : Compilation can't be completed because some library classes are missing.
```
### Version Information
All details, sample repo, and more information are here: https://github.com/OneSignal/OneSignal-Xamarin-SDK/issues/185
Why R8 is fired when mulidex is enabled? If that is default behavior, then visual studio csproj editor should chcck "codeshrinker" to d8 automatically when selecting multidex, otherwise developer don't know whats happening.
|
1.0
|
R8 throws an error when code shrinker is set to none - ### Steps to Reproduce
0. Create empty project
1. Install package com.onesginal,
2. Set DexCompiler to d8
3. Set Code Shrinker to none
4. Enable multidex
### Expected Behavior
There are no errors from R8 (because code shrinker is set to none)
### Actual Behavior
R8 Fires error
```
2>R8 : warning : Missing class: com.google.android.gms.location.LocationListener
2>R8 : error : Compilation can't be completed because some library classes are missing.
```
### Version Information
All details, sample repo, and more information are here: https://github.com/OneSignal/OneSignal-Xamarin-SDK/issues/185
Why R8 is fired when mulidex is enabled? If that is default behavior, then visual studio csproj editor should chcck "codeshrinker" to d8 automatically when selecting multidex, otherwise developer don't know whats happening.
|
build
|
throws an error when code shrinker is set to none steps to reproduce create empty project install package com onesginal set dexcompiler to set code shrinker to none enable multidex expected behavior there are no errors from because code shrinker is set to none actual behavior fires error warning missing class com google android gms location locationlistener error compilation can t be completed because some library classes are missing version information all details sample repo and more information are here why is fired when mulidex is enabled if that is default behavior then visual studio csproj editor should chcck codeshrinker to automatically when selecting multidex otherwise developer don t know whats happening
| 1
|
22,608
| 11,685,931,389
|
IssuesEvent
|
2020-03-05 09:58:42
|
Automattic/jetpack
|
https://api.github.com/repos/Automattic/jetpack
|
closed
|
Tiled Gallery block: images are served with Photon, but not resized
|
Gutenberg Performance Tiled Galleries [Block] Tiled Gallery [Pri] High [Type] Enhancement
|
#### Steps to reproduce the issue
1. Add a Tiled Gallery with Gutenberg using large, hi-res images.
2. The images on the back end and front end aren't resized with Photon. They are loading in the large, full-sized images instead. This causes the browser to freeze up and leads to slow loading on the front end.
#### What I expected
Photon should have added the "w" and "h" attributes to load in smaller versions of the images. This works correctly when using Tiled Gallery and the Classic Editor.
#### What happened instead
Tiled Gallery (Gutenberg) loads in the full-sized images. This causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large.
|
True
|
Tiled Gallery block: images are served with Photon, but not resized - #### Steps to reproduce the issue
1. Add a Tiled Gallery with Gutenberg using large, hi-res images.
2. The images on the back end and front end aren't resized with Photon. They are loading in the large, full-sized images instead. This causes the browser to freeze up and leads to slow loading on the front end.
#### What I expected
Photon should have added the "w" and "h" attributes to load in smaller versions of the images. This works correctly when using Tiled Gallery and the Classic Editor.
#### What happened instead
Tiled Gallery (Gutenberg) loads in the full-sized images. This causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large.
|
non_build
|
tiled gallery block images are served with photon but not resized steps to reproduce the issue add a tiled gallery with gutenberg using large hi res images the images on the back end and front end aren t resized with photon they are loading in the large full sized images instead this causes the browser to freeze up and leads to slow loading on the front end what i expected photon should have added the w and h attributes to load in smaller versions of the images this works correctly when using tiled gallery and the classic editor what happened instead tiled gallery gutenberg loads in the full sized images this causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large
| 0
|
53,634
| 13,189,973,272
|
IssuesEvent
|
2020-08-13 09:22:30
|
googleapis/java-bigtable-hbase
|
https://api.github.com/repos/googleapis/java-bigtable-hbase
|
opened
|
bigtable.grpc.scanner.ReadRowsAcceptanceTest: many tests failed
|
buildcop: issue priority: p1 type: bug
|
Many tests failed at the same time in this package.
* I will close this issue when there are no more failures in this package _and_
there is at least one pass.
* No new issues will be filed for this package until this issue is closed.
* If there are already issues for individual test cases, I will close them when
the corresponding test passes. You can close them earlier, if you prefer, and
I won't reopen them while this issue is still open.
Here are the tests that failed:
* test[invalidNoCommit]
* test[invalidNoCellKeyBeforeCommit]
* test[invalidNoCellKeyBeforeValue]
* test[invalidNewColFamilyMustSpecifyQualifier]
* test[bareCommitImpliesTs=0]
* test[simpleRowWithTimestamp]
* test[missingTimestamp,ImpliedTs=0]
* test[emptyCellValue]
* test[twoUnsplitCells]
* test[twoQualifiers]
* test[twoFamilies]
* test[withLabels]
* test[splitCell,BareCommit]
* test[splitCell]
* test[splitFourWays]
* test[twoSplitCells]
* test[multiQualifierSplits]
* test[multiQualifierMultiSplit]
* test[multiFamilySplit]
* test[invalidNoCommitBetweenRows]
* test[invalidNoCommitAfterFirstRow]
* test[invalidLastRowMissingCommit]
* test[invalidDuplicateRowKey]
* test[invalidNewRowMissingRowKey]
* test[twoRows]
* test[twoRowsImplicitTimestamp]
* test[twoRowsEmptyValue]
* test[twoRows,OneWithMultipleCells]
* test[twoRows,MultipleCells]
* test[twoRows,MultipleCells,MultipleFamilies]
* test[twoRows,FourCells,2Labels]
* test[twoRowsWithSplits,SameTimestamp]
* test[invalidBareReset]
* test[invalidBadReset,NoCommit]
* test[invalidMissingKeyAfterReset]
* test[noDataAfterReset]
* test[simpleReset]
* test[resetToNewVal]
* test[resetToNewQual]
* test[resetWithSplits]
* test[resetTwoCells]
* test[twoResets]
* test[resetThenTwoCells]
* test[resetToNewRow]
* test[resetInBetweenChunks]
* test[invalidResetWithChunk]
* test[invalidCommitWithChunk]
* test[emptyCellChunk]
-----
commit: 81641f77b42506cdd7f0ddcb2192b5f0151aa413
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1c4ccf74-76a3-4494-a56d-e98266b137a0), [Sponge](http://sponge2/1c4ccf74-76a3-4494-a56d-e98266b137a0)
status: failed
|
1.0
|
bigtable.grpc.scanner.ReadRowsAcceptanceTest: many tests failed - Many tests failed at the same time in this package.
* I will close this issue when there are no more failures in this package _and_
there is at least one pass.
* No new issues will be filed for this package until this issue is closed.
* If there are already issues for individual test cases, I will close them when
the corresponding test passes. You can close them earlier, if you prefer, and
I won't reopen them while this issue is still open.
Here are the tests that failed:
* test[invalidNoCommit]
* test[invalidNoCellKeyBeforeCommit]
* test[invalidNoCellKeyBeforeValue]
* test[invalidNewColFamilyMustSpecifyQualifier]
* test[bareCommitImpliesTs=0]
* test[simpleRowWithTimestamp]
* test[missingTimestamp,ImpliedTs=0]
* test[emptyCellValue]
* test[twoUnsplitCells]
* test[twoQualifiers]
* test[twoFamilies]
* test[withLabels]
* test[splitCell,BareCommit]
* test[splitCell]
* test[splitFourWays]
* test[twoSplitCells]
* test[multiQualifierSplits]
* test[multiQualifierMultiSplit]
* test[multiFamilySplit]
* test[invalidNoCommitBetweenRows]
* test[invalidNoCommitAfterFirstRow]
* test[invalidLastRowMissingCommit]
* test[invalidDuplicateRowKey]
* test[invalidNewRowMissingRowKey]
* test[twoRows]
* test[twoRowsImplicitTimestamp]
* test[twoRowsEmptyValue]
* test[twoRows,OneWithMultipleCells]
* test[twoRows,MultipleCells]
* test[twoRows,MultipleCells,MultipleFamilies]
* test[twoRows,FourCells,2Labels]
* test[twoRowsWithSplits,SameTimestamp]
* test[invalidBareReset]
* test[invalidBadReset,NoCommit]
* test[invalidMissingKeyAfterReset]
* test[noDataAfterReset]
* test[simpleReset]
* test[resetToNewVal]
* test[resetToNewQual]
* test[resetWithSplits]
* test[resetTwoCells]
* test[twoResets]
* test[resetThenTwoCells]
* test[resetToNewRow]
* test[resetInBetweenChunks]
* test[invalidResetWithChunk]
* test[invalidCommitWithChunk]
* test[emptyCellChunk]
-----
commit: 81641f77b42506cdd7f0ddcb2192b5f0151aa413
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1c4ccf74-76a3-4494-a56d-e98266b137a0), [Sponge](http://sponge2/1c4ccf74-76a3-4494-a56d-e98266b137a0)
status: failed
|
build
|
bigtable grpc scanner readrowsacceptancetest many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test commit buildurl status failed
| 1
|
95,785
| 27,616,694,522
|
IssuesEvent
|
2023-03-09 19:58:17
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Gradlew build fails on multiple first party packages
|
platform-android plugin t: gradle a: build passed first triage P4
|
## Steps to Reproduce
1. In your project folder, do `cd android ` and then `gradlew build`
**Expected results:**
Successful build
**Actual results:**
Build fails
P.S. - I don't want the mentioned solution. I need to abort on lint error.
<details>
<summary>Cmd Output</summary>
```
> Task :local_auth:lint FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':local_auth:lint'.
> Lint found errors in the project; aborting build.
Fix the issues identified by lint, or add the following to your build script to proceed with errors:
...
android {
lintOptions {
abortOnError false
}
}
...
Errors found:
{MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:3: Error: @android:style/Theme.Material.Dialog.Alert requires API level 21 (current min is 16) [NewApi]
<style name="AlertDialogCustom" parent="@android:style/Theme.Material.Dialog.Alert">
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
{MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:7: Error: android:colorAccent requires API level 21 (current min is 16) [NewApi]
<item name="android:colorAccent">#FF009688</item>
~~~~~~~~~~~~~~~~~~~~~~~~~~
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
```
</details>
<details>
<summary>Flutter Doctor</summary>
```
[√] Flutter (Channel stable, 2.2.2, on Microsoft Windows [Version 10.0.19042.928], locale en-IN)
• Flutter version 2.2.2 at {MyPath}
• Framework revision d79295af24 (6 days ago), 2021-06-11 08:56:01 -0700
• Engine revision 91c9fc8fe0
• Dart version 2.13.3
[√] Android toolchain - develop for Android devices (Android SDK version 30.0.3)
• Android SDK at {MyPath}
• Platform android-30, build-tools 30.0.3
• ANDROID_HOME = {MyPath}
• Java binary at: {MyPath}
• Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174)
• All Android licenses accepted.
[X] Chrome - develop for the web (Cannot find Chrome executable at .\Google\Chrome\Application\chrome.exe)
! Cannot find Chrome. Try setting CHROME_EXECUTABLE to a Chrome executable.
[√] Android Studio
• Android Studio at {MyPath}
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
• android-studio-dir = {MyPath}
• Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174)
[√] Connected device (1 available)
• Edge (web) • edge • web-javascript • Microsoft Edge 92.0.902.9 (unsupported)
! Doctor found issues in 1 category.
```
</details>
|
1.0
|
Gradlew build fails on multiple first party packages - ## Steps to Reproduce
1. In your project folder, do `cd android ` and then `gradlew build`
**Expected results:**
Successful build
**Actual results:**
Build fails
P.S. - I don't want the mentioned solution. I need to abort on lint error.
<details>
<summary>Cmd Output</summary>
```
> Task :local_auth:lint FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':local_auth:lint'.
> Lint found errors in the project; aborting build.
Fix the issues identified by lint, or add the following to your build script to proceed with errors:
...
android {
lintOptions {
abortOnError false
}
}
...
Errors found:
{MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:3: Error: @android:style/Theme.Material.Dialog.Alert requires API level 21 (current min is 16) [NewApi]
<style name="AlertDialogCustom" parent="@android:style/Theme.Material.Dialog.Alert">
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
{MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:7: Error: android:colorAccent requires API level 21 (current min is 16) [NewApi]
<item name="android:colorAccent">#FF009688</item>
~~~~~~~~~~~~~~~~~~~~~~~~~~
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
```
</details>
<details>
<summary>Flutter Doctor</summary>
```
[√] Flutter (Channel stable, 2.2.2, on Microsoft Windows [Version 10.0.19042.928], locale en-IN)
• Flutter version 2.2.2 at {MyPath}
• Framework revision d79295af24 (6 days ago), 2021-06-11 08:56:01 -0700
• Engine revision 91c9fc8fe0
• Dart version 2.13.3
[√] Android toolchain - develop for Android devices (Android SDK version 30.0.3)
• Android SDK at {MyPath}
• Platform android-30, build-tools 30.0.3
• ANDROID_HOME = {MyPath}
• Java binary at: {MyPath}
• Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174)
• All Android licenses accepted.
[X] Chrome - develop for the web (Cannot find Chrome executable at .\Google\Chrome\Application\chrome.exe)
! Cannot find Chrome. Try setting CHROME_EXECUTABLE to a Chrome executable.
[√] Android Studio
• Android Studio at {MyPath}
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
• android-studio-dir = {MyPath}
• Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174)
[√] Connected device (1 available)
• Edge (web) • edge • web-javascript • Microsoft Edge 92.0.902.9 (unsupported)
! Doctor found issues in 1 category.
```
</details>
|
build
|
gradlew build fails on multiple first party packages steps to reproduce in your project folder do cd android and then gradlew build expected results successful build actual results build fails p s i don t want the mentioned solution i need to abort on lint error cmd output task local auth lint failed failure build failed with an exception what went wrong execution failed for task local auth lint lint found errors in the project aborting build fix the issues identified by lint or add the following to your build script to proceed with errors android lintoptions abortonerror false errors found mypath pub cache hosted pub dartlang org local auth android src main res values styles xml error android style theme material dialog alert requires api level current min is mypath pub cache hosted pub dartlang org local auth android src main res values styles xml error android coloraccent requires api level current min is try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights flutter doctor flutter channel stable on microsoft windows locale en in • flutter version at mypath • framework revision days ago • engine revision • dart version android toolchain develop for android devices android sdk version • android sdk at mypath • platform android build tools • android home mypath • java binary at mypath • java version openjdk runtime environment build • all android licenses accepted chrome develop for the web cannot find chrome executable at google chrome application chrome exe cannot find chrome try setting chrome executable to a chrome executable android studio • android studio at mypath • flutter plugin can be installed from • dart plugin can be installed from • android studio dir mypath • java version openjdk runtime environment build connected device available • edge web • edge • web javascript • microsoft edge unsupported doctor found issues in category
| 1
|
199,259
| 6,987,368,868
|
IssuesEvent
|
2017-12-14 08:58:51
|
sunpy/sunpy
|
https://api.github.com/repos/sunpy/sunpy
|
opened
|
Decorator to be able to prepend and/or append to doc strings
|
Effort Low Feature Request Package Novice Priority Low Refactoring
|
Once #2369 is merged in we have under `sunpy/sun/sun.py`, a decorator (called `add_parameter_docs`) that we use to append to the functions under `sun.py` as they all take the same parameter input.
We would like this to moved under to `sunpy/util/decorators.py` where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and/or append argument to the decorator that would add it to a decorated function's docstring. Maybe `.format` on the docstring if a dictionary is provided as an argument to the decorator.
|
1.0
|
Decorator to be able to prepend and/or append to doc strings - Once #2369 is merged in we have under `sunpy/sun/sun.py`, a decorator (called `add_parameter_docs`) that we use to append to the functions under `sun.py` as they all take the same parameter input.
We would like this to moved under to `sunpy/util/decorators.py` where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and/or append argument to the decorator that would add it to a decorated function's docstring. Maybe `.format` on the docstring if a dictionary is provided as an argument to the decorator.
|
non_build
|
decorator to be able to prepend and or append to doc strings once is merged in we have under sunpy sun sun py a decorator called add parameter docs that we use to append to the functions under sun py as they all take the same parameter input we would like this to moved under to sunpy util decorators py where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and or append argument to the decorator that would add it to a decorated function s docstring maybe format on the docstring if a dictionary is provided as an argument to the decorator
| 0
|
92,094
| 26,578,723,822
|
IssuesEvent
|
2023-01-22 06:15:24
|
runatlantis/atlantis
|
https://api.github.com/repos/runatlantis/atlantis
|
closed
|
Deprecate `set-output` in favor of `$GITHUB_OUTPUT`
|
bug build github-actions
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you!
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request.
* If you are interested in working on this issue or have submitted a pull request, please leave a comment.
<!--- Thank you for keeping this note for the community --->
---
<!--- When filing a bug, please include the following headings if possible. Any example text in this template can be deleted. --->
### Overview of the Issue
<!--- Please describe the issue you are having and how you encountered the problem. --->
> To avoid untrusted logged data to use save-stateand set-output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output.
https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/#examples
### Reproduction Steps
<!---
In order to effectively and quickly resolve the issue, please provide exact steps that allow us the reproduce the problem. If no steps are provided, then it will likely take longer to get the issue resolved.
--->
n/a
### Logs
<!---
Provide log files from Atlantis server
<details>
<summary>Logs</summary>
```
log output
```
</details>
--->
n/a
### Environment details
<!---
If not already included, please provide the following:
- Atlantis version:
- If not running the latest Atlantis version have you tried to reproduce this issue on the latest version:
- Atlantis flags:
Atlantis server-side config file:
```
# config file
```
Repo `atlantis.yaml` file:
```
# config file
```
Any other information you can provide about the environment/deployment.
--->
n/a
### Additional Context
<!---
Additional context on the problem. Docs, links to blogs, or other material that lead you to discover this issue or were helpful in troubleshooting the issue.
--->
See link above
I only see it `set-output` in one location
https://github.com/runatlantis/atlantis/blob/d620d5de98b0178a16429fa89095507109b4a815/.github/workflows/release.yml#L35
I do not see `save-state` anywhere.
|
1.0
|
Deprecate `set-output` in favor of `$GITHUB_OUTPUT` - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you!
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request.
* If you are interested in working on this issue or have submitted a pull request, please leave a comment.
<!--- Thank you for keeping this note for the community --->
---
<!--- When filing a bug, please include the following headings if possible. Any example text in this template can be deleted. --->
### Overview of the Issue
<!--- Please describe the issue you are having and how you encountered the problem. --->
> To avoid untrusted logged data to use save-stateand set-output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output.
https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/#examples
### Reproduction Steps
<!---
In order to effectively and quickly resolve the issue, please provide exact steps that allow us the reproduce the problem. If no steps are provided, then it will likely take longer to get the issue resolved.
--->
n/a
### Logs
<!---
Provide log files from Atlantis server
<details>
<summary>Logs</summary>
```
log output
```
</details>
--->
n/a
### Environment details
<!---
If not already included, please provide the following:
- Atlantis version:
- If not running the latest Atlantis version have you tried to reproduce this issue on the latest version:
- Atlantis flags:
Atlantis server-side config file:
```
# config file
```
Repo `atlantis.yaml` file:
```
# config file
```
Any other information you can provide about the environment/deployment.
--->
n/a
### Additional Context
<!---
Additional context on the problem. Docs, links to blogs, or other material that lead you to discover this issue or were helpful in troubleshooting the issue.
--->
See link above
I only see it `set-output` in one location
https://github.com/runatlantis/atlantis/blob/d620d5de98b0178a16429fa89095507109b4a815/.github/workflows/release.yml#L35
I do not see `save-state` anywhere.
|
build
|
deprecate set output in favor of github output community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request searching for pre existing feature requests helps us consolidate datapoints for identical requirements into a single place thank you please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment overview of the issue to avoid untrusted logged data to use save stateand set output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output reproduction steps in order to effectively and quickly resolve the issue please provide exact steps that allow us the reproduce the problem if no steps are provided then it will likely take longer to get the issue resolved n a logs provide log files from atlantis server logs log output n a environment details if not already included please provide the following atlantis version if not running the latest atlantis version have you tried to reproduce this issue on the latest version atlantis flags atlantis server side config file config file repo atlantis yaml file config file any other information you can provide about the environment deployment n a additional context additional context on the problem docs links to blogs or other material that lead you to discover this issue or were helpful in troubleshooting the issue see link above i only see it set output in one location i do not see save state anywhere
| 1
|
437,013
| 12,558,321,970
|
IssuesEvent
|
2020-06-07 15:34:02
|
tensorfork/tensorfork
|
https://api.github.com/repos/tensorfork/tensorfork
|
opened
|
BigGAN: D data augmentations per Zhao et al 2020
|
enhancement good first issue priority: high
|
The new paper ["Image Augmentations for GAN Training"](https://arxiv.org/abs/2006.02595#google), Zhao et al 2020b, reports:
> Data augmentations have been widely studied to improve the accuracy and robustness of classifiers. However, the potential of image augmentation in improving GAN models for image synthesis has not been thoroughly investigated in previous studies. In this work, we systematically study the effectiveness of various existing augmentation techniques for GAN training in a variety of settings. We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations, improving the fidelity of the generated images substantially. Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results if we use augmentations on both real and generated images. When this GAN training is combined with other augmentation-based regularization techniques, such as contrastive loss and consistency regularization, the augmentations further improve the quality of generated images. We provide new state-of-the-art results for conditional generation on CIFAR-10 with both consistency loss and contrastive loss as additional regularizations.
This follows up Zhao et al 2020a, ["Improved Consistency Regularization for GANs"](https://arxiv.org/abs/2002.04724), which we've discussed at length before.
The major finding of 2020b here is two-part:
1. *no* data augmentation helps when applied *only* to real samples, and they often hurt, a lot!
](https://user-images.githubusercontent.com/352559/83972469-4abe9000-a8ae-11ea-8f64-1bd23dabfb97.png)
Note that ProGAN/StyleGAN, Mooch BigGAN, compare_gan, and MSG-GAN either do no data augmentations whatsoever, or they apply very limited data augmentation (flip/mirror and/or random cropping) to real images as part of the data loading pipeline of reals - and never inside the training loop where it could be applied to fakes too. We too currently do only random cropping as a hack to deal with non-square (real) images, but we seem to get better results when doing rescale+black-background (on reals).
2. *several* of the same data augmentations help when applied to *both* real & fake samples, and they often help, a lot!
](https://user-images.githubusercontent.com/352559/83972481-6aee4f00-a8ae-11ea-8e3a-8782cbc51535.png)
Even for a vanilla GAN, using the right data augmentation in the right way can improve FID by as much as a quarter to a half, with no other change whatsoever. The only difference is that the augmentation is done inside the training loop to both real & fakes, instead of just fakes. (2020b uses a single data augmentation at a time, applied to the entire minibatch. There seems to be no reason you couldn't use all of the good data augmentations, and pick one at random per iteration to apply to an entire minibatch, to get even larger benefits.)
Specifically, the 'spatial' data augmentations like zooming in/out or shifting the image up/down/left/right (wrapping around) help the most, while data augmentations like adding pixel noise or distorting colors do not do anything on net. (They do not consider flipping or some other possible ones like Sobel distortion, which the SimCLR paper considered.)
More details on the helpful augmentation implementations:
](https://user-images.githubusercontent.com/352559/83972533-aa1ca000-a8ae-11ea-981c-c9b91db00c5f.png)
The intuition from 2020a is that if you do data augmentation only on reals, D will learn that reals are distorted in such a way, while the G samples are undistorted, and so the data augmentation, instead of teaching D more meaningful representations of images, instead misleads D, resulting in 'leakage' where D makes G start to generate distorted samples too. This is visible in generated samples, wastes training, and apparently is bad enough that it eliminates the benefits that data augmentation should have given you. (This may account for why so many GANs omit data augmentation entirely, or use an extremely weaksauce one like flipping - flipping can't result in 'leakage' because usually, a flipped image is just another valid image and is not 'distorted'. If you don't use data augmentation, it can't help you, but it also can't hurt you when you use it wrong.)
But if you do it on both reals & fakes, D no longer learns the wrong thing. It merely gets a much more challenging task that it can learn from better, as it must distinguish real from fake even when the images are distorted in various ways, forcing it to generalize more.
On the downside: while the ablations are fairly thorough and they do many runs so it's unlikely to be the usual GAN bugaboo of very high variance results and they test on meaningful architectures like BigGAN, so the results look promising, the data augmentation is tested only on CIFAR-10, which is a very small and simple data compared to Danbooru, so there's a chance it won't work nearly as well for us as for their vanilla GAN, and the biggest benefits come from adding in BCR and SimCLR (#11 ), which we don't have yet.
---
This is a remarkably simple improvement we should add to BigGAN immediately to test out. Since it doesn't modify the dataset or arch, we can even add it to current runs.
We probably want (in descending order of how much it helps in Figure 3): ZoomIn, ZoomOut, TranslationX, TranslationY, TranslationX+Y, mirroring/flipping, and CutOut. (CutMix is a lot harder to implement since it requires combing multiple images, and it appears to potentially hurt, so we'll skip that one.)
1. ZoomIn is just a random crop and then a resize up to fit
2. ZoomOut resizes the whole image, and then fills in the padding not with 0s, but with the edge of the image. (This makes it different from [`tf.contrib.image.translate`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/contrib/image/translate), which pads with 0s; there is a [`tf.pad`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/pad) which has a 'mirror' option which I think may be what is done here, but we need an image version.)
3. TranslationX: shifts the image by a few pixels left/right; the excess pixels get wrapped around to the other side
4. TranslationY: same thing, up/down, wrapping around
5. TranslationX+Y: shift both left/right and up/down (so can be implemented by calling `TranslationX(TranslationY(image))`)
6. CutOut: superimposes a small black square at a random point in the image, erasing the existing pixels; apparently implemented as [`tfa.image.cutout`](https://www.tensorflow.org/addons/api_docs/python/tfa/image/cutout)
7. mirroring: the usual, something like `tf.image.flip_left_right`
|
1.0
|
BigGAN: D data augmentations per Zhao et al 2020 - The new paper ["Image Augmentations for GAN Training"](https://arxiv.org/abs/2006.02595#google), Zhao et al 2020b, reports:
> Data augmentations have been widely studied to improve the accuracy and robustness of classifiers. However, the potential of image augmentation in improving GAN models for image synthesis has not been thoroughly investigated in previous studies. In this work, we systematically study the effectiveness of various existing augmentation techniques for GAN training in a variety of settings. We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations, improving the fidelity of the generated images substantially. Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results if we use augmentations on both real and generated images. When this GAN training is combined with other augmentation-based regularization techniques, such as contrastive loss and consistency regularization, the augmentations further improve the quality of generated images. We provide new state-of-the-art results for conditional generation on CIFAR-10 with both consistency loss and contrastive loss as additional regularizations.
This follows up Zhao et al 2020a, ["Improved Consistency Regularization for GANs"](https://arxiv.org/abs/2002.04724), which we've discussed at length before.
The major finding of 2020b here is two-part:
1. *no* data augmentation helps when applied *only* to real samples, and they often hurt, a lot!
](https://user-images.githubusercontent.com/352559/83972469-4abe9000-a8ae-11ea-8f64-1bd23dabfb97.png)
Note that ProGAN/StyleGAN, Mooch BigGAN, compare_gan, and MSG-GAN either do no data augmentations whatsoever, or they apply very limited data augmentation (flip/mirror and/or random cropping) to real images as part of the data loading pipeline of reals - and never inside the training loop where it could be applied to fakes too. We too currently do only random cropping as a hack to deal with non-square (real) images, but we seem to get better results when doing rescale+black-background (on reals).
2. *several* of the same data augmentations help when applied to *both* real & fake samples, and they often help, a lot!
](https://user-images.githubusercontent.com/352559/83972481-6aee4f00-a8ae-11ea-8e3a-8782cbc51535.png)
Even for a vanilla GAN, using the right data augmentation in the right way can improve FID by as much as a quarter to a half, with no other change whatsoever. The only difference is that the augmentation is done inside the training loop to both real & fakes, instead of just fakes. (2020b uses a single data augmentation at a time, applied to the entire minibatch. There seems to be no reason you couldn't use all of the good data augmentations, and pick one at random per iteration to apply to an entire minibatch, to get even larger benefits.)
Specifically, the 'spatial' data augmentations like zooming in/out or shifting the image up/down/left/right (wrapping around) help the most, while data augmentations like adding pixel noise or distorting colors do not do anything on net. (They do not consider flipping or some other possible ones like Sobel distortion, which the SimCLR paper considered.)
More details on the helpful augmentation implementations:
](https://user-images.githubusercontent.com/352559/83972533-aa1ca000-a8ae-11ea-981c-c9b91db00c5f.png)
The intuition from 2020a is that if you do data augmentation only on reals, D will learn that reals are distorted in such a way, while the G samples are undistorted, and so the data augmentation, instead of teaching D more meaningful representations of images, instead misleads D, resulting in 'leakage' where D makes G start to generate distorted samples too. This is visible in generated samples, wastes training, and apparently is bad enough that it eliminates the benefits that data augmentation should have given you. (This may account for why so many GANs omit data augmentation entirely, or use an extremely weaksauce one like flipping - flipping can't result in 'leakage' because usually, a flipped image is just another valid image and is not 'distorted'. If you don't use data augmentation, it can't help you, but it also can't hurt you when you use it wrong.)
But if you do it on both reals & fakes, D no longer learns the wrong thing. It merely gets a much more challenging task that it can learn from better, as it must distinguish real from fake even when the images are distorted in various ways, forcing it to generalize more.
On the downside: while the ablations are fairly thorough and they do many runs so it's unlikely to be the usual GAN bugaboo of very high variance results and they test on meaningful architectures like BigGAN, so the results look promising, the data augmentation is tested only on CIFAR-10, which is a very small and simple data compared to Danbooru, so there's a chance it won't work nearly as well for us as for their vanilla GAN, and the biggest benefits come from adding in BCR and SimCLR (#11 ), which we don't have yet.
---
This is a remarkably simple improvement we should add to BigGAN immediately to test out. Since it doesn't modify the dataset or arch, we can even add it to current runs.
We probably want (in descending order of how much it helps in Figure 3): ZoomIn, ZoomOut, TranslationX, TranslationY, TranslationX+Y, mirroring/flipping, and CutOut. (CutMix is a lot harder to implement since it requires combing multiple images, and it appears to potentially hurt, so we'll skip that one.)
1. ZoomIn is just a random crop and then a resize up to fit
2. ZoomOut resizes the whole image, and then fills in the padding not with 0s, but with the edge of the image. (This makes it different from [`tf.contrib.image.translate`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/contrib/image/translate), which pads with 0s; there is a [`tf.pad`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/pad) which has a 'mirror' option which I think may be what is done here, but we need an image version.)
3. TranslationX: shifts the image by a few pixels left/right; the excess pixels get wrapped around to the other side
4. TranslationY: same thing, up/down, wrapping around
5. TranslationX+Y: shift both left/right and up/down (so can be implemented by calling `TranslationX(TranslationY(image))`)
6. CutOut: superimposes a small black square at a random point in the image, erasing the existing pixels; apparently implemented as [`tfa.image.cutout`](https://www.tensorflow.org/addons/api_docs/python/tfa/image/cutout)
7. mirroring: the usual, something like `tf.image.flip_left_right`
|
non_build
|
biggan d data augmentations per zhao et al the new paper zhao et al reports data augmentations have been widely studied to improve the accuracy and robustness of classifiers however the potential of image augmentation in improving gan models for image synthesis has not been thoroughly investigated in previous studies in this work we systematically study the effectiveness of various existing augmentation techniques for gan training in a variety of settings we provide insights and guidelines on how to augment images for both vanilla gans and gans with regularizations improving the fidelity of the generated images substantially surprisingly we find that vanilla gans attain generation quality on par with recent state of the art results if we use augmentations on both real and generated images when this gan training is combined with other augmentation based regularization techniques such as contrastive loss and consistency regularization the augmentations further improve the quality of generated images we provide new state of the art results for conditional generation on cifar with both consistency loss and contrastive loss as additional regularizations this follows up zhao et al which we ve discussed at length before the major finding of here is two part no data augmentation helps when applied only to real samples and they often hurt a lot note that progan stylegan mooch biggan compare gan and msg gan either do no data augmentations whatsoever or they apply very limited data augmentation flip mirror and or random cropping to real images as part of the data loading pipeline of reals and never inside the training loop where it could be applied to fakes too we too currently do only random cropping as a hack to deal with non square real images but we seem to get better results when doing rescale black background on reals several of the same data augmentations help when applied to both real fake samples and they often help a lot even for a vanilla gan using the right data augmentation in the right way can improve fid by as much as a quarter to a half with no other change whatsoever the only difference is that the augmentation is done inside the training loop to both real fakes instead of just fakes uses a single data augmentation at a time applied to the entire minibatch there seems to be no reason you couldn t use all of the good data augmentations and pick one at random per iteration to apply to an entire minibatch to get even larger benefits specifically the spatial data augmentations like zooming in out or shifting the image up down left right wrapping around help the most while data augmentations like adding pixel noise or distorting colors do not do anything on net they do not consider flipping or some other possible ones like sobel distortion which the simclr paper considered more details on the helpful augmentation implementations the intuition from is that if you do data augmentation only on reals d will learn that reals are distorted in such a way while the g samples are undistorted and so the data augmentation instead of teaching d more meaningful representations of images instead misleads d resulting in leakage where d makes g start to generate distorted samples too this is visible in generated samples wastes training and apparently is bad enough that it eliminates the benefits that data augmentation should have given you this may account for why so many gans omit data augmentation entirely or use an extremely weaksauce one like flipping flipping can t result in leakage because usually a flipped image is just another valid image and is not distorted if you don t use data augmentation it can t help you but it also can t hurt you when you use it wrong but if you do it on both reals fakes d no longer learns the wrong thing it merely gets a much more challenging task that it can learn from better as it must distinguish real from fake even when the images are distorted in various ways forcing it to generalize more on the downside while the ablations are fairly thorough and they do many runs so it s unlikely to be the usual gan bugaboo of very high variance results and they test on meaningful architectures like biggan so the results look promising the data augmentation is tested only on cifar which is a very small and simple data compared to danbooru so there s a chance it won t work nearly as well for us as for their vanilla gan and the biggest benefits come from adding in bcr and simclr which we don t have yet this is a remarkably simple improvement we should add to biggan immediately to test out since it doesn t modify the dataset or arch we can even add it to current runs we probably want in descending order of how much it helps in figure zoomin zoomout translationx translationy translationx y mirroring flipping and cutout cutmix is a lot harder to implement since it requires combing multiple images and it appears to potentially hurt so we ll skip that one zoomin is just a random crop and then a resize up to fit zoomout resizes the whole image and then fills in the padding not with but with the edge of the image this makes it different from which pads with there is a which has a mirror option which i think may be what is done here but we need an image version translationx shifts the image by a few pixels left right the excess pixels get wrapped around to the other side translationy same thing up down wrapping around translationx y shift both left right and up down so can be implemented by calling translationx translationy image cutout superimposes a small black square at a random point in the image erasing the existing pixels apparently implemented as mirroring the usual something like tf image flip left right
| 0
|
82,776
| 23,874,693,436
|
IssuesEvent
|
2022-09-07 17:51:52
|
mfem/mfem
|
https://api.github.com/repos/mfem/mfem
|
closed
|
Add support for SUNDIALS v6
|
building WIP libraries todo
|
Version 6 of SUNDIALS made some interface changes that are breaking the integration with MFEM.
Addition: When addressing this, we should also address an issue seen when running `ex16p` when HYPRE is built with GPU support -- there is a repeated registration of the same host pointer which leads to problems.
Fix an issue: `SUNDIALS_DIR` is set by default to `@MFEM_DIR@/../sundials-5.0.0/instdir` which does not work since we try to expand the path with `$(wildcard ...)` before replacing `@MFEM_DIR@` with the real path.
|
1.0
|
Add support for SUNDIALS v6 - Version 6 of SUNDIALS made some interface changes that are breaking the integration with MFEM.
Addition: When addressing this, we should also address an issue seen when running `ex16p` when HYPRE is built with GPU support -- there is a repeated registration of the same host pointer which leads to problems.
Fix an issue: `SUNDIALS_DIR` is set by default to `@MFEM_DIR@/../sundials-5.0.0/instdir` which does not work since we try to expand the path with `$(wildcard ...)` before replacing `@MFEM_DIR@` with the real path.
|
build
|
add support for sundials version of sundials made some interface changes that are breaking the integration with mfem addition when addressing this we should also address an issue seen when running when hypre is built with gpu support there is a repeated registration of the same host pointer which leads to problems fix an issue sundials dir is set by default to mfem dir sundials instdir which does not work since we try to expand the path with wildcard before replacing mfem dir with the real path
| 1
|
17,618
| 6,485,770,224
|
IssuesEvent
|
2017-08-19 13:41:06
|
ElektraInitiative/libelektra
|
https://api.github.com/repos/ElektraInitiative/libelektra
|
closed
|
debian: errors installing+using
|
build low priority
|
The build server finally built https://build.libelektra.org/job/elektra-git-buildpackage-jessie/lastSuccessfulBuild/console
:+1:
Unfortunately, there are errors in installing:
```
Preparing to unpack .../elektra-qt-gui_0.8.19-1.5721_amd64.deb ...
Unpacking elektra-qt-gui (0.8.19-1.5721) ...
dpkg: error processing archive /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb (--unpack):
trying to overwrite '/usr/share/man/man1/kdb-qt-gui.1.gz', which is also in package elektra-bin 0.8.19-1.5721
Processing triggers for man-db (2.7.0.2-5) ...
Errors were encountered while processing:
/var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb
E: Sub-process /usr/bin/dpkg returned an error code (1)
Failed to perform requested operation on package. Trying to recover:
Press Return to continue.
```
and install-docs fails:
```
install-docs --verbose --check /usr/share/doc-base/elektra-doc
Warning in `/usr/share/doc-base/elektra-doc', line 8: file `/usr/share/doc/elektra-doc/html/index.html' does not exist.
Error in `/usr/share/doc-base/elektra-doc', line 8: all `Format' sections are invalid.
/usr/share/doc-base/elektra-doc: Fatal error found, the file won't be registered.
```
and using it with cmake fails, too (can be tested with `ctest -V -R external` which runs `cmake` in the `examples/external/cmake` project):
```
CMake Error at /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake:98 (message):
The imported target "elektra-utility" references the file
"/usr/lib/x86_64-linux-gnu/libelektra-utility.so.0.8.19"
but this file does not exist. Possible reasons include:
* The file was deleted, renamed, or moved to another location.
* An install or uninstall procedure did not complete successfully.
* The installation package was faulty and contained
"/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake"
but not all the files it references.
Call Stack (most recent call first):
/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraConfig.cmake:16 (include)
CMakeLists.txt:5 (find_package)
-- Configuring incomplete, errors occurred
```
Btw. at least aptitude needed a lot of help to upgrade to the new packages.
Can you test on PCS?
Thank you, great job nevertheless. It is very nice to have debug symbols and hopefully also the qt-gui soon ;)
|
1.0
|
debian: errors installing+using - The build server finally built https://build.libelektra.org/job/elektra-git-buildpackage-jessie/lastSuccessfulBuild/console
:+1:
Unfortunately, there are errors in installing:
```
Preparing to unpack .../elektra-qt-gui_0.8.19-1.5721_amd64.deb ...
Unpacking elektra-qt-gui (0.8.19-1.5721) ...
dpkg: error processing archive /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb (--unpack):
trying to overwrite '/usr/share/man/man1/kdb-qt-gui.1.gz', which is also in package elektra-bin 0.8.19-1.5721
Processing triggers for man-db (2.7.0.2-5) ...
Errors were encountered while processing:
/var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb
E: Sub-process /usr/bin/dpkg returned an error code (1)
Failed to perform requested operation on package. Trying to recover:
Press Return to continue.
```
and install-docs fails:
```
install-docs --verbose --check /usr/share/doc-base/elektra-doc
Warning in `/usr/share/doc-base/elektra-doc', line 8: file `/usr/share/doc/elektra-doc/html/index.html' does not exist.
Error in `/usr/share/doc-base/elektra-doc', line 8: all `Format' sections are invalid.
/usr/share/doc-base/elektra-doc: Fatal error found, the file won't be registered.
```
and using it with cmake fails, too (can be tested with `ctest -V -R external` which runs `cmake` in the `examples/external/cmake` project):
```
CMake Error at /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake:98 (message):
The imported target "elektra-utility" references the file
"/usr/lib/x86_64-linux-gnu/libelektra-utility.so.0.8.19"
but this file does not exist. Possible reasons include:
* The file was deleted, renamed, or moved to another location.
* An install or uninstall procedure did not complete successfully.
* The installation package was faulty and contained
"/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake"
but not all the files it references.
Call Stack (most recent call first):
/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraConfig.cmake:16 (include)
CMakeLists.txt:5 (find_package)
-- Configuring incomplete, errors occurred
```
Btw. at least aptitude needed a lot of help to upgrade to the new packages.
Can you test on PCS?
Thank you, great job nevertheless. It is very nice to have debug symbols and hopefully also the qt-gui soon ;)
|
build
|
debian errors installing using the build server finally built unfortunately there are errors in installing preparing to unpack elektra qt gui deb unpacking elektra qt gui dpkg error processing archive var cache apt archives elektra qt gui deb unpack trying to overwrite usr share man kdb qt gui gz which is also in package elektra bin processing triggers for man db errors were encountered while processing var cache apt archives elektra qt gui deb e sub process usr bin dpkg returned an error code failed to perform requested operation on package trying to recover press return to continue and install docs fails install docs verbose check usr share doc base elektra doc warning in usr share doc base elektra doc line file usr share doc elektra doc html index html does not exist error in usr share doc base elektra doc line all format sections are invalid usr share doc base elektra doc fatal error found the file won t be registered and using it with cmake fails too can be tested with ctest v r external which runs cmake in the examples external cmake project cmake error at usr lib linux gnu cmake elektra elektratargetslibelektra cmake message the imported target elektra utility references the file usr lib linux gnu libelektra utility so but this file does not exist possible reasons include the file was deleted renamed or moved to another location an install or uninstall procedure did not complete successfully the installation package was faulty and contained usr lib linux gnu cmake elektra elektratargetslibelektra cmake but not all the files it references call stack most recent call first usr lib linux gnu cmake elektra elektraconfig cmake include cmakelists txt find package configuring incomplete errors occurred btw at least aptitude needed a lot of help to upgrade to the new packages can you test on pcs thank you great job nevertheless it is very nice to have debug symbols and hopefully also the qt gui soon
| 1
|
53,917
| 6,774,515,699
|
IssuesEvent
|
2017-10-27 10:40:31
|
nawissor/stomer
|
https://api.github.com/repos/nawissor/stomer
|
closed
|
Front page banner
|
Design
|
Create a stich in photoshop and include a few images for the banner on the front page. See this howto --> https://digital-photography-school.com/how-to-make-a-photo-collage-banner-for-your-photoblog-or-website/
|
1.0
|
Front page banner - Create a stich in photoshop and include a few images for the banner on the front page. See this howto --> https://digital-photography-school.com/how-to-make-a-photo-collage-banner-for-your-photoblog-or-website/
|
non_build
|
front page banner create a stich in photoshop and include a few images for the banner on the front page see this howto
| 0
|
84,811
| 24,429,745,345
|
IssuesEvent
|
2022-10-06 07:12:25
|
artichoke/artichoke
|
https://api.github.com/repos/artichoke/artichoke
|
closed
|
Implement a build of artichoke that can be used to bootstrap the build with no Ruby dependency
|
A-core A-build A-ruby-core S-speculative
|
Once Core is complete, it should be possible to produce a Core-only build of artichoke that can be used to bootstrap artichoke-core and mruby-sys.
|
1.0
|
Implement a build of artichoke that can be used to bootstrap the build with no Ruby dependency - Once Core is complete, it should be possible to produce a Core-only build of artichoke that can be used to bootstrap artichoke-core and mruby-sys.
|
build
|
implement a build of artichoke that can be used to bootstrap the build with no ruby dependency once core is complete it should be possible to produce a core only build of artichoke that can be used to bootstrap artichoke core and mruby sys
| 1
|
21,135
| 6,986,235,177
|
IssuesEvent
|
2017-12-14 02:09:40
|
openshiftio/openshift.io
|
https://api.github.com/repos/openshiftio/openshift.io
|
opened
|
Pipeline crashes and fails to build
|
area/pipelines SEV2-high team/build-cd type/bug
|
The pipelines fail and require a restart of the jenkins node/pod thing or GKE or something. Once whatever it is gets restarted the build passes. This thing stops working a lot. Why is it so fragile? We need this to be up and running all the time.
Example
PR: https://github.com/fabric8-ui/fabric8-ui/pull/2302
Logs: https://jenkins.cd.test.fabric8.io/job/fabric8-ui/job/fabric8-ui/view/change-requests/job/PR-2302/5/console
|
1.0
|
Pipeline crashes and fails to build - The pipelines fail and require a restart of the jenkins node/pod thing or GKE or something. Once whatever it is gets restarted the build passes. This thing stops working a lot. Why is it so fragile? We need this to be up and running all the time.
Example
PR: https://github.com/fabric8-ui/fabric8-ui/pull/2302
Logs: https://jenkins.cd.test.fabric8.io/job/fabric8-ui/job/fabric8-ui/view/change-requests/job/PR-2302/5/console
|
build
|
pipeline crashes and fails to build the pipelines fail and require a restart of the jenkins node pod thing or gke or something once whatever it is gets restarted the build passes this thing stops working a lot why is it so fragile we need this to be up and running all the time example pr logs
| 1
|
1,329
| 2,729,697,462
|
IssuesEvent
|
2015-04-16 10:16:08
|
AdaptiveMe/adaptive-arp-javascript
|
https://api.github.com/repos/AdaptiveMe/adaptive-arp-javascript
|
closed
|
Push Adaptive Javascript Definitions to Bower Repository
|
build
|
Every Adaptive ARP new version push the Javascript Library to the repo created on #2 with a new tag in order to used for the new applications.
The files has to be:
adaptive.js (generated with ts compiler)
adaptive.min.js (generated with Uglify)
adaptive.min.js.map (generated with Uglify)
adaptive.d.ts
|
1.0
|
Push Adaptive Javascript Definitions to Bower Repository - Every Adaptive ARP new version push the Javascript Library to the repo created on #2 with a new tag in order to used for the new applications.
The files has to be:
adaptive.js (generated with ts compiler)
adaptive.min.js (generated with Uglify)
adaptive.min.js.map (generated with Uglify)
adaptive.d.ts
|
build
|
push adaptive javascript definitions to bower repository every adaptive arp new version push the javascript library to the repo created on with a new tag in order to used for the new applications the files has to be adaptive js generated with ts compiler adaptive min js generated with uglify adaptive min js map generated with uglify adaptive d ts
| 1
|
297,721
| 25,758,644,523
|
IssuesEvent
|
2022-12-08 18:27:07
|
openedx/build-test-release-wg
|
https://api.github.com/repos/openedx/build-test-release-wg
|
closed
|
olive-testing (TC_AUTHOR_45): Advanced problem types
|
bug release testing
|
# Passing Tests
- [x] Blank advanced Problem
- [x] Circuit Schematic Builder
- [x] Custom JavaScript Display and Grading
- [x] Custom Python-Evaluated Input
- [x] Drag and Drop
- [ ] Drag and Drop (Deprecated Version)
- [x] Image Mapped Input
- [x] Math Expression Input
- [x] Problem with Adaptive Hint
- [x] Staff Graded Points
|
1.0
|
olive-testing (TC_AUTHOR_45): Advanced problem types - # Passing Tests
- [x] Blank advanced Problem
- [x] Circuit Schematic Builder
- [x] Custom JavaScript Display and Grading
- [x] Custom Python-Evaluated Input
- [x] Drag and Drop
- [ ] Drag and Drop (Deprecated Version)
- [x] Image Mapped Input
- [x] Math Expression Input
- [x] Problem with Adaptive Hint
- [x] Staff Graded Points
|
non_build
|
olive testing tc author advanced problem types passing tests blank advanced problem circuit schematic builder custom javascript display and grading custom python evaluated input drag and drop drag and drop deprecated version image mapped input math expression input problem with adaptive hint staff graded points
| 0
|
38,291
| 10,166,661,977
|
IssuesEvent
|
2019-08-07 16:18:14
|
sactory/sactory
|
https://api.github.com/repos/sactory/sactory
|
closed
|
Custom events, including `append` and `remove` are not triggered on Firefox
|
bug runtime:builder
|
```html
<div +append=console.log />
```
does not log anything
|
1.0
|
Custom events, including `append` and `remove` are not triggered on Firefox - ```html
<div +append=console.log />
```
does not log anything
|
build
|
custom events including append and remove are not triggered on firefox html does not log anything
| 1
|
209,672
| 23,730,739,875
|
IssuesEvent
|
2022-08-31 01:18:33
|
jiw065/jiw065.github.io
|
https://api.github.com/repos/jiw065/jiw065.github.io
|
opened
|
CVE-2020-11022 (Medium) detected in jquery-3.1.1.min.js
|
security vulnerability
|
## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p>
<p>Path to dependency file: /MyPortfolio/index.html</p>
<p>Path to vulnerable library: /MyPortfolio/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.1.1.min.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11022 (Medium) detected in jquery-3.1.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p>
<p>Path to dependency file: /MyPortfolio/index.html</p>
<p>Path to vulnerable library: /MyPortfolio/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.1.1.min.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file myportfolio index html path to vulnerable library myportfolio index html dependency hierarchy x jquery min js vulnerable library vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
| 0
|
91,257
| 26,339,746,127
|
IssuesEvent
|
2023-01-10 16:46:47
|
dotnet/arcade
|
https://api.github.com/repos/dotnet/arcade
|
closed
|
Build failed: dotnet-arcade-validation-official/main #20230110.1
|
Build Failed
|
Build [#20230110.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2084297) partiallySucceeded
## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded
### Summary
**Finished** - Tue, 10 Jan 2023 15:47:10 GMT
**Duration** - 116 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Promote Arcade to '.NET Eng - Latest' channel
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'runtime' repository was not successful.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'aspnetcore' repository was not successful.
#### Publish using Darc
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/312) - There was a failure in sending the provision message: A timeout occurred while sending request to the remote provider.
### Changes
- [07665182](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/07665182e45fc1aaf5a04ac9fcb73a4053259e5a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230110.1 (#3606)
- [bbb340e6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/bbb340e689757c2320e66379ab5ae2d9dede73e2) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.3 (#3605)
- [1d04b3e1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/1d04b3e194a01a1fa590deeca7f5d2308874c0df) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.2 (#3604)
|
1.0
|
Build failed: dotnet-arcade-validation-official/main #20230110.1 - Build [#20230110.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2084297) partiallySucceeded
## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded
### Summary
**Finished** - Tue, 10 Jan 2023 15:47:10 GMT
**Duration** - 116 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Promote Arcade to '.NET Eng - Latest' channel
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'runtime' repository was not successful.
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'aspnetcore' repository was not successful.
#### Publish using Darc
- :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/312) - There was a failure in sending the provision message: A timeout occurred while sending request to the remote provider.
### Changes
- [07665182](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/07665182e45fc1aaf5a04ac9fcb73a4053259e5a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230110.1 (#3606)
- [bbb340e6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/bbb340e689757c2320e66379ab5ae2d9dede73e2) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.3 (#3605)
- [1d04b3e1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/1d04b3e194a01a1fa590deeca7f5d2308874c0df) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.2 (#3604)
|
build
|
build failed dotnet arcade validation official main build partiallysucceeded warning internal dotnet arcade validation official partiallysucceeded summary finished tue jan gmt duration minutes requested for dotnet bot reason batchedci details promote arcade to net eng latest channel warning the latest build on main branch for the runtime repository was not successful warning the latest build on main branch for the aspnetcore repository was not successful publish using darc warning there was a failure in sending the provision message a timeout occurred while sending request to the remote provider changes dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from build
| 1
|
10,256
| 4,028,973,861
|
IssuesEvent
|
2016-05-18 08:53:05
|
zturtleman/spearmint
|
https://api.github.com/repos/zturtleman/spearmint
|
opened
|
Activate entity regression on nodm7
|
bug gamecode
|
Activate entity regression on [Claustrophobopolis 2003](http://de.ws.q3df.org/map/nodm7/) reportedly caused by https://github.com/zturtleman/mint-arena/compare/f10b0870f6df...116ec08b3843 changes for #220.
|
1.0
|
Activate entity regression on nodm7 - Activate entity regression on [Claustrophobopolis 2003](http://de.ws.q3df.org/map/nodm7/) reportedly caused by https://github.com/zturtleman/mint-arena/compare/f10b0870f6df...116ec08b3843 changes for #220.
|
non_build
|
activate entity regression on activate entity regression on reportedly caused by changes for
| 0
|
86,116
| 10,474,314,009
|
IssuesEvent
|
2019-09-23 14:17:38
|
weaveworks/wks-quickstart-firekube
|
https://api.github.com/repos/weaveworks/wks-quickstart-firekube
|
closed
|
Add a step that shows gitops in action
|
documentation enhancement
|
Once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster.
|
1.0
|
Add a step that shows gitops in action - Once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster.
|
non_build
|
add a step that shows gitops in action once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster
| 0
|
312,256
| 26,852,300,169
|
IssuesEvent
|
2023-02-03 12:02:40
|
galaxyproject/training-material
|
https://api.github.com/repos/galaxyproject/training-material
|
closed
|
Rule Based Uploader: Advanced - Outdated Info
|
bug galaxy-interface galaxy-release-testing
|
Release Testing 23.0
* “Building URLs from Accession Information”, Hands-on: Uniport Outdated Info and Screenshots
- Format can now be found under Download. Format menu point no longer exists
- Page now has an api url generator, which may be a better fit for this tutorial. It is found under the parent section, and is formated: `https://rest.uniprot.org/uniprotkb/stream?format=fasta&query=accession%{identifier}`
* "Apply Rules to Existing Collections", Step 3: Click `Build` should be Click `Upload`
* Outdated Screenshots showing the History
|
1.0
|
Rule Based Uploader: Advanced - Outdated Info - Release Testing 23.0
* “Building URLs from Accession Information”, Hands-on: Uniport Outdated Info and Screenshots
- Format can now be found under Download. Format menu point no longer exists
- Page now has an api url generator, which may be a better fit for this tutorial. It is found under the parent section, and is formated: `https://rest.uniprot.org/uniprotkb/stream?format=fasta&query=accession%{identifier}`
* "Apply Rules to Existing Collections", Step 3: Click `Build` should be Click `Upload`
* Outdated Screenshots showing the History
|
non_build
|
rule based uploader advanced outdated info release testing “building urls from accession information” hands on uniport outdated info and screenshots format can now be found under download format menu point no longer exists page now has an api url generator which may be a better fit for this tutorial it is found under the parent section and is formated apply rules to existing collections step click build should be click upload outdated screenshots showing the history
| 0
|
17,886
| 3,644,828,375
|
IssuesEvent
|
2016-02-15 11:47:20
|
hu-semesterprojekt-genombrowser/Alazen-Middleware
|
https://api.github.com/repos/hu-semesterprojekt-genombrowser/Alazen-Middleware
|
closed
|
Ant File anpassen
|
enhancement testing
|
Die Datei build.xml muss an eure Struktur angepasst werden.
Einer Erklärung, wie Ant funktioniert, ist unter [https://ant.apache.org/manual/tutorial-HelloWorldWithAnt.html] zu finden.
Es fehlt doch:
* Tests bauen und ausführen
* ~~Dokumentation bauen~~
|
1.0
|
Ant File anpassen - Die Datei build.xml muss an eure Struktur angepasst werden.
Einer Erklärung, wie Ant funktioniert, ist unter [https://ant.apache.org/manual/tutorial-HelloWorldWithAnt.html] zu finden.
Es fehlt doch:
* Tests bauen und ausführen
* ~~Dokumentation bauen~~
|
non_build
|
ant file anpassen die datei build xml muss an eure struktur angepasst werden einer erklärung wie ant funktioniert ist unter zu finden es fehlt doch tests bauen und ausführen dokumentation bauen
| 0
|
265,262
| 23,156,535,719
|
IssuesEvent
|
2022-07-29 13:32:07
|
simpleworks-gmbh/staf
|
https://api.github.com/repos/simpleworks-gmbh/staf
|
closed
|
Proxy can't connect to the expected port
|
bug framework testing v0.120.13
|
If one starts several API Testcase, from another Testcase (manually via executeTestcase Method).
The configured port, will be used for all Proxy Server instances., Only the first Proxy will run, but the other proxies (for the APITestcases left) won't run.
That will fail with the error message
2022-07-11 14:34:16.481 ERROR HttpClient (HttpClient.java:103) - can't set up Response.
java.net.ConnectException: Failed to connect to /0.0.0.0:8893
at okhttp3.internal.io.RealConnection.connectSocket(RealConnection.java:187) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.io.RealConnection.buildTunneledConnection(RealConnection.java:151) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.io.RealConnection.connect(RealConnection.java:108) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.findConnection(StreamAllocation.java:187) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.findHealthyConnection(StreamAllocation.java:123) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.newStream(StreamAllocation.java:93) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.HttpEngine.connect(HttpEngine.java:296) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.HttpEngine.sendRequest(HttpEngine.java:248) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.getResponse(RealCall.java:243) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall$ApplicationInterceptorChain.proceed(RealCall.java:201) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:163) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.execute(RealCall.java:57) ~[okhttp-3.3.1.jar:?]
at de.simpleworks.staf.framework.api.httpclient.HttpClient.doRequest(HttpClient.java:89) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.doRequest(APITestCase.java:323) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.checkRequest(APITestCase.java:258) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.executeTestStep(APITestCase.java:498) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.commons.TestCase.executeTestcase(TestCase.java:325) [framework-0.120.9.jar:?]
at package.executeTestcase(TestCaseIntegration.kt:48) [main/:?]
at package.step5(TEST_CASE_ID.kt:112) [test/:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) [junit-4.13.1.jar:4.13.1]
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) [junit-4.13.1.jar:4.13.1]
at de.simpleworks.staf.module.junit4.STAFInvokeMethod.evaluate(STAFInvokeMethod.java:88) [junit4-module-0.120.9.jar:?]
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.run(ParentRunner.java:413) [junit-4.13.1.jar:4.13.1]
at org.junit.runner.JUnitCore.run(JUnitCore.java:137) [junit-4.13.1.jar:4.13.1]
at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69) [junit-rt.jar:?]
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) [junit-rt.jar:?]
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) [junit-rt.jar:?]
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) [junit-rt.jar:?]
|
1.0
|
Proxy can't connect to the expected port - If one starts several API Testcase, from another Testcase (manually via executeTestcase Method).
The configured port, will be used for all Proxy Server instances., Only the first Proxy will run, but the other proxies (for the APITestcases left) won't run.
That will fail with the error message
2022-07-11 14:34:16.481 ERROR HttpClient (HttpClient.java:103) - can't set up Response.
java.net.ConnectException: Failed to connect to /0.0.0.0:8893
at okhttp3.internal.io.RealConnection.connectSocket(RealConnection.java:187) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.io.RealConnection.buildTunneledConnection(RealConnection.java:151) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.io.RealConnection.connect(RealConnection.java:108) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.findConnection(StreamAllocation.java:187) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.findHealthyConnection(StreamAllocation.java:123) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.StreamAllocation.newStream(StreamAllocation.java:93) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.HttpEngine.connect(HttpEngine.java:296) ~[okhttp-3.3.1.jar:?]
at okhttp3.internal.http.HttpEngine.sendRequest(HttpEngine.java:248) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.getResponse(RealCall.java:243) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall$ApplicationInterceptorChain.proceed(RealCall.java:201) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:163) ~[okhttp-3.3.1.jar:?]
at okhttp3.RealCall.execute(RealCall.java:57) ~[okhttp-3.3.1.jar:?]
at de.simpleworks.staf.framework.api.httpclient.HttpClient.doRequest(HttpClient.java:89) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.doRequest(APITestCase.java:323) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.checkRequest(APITestCase.java:258) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.api.APITestCase.executeTestStep(APITestCase.java:498) [framework-0.120.9.jar:?]
at de.simpleworks.staf.framework.elements.commons.TestCase.executeTestcase(TestCase.java:325) [framework-0.120.9.jar:?]
at package.executeTestcase(TestCaseIntegration.kt:48) [main/:?]
at package.step5(TEST_CASE_ID.kt:112) [test/:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) [junit-4.13.1.jar:4.13.1]
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) [junit-4.13.1.jar:4.13.1]
at de.simpleworks.staf.module.junit4.STAFInvokeMethod.evaluate(STAFInvokeMethod.java:88) [junit4-module-0.120.9.jar:?]
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1]
at org.junit.runners.ParentRunner.run(ParentRunner.java:413) [junit-4.13.1.jar:4.13.1]
at org.junit.runner.JUnitCore.run(JUnitCore.java:137) [junit-4.13.1.jar:4.13.1]
at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69) [junit-rt.jar:?]
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) [junit-rt.jar:?]
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) [junit-rt.jar:?]
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) [junit-rt.jar:?]
|
non_build
|
proxy can t connect to the expected port if one starts several api testcase from another testcase manually via executetestcase method the configured port will be used for all proxy server instances only the first proxy will run but the other proxies for the apitestcases left won t run that will fail with the error message error httpclient httpclient java can t set up response java net connectexception failed to connect to at internal io realconnection connectsocket realconnection java at internal io realconnection buildtunneledconnection realconnection java at internal io realconnection connect realconnection java at internal http streamallocation findconnection streamallocation java at internal http streamallocation findhealthyconnection streamallocation java at internal http streamallocation newstream streamallocation java at internal http httpengine connect httpengine java at internal http httpengine sendrequest httpengine java at realcall getresponse realcall java at realcall applicationinterceptorchain proceed realcall java at realcall getresponsewithinterceptorchain realcall java at realcall execute realcall java at de simpleworks staf framework api httpclient httpclient dorequest httpclient java at de simpleworks staf framework elements api apitestcase dorequest apitestcase java at de simpleworks staf framework elements api apitestcase checkrequest apitestcase java at de simpleworks staf framework elements api apitestcase executeteststep apitestcase java at de simpleworks staf framework elements commons testcase executetestcase testcase java at package executetestcase testcaseintegration kt at package test case id kt at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at de simpleworks staf module stafinvokemethod evaluate stafinvokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runner junitcore run junitcore java at com intellij startrunnerwithargs java at com intellij rt junit ideatestrunner repeater startrunnerwithargs ideatestrunner java at com intellij rt junit junitstarter preparestreamsandstart junitstarter java at com intellij rt junit junitstarter main junitstarter java
| 0
|
159,537
| 20,068,329,610
|
IssuesEvent
|
2022-02-04 01:12:23
|
panasalap/frameworks_av
|
https://api.github.com/repos/panasalap/frameworks_av
|
closed
|
CVE-2020-0169 (Medium) detected in avandroid-10.0.0_r46 - autoclosed
|
security vulnerability
|
## CVE-2020-0169 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r46</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/frameworks_av/commit/0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b">0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/media/extractors/midi/MidiExtractor.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In RTTTL_Event of eas_rtttl.c, there is possible resource exhaustion due to a missing bounds check. This could lead to remote denial of service with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-123700383
<p>Publish Date: 2020-06-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0169>CVE-2020-0169</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37">https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37</a></p>
<p>Release Date: 2020-06-11</p>
<p>Fix Resolution: android-10.0.0_r37</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-0169 (Medium) detected in avandroid-10.0.0_r46 - autoclosed - ## CVE-2020-0169 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r46</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/frameworks_av/commit/0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b">0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/media/extractors/midi/MidiExtractor.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In RTTTL_Event of eas_rtttl.c, there is possible resource exhaustion due to a missing bounds check. This could lead to remote denial of service with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-123700383
<p>Publish Date: 2020-06-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0169>CVE-2020-0169</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37">https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37</a></p>
<p>Release Date: 2020-06-11</p>
<p>Fix Resolution: android-10.0.0_r37</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve medium detected in avandroid autoclosed cve medium severity vulnerability vulnerable library avandroid library home page a href found in head commit a href found in base branch master vulnerable source files media extractors midi midiextractor cpp vulnerability details in rtttl event of eas rtttl c there is possible resource exhaustion due to a missing bounds check this could lead to remote denial of service with no additional execution privileges needed user interaction is needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with whitesource
| 0
|
458,045
| 13,167,428,812
|
IssuesEvent
|
2020-08-11 10:15:15
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
[0.9.0.0 - dev-23] Modding - Server crash on config manipulation
|
Category: Modkit Priority: Medium
|
While developing DiscordLink is noticed the server would crash when any element (that is not a collection) was changed in the config via the server UI.
For booleans, this happens right away and for other types it crashes upon hitting enter while the input field has focus.
This is a huge issue for the 9.0 update of the mod as we recommend our users to do configuration via the server interface.
I've made a minimal project that reproduces the issue. I used DLLs fetched from the temp directory of the dev-23 server for this.
[EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033963/EcoServerCrashMVP.zip)
Here is the DLL output of the above project (compiled with debug flags)
[EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033974/EcoServerCrashMVP.zip)
Steps to reproduce:
1. Add the mod DLL to the server.
2. Start the server.
3. In the server UI, navigate to EcoServerCrashMVP.MVP (probably did something odd to get that name but that shouldn't be related to the crash).
4. Change the value of the "The Bool" field.
5. Observe NullReferenceException outside of the plugin code.
--------------------------
Exception + Stack trace
--------------------------
```
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Server.UI.ServerMainForm.<>c__DisplayClass7_0.<BuildEditor>b__0(Object o, PropertyValueChangedEventArgs e)
at System.Windows.Forms.PropertyGrid.OnPropertyValueChanged(PropertyValueChangedEventArgs e)
at System.Windows.Forms.PropertyGrid.OnPropertyValueSet(GridItem changedItem, Object oldValue)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.CommitValue(GridEntry ipeCur, Object value, Boolean closeDropDown)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.ProcessEnumUpAndDown(GridEntry gridEntry, Keys keyCode, Boolean closeDropDown)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.OnEditKeyDown(Object sender, KeyEventArgs ke)
at System.Windows.Forms.Control.OnKeyDown(KeyEventArgs e)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.OnKeyDown(KeyEventArgs ke)
at System.Windows.Forms.Control.ProcessKeyEventArgs(Message& m)
at System.Windows.Forms.Control.ProcessKeyMessage(Message& m)
at System.Windows.Forms.Control.WndProc(Message& m)
at System.Windows.Forms.TextBox.WndProc(Message& m)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.WndProc(Message& m)
at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
```
--------------------------
Plugin code reference
--------------------------
```
using System.ComponentModel;
using Eco.Core.Plugins;
using Eco.Core.Plugins.Interfaces;
using Eco.Core.Utils;
namespace EcoServerCrashMVP
{
public class MVP : IModKitPlugin, IConfigurablePlugin, IInitializablePlugin
{
public ThreadSafeAction<object, string> ParamChanged { get; set; }
private PluginConfig<TinyConfig> _configOptions;
public string GetStatus()
{
return _configOptions.Config.TheBool ? "Bool is True" : "Bool is False";
}
public IPluginConfig PluginConfig
{
get { return _configOptions; }
}
public object GetEditObject()
{
return _configOptions.Config;
}
public void OnEditObjectChanged(object o, string param)
{
}
public void Initialize(TimedTask timer)
{
_configOptions = new PluginConfig<TinyConfig>("CrashMVP");
}
}
public class TinyConfig
{
[Description("Flip to crash!"), Category("Crashing")]
public bool TheBool { get; set; } = false;
}
}
```
|
1.0
|
[0.9.0.0 - dev-23] Modding - Server crash on config manipulation - While developing DiscordLink is noticed the server would crash when any element (that is not a collection) was changed in the config via the server UI.
For booleans, this happens right away and for other types it crashes upon hitting enter while the input field has focus.
This is a huge issue for the 9.0 update of the mod as we recommend our users to do configuration via the server interface.
I've made a minimal project that reproduces the issue. I used DLLs fetched from the temp directory of the dev-23 server for this.
[EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033963/EcoServerCrashMVP.zip)
Here is the DLL output of the above project (compiled with debug flags)
[EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033974/EcoServerCrashMVP.zip)
Steps to reproduce:
1. Add the mod DLL to the server.
2. Start the server.
3. In the server UI, navigate to EcoServerCrashMVP.MVP (probably did something odd to get that name but that shouldn't be related to the crash).
4. Change the value of the "The Bool" field.
5. Observe NullReferenceException outside of the plugin code.
--------------------------
Exception + Stack trace
--------------------------
```
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Server.UI.ServerMainForm.<>c__DisplayClass7_0.<BuildEditor>b__0(Object o, PropertyValueChangedEventArgs e)
at System.Windows.Forms.PropertyGrid.OnPropertyValueChanged(PropertyValueChangedEventArgs e)
at System.Windows.Forms.PropertyGrid.OnPropertyValueSet(GridItem changedItem, Object oldValue)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.CommitValue(GridEntry ipeCur, Object value, Boolean closeDropDown)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.ProcessEnumUpAndDown(GridEntry gridEntry, Keys keyCode, Boolean closeDropDown)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.OnEditKeyDown(Object sender, KeyEventArgs ke)
at System.Windows.Forms.Control.OnKeyDown(KeyEventArgs e)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.OnKeyDown(KeyEventArgs ke)
at System.Windows.Forms.Control.ProcessKeyEventArgs(Message& m)
at System.Windows.Forms.Control.ProcessKeyMessage(Message& m)
at System.Windows.Forms.Control.WndProc(Message& m)
at System.Windows.Forms.TextBox.WndProc(Message& m)
at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.WndProc(Message& m)
at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
```
--------------------------
Plugin code reference
--------------------------
```
using System.ComponentModel;
using Eco.Core.Plugins;
using Eco.Core.Plugins.Interfaces;
using Eco.Core.Utils;
namespace EcoServerCrashMVP
{
public class MVP : IModKitPlugin, IConfigurablePlugin, IInitializablePlugin
{
public ThreadSafeAction<object, string> ParamChanged { get; set; }
private PluginConfig<TinyConfig> _configOptions;
public string GetStatus()
{
return _configOptions.Config.TheBool ? "Bool is True" : "Bool is False";
}
public IPluginConfig PluginConfig
{
get { return _configOptions; }
}
public object GetEditObject()
{
return _configOptions.Config;
}
public void OnEditObjectChanged(object o, string param)
{
}
public void Initialize(TimedTask timer)
{
_configOptions = new PluginConfig<TinyConfig>("CrashMVP");
}
}
public class TinyConfig
{
[Description("Flip to crash!"), Category("Crashing")]
public bool TheBool { get; set; } = false;
}
}
```
|
non_build
|
modding server crash on config manipulation while developing discordlink is noticed the server would crash when any element that is not a collection was changed in the config via the server ui for booleans this happens right away and for other types it crashes upon hitting enter while the input field has focus this is a huge issue for the update of the mod as we recommend our users to do configuration via the server interface i ve made a minimal project that reproduces the issue i used dlls fetched from the temp directory of the dev server for this here is the dll output of the above project compiled with debug flags steps to reproduce add the mod dll to the server start the server in the server ui navigate to ecoservercrashmvp mvp probably did something odd to get that name but that shouldn t be related to the crash change the value of the the bool field observe nullreferenceexception outside of the plugin code exception stack trace system nullreferenceexception object reference not set to an instance of an object at eco server ui servermainform c b object o propertyvaluechangedeventargs e at system windows forms propertygrid onpropertyvaluechanged propertyvaluechangedeventargs e at system windows forms propertygrid onpropertyvalueset griditem changeditem object oldvalue at system windows forms propertygridinternal propertygridview commitvalue gridentry ipecur object value boolean closedropdown at system windows forms propertygridinternal propertygridview processenumupanddown gridentry gridentry keys keycode boolean closedropdown at system windows forms propertygridinternal propertygridview oneditkeydown object sender keyeventargs ke at system windows forms control onkeydown keyeventargs e at system windows forms propertygridinternal propertygridview gridviewedit onkeydown keyeventargs ke at system windows forms control processkeyeventargs message m at system windows forms control processkeymessage message m at system windows forms control wndproc message m at system windows forms textbox wndproc message m at system windows forms propertygridinternal propertygridview gridviewedit wndproc message m at system windows forms nativewindow callback intptr hwnd msg intptr wparam intptr lparam plugin code reference using system componentmodel using eco core plugins using eco core plugins interfaces using eco core utils namespace ecoservercrashmvp public class mvp imodkitplugin iconfigurableplugin iinitializableplugin public threadsafeaction paramchanged get set private pluginconfig configoptions public string getstatus return configoptions config thebool bool is true bool is false public ipluginconfig pluginconfig get return configoptions public object geteditobject return configoptions config public void oneditobjectchanged object o string param public void initialize timedtask timer configoptions new pluginconfig crashmvp public class tinyconfig public bool thebool get set false
| 0
|
5,081
| 7,603,776,161
|
IssuesEvent
|
2018-04-29 17:59:29
|
hawthorne-git/hawthorne-admin
|
https://api.github.com/repos/hawthorne-git/hawthorne-admin
|
opened
|
Checkup Job
|
Bug Priority - High Effort Required - Medium v2 Requirement
|
Validate that each product has an image, and its image is valid (EX: for re-sizing)
|
1.0
|
Checkup Job - Validate that each product has an image, and its image is valid (EX: for re-sizing)
|
non_build
|
checkup job validate that each product has an image and its image is valid ex for re sizing
| 0
|
104,415
| 4,211,085,495
|
IssuesEvent
|
2016-06-29 12:25:29
|
play2-maven-plugin/play2-maven-plugin
|
https://api.github.com/repos/play2-maven-plugin/play2-maven-plugin
|
closed
|
Upgrade minimum supported Maven version to 3.0
|
Component-Maven-Plugin Priority-Medium
|
It will be required for running Play! in development mode with hot reloading (issue #6).
|
1.0
|
Upgrade minimum supported Maven version to 3.0 - It will be required for running Play! in development mode with hot reloading (issue #6).
|
non_build
|
upgrade minimum supported maven version to it will be required for running play in development mode with hot reloading issue
| 0
|
85,167
| 10,429,827,736
|
IssuesEvent
|
2019-09-17 04:26:56
|
StarChart-Labs/operations
|
https://api.github.com/repos/StarChart-Labs/operations
|
opened
|
Standard Release Scope Handling
|
documentation enhancement
|
Document a standard practice for determining the scope a change requires in release versioning (major, minor, micro), and a standardized implementation for specifying it.
Currently, flare-plugins defines a decent starting implementation where there are labels for major, minor, and micro release scopes, and the "next-release" milestone always includes the minimum scope requried
|
1.0
|
Standard Release Scope Handling - Document a standard practice for determining the scope a change requires in release versioning (major, minor, micro), and a standardized implementation for specifying it.
Currently, flare-plugins defines a decent starting implementation where there are labels for major, minor, and micro release scopes, and the "next-release" milestone always includes the minimum scope requried
|
non_build
|
standard release scope handling document a standard practice for determining the scope a change requires in release versioning major minor micro and a standardized implementation for specifying it currently flare plugins defines a decent starting implementation where there are labels for major minor and micro release scopes and the next release milestone always includes the minimum scope requried
| 0
|
52,114
| 12,878,892,334
|
IssuesEvent
|
2020-07-11 19:00:11
|
PowerShell/PowerShell
|
https://api.github.com/repos/PowerShell/PowerShell
|
closed
|
Azure Pipelines clock issues
|
Area-Build Issue-Question Resolution-Answered Resolution-External
|
In build 55800 the clock keeps jumping back and forth about 16 hours:
2020-06-18T16:46:46.1993748Z
2020-06-19T09:28:05.5534482Z
https://dev.azure.com/powershell/2972bb5c-f20c-4a60-8bd9-00ffe9987edc/_apis/build/builds/55800/logs/35
|
1.0
|
Azure Pipelines clock issues - In build 55800 the clock keeps jumping back and forth about 16 hours:
2020-06-18T16:46:46.1993748Z
2020-06-19T09:28:05.5534482Z
https://dev.azure.com/powershell/2972bb5c-f20c-4a60-8bd9-00ffe9987edc/_apis/build/builds/55800/logs/35
|
build
|
azure pipelines clock issues in build the clock keeps jumping back and forth about hours
| 1
|
27,989
| 8,058,156,988
|
IssuesEvent
|
2018-08-02 17:34:22
|
PowerShell/PowerShell
|
https://api.github.com/repos/PowerShell/PowerShell
|
closed
|
Microsoft.PowerShell.SDK version 6.1.0-preview.4 fails at InitialSessionState.CreateDefault()
|
Area-Build
|
I'm unable to use preview.4 of the SDK with my test project. preview.3 worked fine. The test project targets .net core 2.1.
Steps to reproduce
------------------
1. Clone https://github.com/jherby2k/AudioWorks.git.
1. Run the tests in the AudioWorks.Commands.Tests project. All should pass.
1. Change the project to reference Microsoft.PowerShell.SDK version 6.1.0-preview.4 (from preview.3)
1. Run the tests again. Most will fail.
Expected behavior
-----------------
all tests pass
Actual behavior
---------------
> One or more errors occurred. (Cannot load PowerShell snap-in Microsoft.PowerShell.Diagnostics because of the following error: Could not load file or assembly '..\Source\Repos\AudioWorks\AudioWorks\tests\AudioWorks.Commands.Tests\bin\Release-Windows\netcoreapp2.1\Microsoft.PowerShell.Commands'. The system cannot find the file specified.
|
1.0
|
Microsoft.PowerShell.SDK version 6.1.0-preview.4 fails at InitialSessionState.CreateDefault() - I'm unable to use preview.4 of the SDK with my test project. preview.3 worked fine. The test project targets .net core 2.1.
Steps to reproduce
------------------
1. Clone https://github.com/jherby2k/AudioWorks.git.
1. Run the tests in the AudioWorks.Commands.Tests project. All should pass.
1. Change the project to reference Microsoft.PowerShell.SDK version 6.1.0-preview.4 (from preview.3)
1. Run the tests again. Most will fail.
Expected behavior
-----------------
all tests pass
Actual behavior
---------------
> One or more errors occurred. (Cannot load PowerShell snap-in Microsoft.PowerShell.Diagnostics because of the following error: Could not load file or assembly '..\Source\Repos\AudioWorks\AudioWorks\tests\AudioWorks.Commands.Tests\bin\Release-Windows\netcoreapp2.1\Microsoft.PowerShell.Commands'. The system cannot find the file specified.
|
build
|
microsoft powershell sdk version preview fails at initialsessionstate createdefault i m unable to use preview of the sdk with my test project preview worked fine the test project targets net core steps to reproduce clone run the tests in the audioworks commands tests project all should pass change the project to reference microsoft powershell sdk version preview from preview run the tests again most will fail expected behavior all tests pass actual behavior one or more errors occurred cannot load powershell snap in microsoft powershell diagnostics because of the following error could not load file or assembly source repos audioworks audioworks tests audioworks commands tests bin release windows microsoft powershell commands the system cannot find the file specified
| 1
|
11,132
| 4,892,462,668
|
IssuesEvent
|
2016-11-18 19:47:59
|
numenta/nupic.core
|
https://api.github.com/repos/numenta/nupic.core
|
closed
|
Reuse VERSION file in Doxygen
|
priority:4 triage type:build type:cleanup type:deployment
|
Currently, the version is specified in both places. The Doxygen file should reuse the VERSION file.
|
1.0
|
Reuse VERSION file in Doxygen - Currently, the version is specified in both places. The Doxygen file should reuse the VERSION file.
|
build
|
reuse version file in doxygen currently the version is specified in both places the doxygen file should reuse the version file
| 1
|
78,676
| 22,345,789,526
|
IssuesEvent
|
2022-06-15 07:40:50
|
isl-org/Open3D
|
https://api.github.com/repos/isl-org/Open3D
|
opened
|
Summarize the issue and your environments (e.g., "Cannot compile on Ubuntu 20.04 with OpenBLAS")
|
build/install issue
|
### Checklist
- [X] I have searched for [similar issues](https://github.com/isl-org/Open3D/issues).
- [X] For Python issues, I have tested with the [latest development wheel](http://www.open3d.org/docs/latest/getting_started.html#development-version-pip).
- [X] I have checked the [release documentation](http://www.open3d.org/docs/release/) and the [latest documentation](http://www.open3d.org/docs/latest/) (for `master` branch).
### Steps to reproduce the issue
#### I first cloned Open3D by:
```
git clone https://github.com/isl-org/Open3D.git
cd Open3D
```
#### Then, I build Open3D (on Ubuntu 20.04, with CUDA 11.5) with:
```
mkdir build
cd build
mkdir build
cd build
cmake .. -G "Visual Studio 15 2017" -A x64 -DBUILD_CUDA_MODULE=ON -DBUILD_COMMON_CUDA_ARCHS=ON -DBUILD_LIBREALSENSE=ON -DBUILD_AZURE_KINECT=ON -DCMAKE_INSTALL_PREFIX="<open3d_install_directory>"
```
### Error message
```shell
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): warning C4522: “open3d::core::Tensor”: 指定了多个赋值运算符 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\PickPointsInteractor.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
and many following same erros:
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 [D:\Program Files\Open3D\build\cpp\open3d\core\core.vcxproj]
```
### Open3D, Python and System information
```markdown
- Operating system: Windows 10 64-bit
- Python version: Python 3.8`
- Open3D version: 15.9.21+g9802d43bc3
- System architecture: x86
- Is this a remote workstation?: no
- How did you install Open3D?: build from source
- Compiler version (if built from source): vs2017 NVIDIA GPU Computing Toolkit/CUDA/v11.0/include (found version "11.0.221")
```
### Additional information
_No response_
|
1.0
|
Summarize the issue and your environments (e.g., "Cannot compile on Ubuntu 20.04 with OpenBLAS") - ### Checklist
- [X] I have searched for [similar issues](https://github.com/isl-org/Open3D/issues).
- [X] For Python issues, I have tested with the [latest development wheel](http://www.open3d.org/docs/latest/getting_started.html#development-version-pip).
- [X] I have checked the [release documentation](http://www.open3d.org/docs/release/) and the [latest documentation](http://www.open3d.org/docs/latest/) (for `master` branch).
### Steps to reproduce the issue
#### I first cloned Open3D by:
```
git clone https://github.com/isl-org/Open3D.git
cd Open3D
```
#### Then, I build Open3D (on Ubuntu 20.04, with CUDA 11.5) with:
```
mkdir build
cd build
mkdir build
cd build
cmake .. -G "Visual Studio 15 2017" -A x64 -DBUILD_CUDA_MODULE=ON -DBUILD_COMMON_CUDA_ARCHS=ON -DBUILD_LIBREALSENSE=ON -DBUILD_AZURE_KINECT=ON -DCMAKE_INSTALL_PREFIX="<open3d_install_directory>"
```
### Error message
```shell
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): warning C4522: “open3d::core::Tensor”: 指定了多个赋值运算符 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\PickPointsInteractor.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj]
and many following same erros:
D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 [D:\Program Files\Open3D\build\cpp\open3d\core\core.vcxproj]
```
### Open3D, Python and System information
```markdown
- Operating system: Windows 10 64-bit
- Python version: Python 3.8`
- Open3D version: 15.9.21+g9802d43bc3
- System architecture: x86
- Is this a remote workstation?: no
- How did you install Open3D?: build from source
- Compiler version (if built from source): vs2017 NVIDIA GPU Computing Toolkit/CUDA/v11.0/include (found version "11.0.221")
```
### Additional information
_No response_
|
build
|
summarize the issue and your environments e g cannot compile on ubuntu with openblas checklist i have searched for for python issues i have tested with the i have checked the and the for master branch steps to reproduce the issue i first cloned by git clone cd then i build on ubuntu with cuda with mkdir build cd build mkdir build cd build cmake g visual studio a dbuild cuda module on dbuild common cuda archs on dbuild librealsense on dbuild azure kinect on dcmake install prefix error message shell d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 编译源文件 d program files cpp visualization gui uiimage cpp d program files cpp core tensor h warning “ core tensor” 指定了多个赋值运算符 编译源文件 d program files cpp visualization gui uiimage cpp d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 编译源文件 d program files cpp visualization gui pickpointsinteractor cpp and many following same erros d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 python and system information markdown operating system windows bit python version python version system architecture is this a remote workstation no how did you install build from source compiler version if built from source nvidia gpu computing toolkit cuda include found version additional information no response
| 1
|
423,741
| 28,931,685,301
|
IssuesEvent
|
2023-05-09 00:16:03
|
OpenINF/wg-a-team
|
https://api.github.com/repos/OpenINF/wg-a-team
|
opened
|
meta: the new Product Integrity WG meta-team
|
documentation help wanted question
|
Our readme states something about a _new Product Integrity WG meta-team_ @jorgebucaran & @smorimoto would probably know more about this as QA & PM, so just opening an issue about it for us to see what needs to be done here.
|
1.0
|
meta: the new Product Integrity WG meta-team - Our readme states something about a _new Product Integrity WG meta-team_ @jorgebucaran & @smorimoto would probably know more about this as QA & PM, so just opening an issue about it for us to see what needs to be done here.
|
non_build
|
meta the new product integrity wg meta team our readme states something about a new product integrity wg meta team jorgebucaran amp smorimoto would probably know more about this as qa amp pm so just opening an issue about it for us to see what needs to be done here
| 0
|
77,972
| 22,060,897,136
|
IssuesEvent
|
2022-05-30 17:39:58
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
Add back `help_on_error` for download-ci-llvm
|
E-easy E-mentor A-rustbuild
|
I accidentally missed this when porting it to rustbuild: https://github.com/rust-lang/rust/pull/95170/files#diff-9e25a089f077eae8c8cefe9586cc07498e64cd950be4c954cb90e4245e5f9fc3L634-L651
It would be nice to add it back.
Mentoring instructions: Add a new `help_on_error` parameter to `Builder::download_component_helper` and pass it down from `download_ci_llvm`:
https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/builder.rs#L872 https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/native.rs#L182
This code moved around in https://github.com/rust-lang/rust/pull/96687 so you may want to base your changes off that PR until it lands.
@rustbot label +E-easy +E-mentor +A-rustbuild
<!-- TRIAGEBOT_START -->
<!-- TRIAGEBOT_ASSIGN_START -->
<!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"binggh"}$$TRIAGEBOT_ASSIGN_DATA_END -->
<!-- TRIAGEBOT_ASSIGN_END -->
<!-- TRIAGEBOT_END -->
|
1.0
|
Add back `help_on_error` for download-ci-llvm - I accidentally missed this when porting it to rustbuild: https://github.com/rust-lang/rust/pull/95170/files#diff-9e25a089f077eae8c8cefe9586cc07498e64cd950be4c954cb90e4245e5f9fc3L634-L651
It would be nice to add it back.
Mentoring instructions: Add a new `help_on_error` parameter to `Builder::download_component_helper` and pass it down from `download_ci_llvm`:
https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/builder.rs#L872 https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/native.rs#L182
This code moved around in https://github.com/rust-lang/rust/pull/96687 so you may want to base your changes off that PR until it lands.
@rustbot label +E-easy +E-mentor +A-rustbuild
<!-- TRIAGEBOT_START -->
<!-- TRIAGEBOT_ASSIGN_START -->
<!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"binggh"}$$TRIAGEBOT_ASSIGN_DATA_END -->
<!-- TRIAGEBOT_ASSIGN_END -->
<!-- TRIAGEBOT_END -->
|
build
|
add back help on error for download ci llvm i accidentally missed this when porting it to rustbuild it would be nice to add it back mentoring instructions add a new help on error parameter to builder download component helper and pass it down from download ci llvm this code moved around in so you may want to base your changes off that pr until it lands rustbot label e easy e mentor a rustbuild
| 1
|
176,976
| 28,308,294,995
|
IssuesEvent
|
2023-04-10 13:12:29
|
bounswe/bounswe2023group5
|
https://api.github.com/repos/bounswe/bounswe2023group5
|
closed
|
Designing Sequence Diagram: Moderator
|
Priority: Critical Type: Design Status: Done
|
### Description
The sequence diagrams are shared among the team members, I will design the use case diagram of moderator. Moderator checks the reports and evaluates them. Moderators also can delete posts, comments and ban users from the forum.
### 👮♀️ Reviewer
Bilal Atım
### ⏰ Deadline
10.04.2023
|
1.0
|
Designing Sequence Diagram: Moderator - ### Description
The sequence diagrams are shared among the team members, I will design the use case diagram of moderator. Moderator checks the reports and evaluates them. Moderators also can delete posts, comments and ban users from the forum.
### 👮♀️ Reviewer
Bilal Atım
### ⏰ Deadline
10.04.2023
|
non_build
|
designing sequence diagram moderator description the sequence diagrams are shared among the team members i will design the use case diagram of moderator moderator checks the reports and evaluates them moderators also can delete posts comments and ban users from the forum 👮♀️ reviewer bilal atım ⏰ deadline
| 0
|
181,195
| 14,008,236,570
|
IssuesEvent
|
2020-10-28 23:11:56
|
alteryx/evalml
|
https://api.github.com/repos/alteryx/evalml
|
closed
|
CircleCI: auth needed for Docker Hub as of Nov 1st
|
task testing
|
Message which @jeremyliweishih posted in slack today (thanks Jeremy!):
> Hi there,
> On November 1st, Docker Hub will begin limiting anonymous image pulls. We want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow.
> Adding Docker authentication to your pipeline config is the easiest way to avoid any service disruptions. If you use the Docker executor or pull Docker images when using the machine executor on CircleCI, we encourage you to authenticate. Because the anonymous API rate limits are based on IP addresses, they will impact CircleCI cloud customers. Authenticated users get higher per-user rate limits, regardless of IP.
> We are currently working on a partnership with Docker to minimize the impact of this change for our users and will share more details as we get them.
> For more information or to leave a question for us, please head over to Discuss or contact support.
> Thanks and happy building,
> - The CircleCI team
Next step: determine what if any action is needed in order for our CI jobs to continue uninterrupted.
@rwedge @gsheni FYI this may affect CI for featuretools/compose/woodwork/others as well
|
1.0
|
CircleCI: auth needed for Docker Hub as of Nov 1st - Message which @jeremyliweishih posted in slack today (thanks Jeremy!):
> Hi there,
> On November 1st, Docker Hub will begin limiting anonymous image pulls. We want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow.
> Adding Docker authentication to your pipeline config is the easiest way to avoid any service disruptions. If you use the Docker executor or pull Docker images when using the machine executor on CircleCI, we encourage you to authenticate. Because the anonymous API rate limits are based on IP addresses, they will impact CircleCI cloud customers. Authenticated users get higher per-user rate limits, regardless of IP.
> We are currently working on a partnership with Docker to minimize the impact of this change for our users and will share more details as we get them.
> For more information or to leave a question for us, please head over to Discuss or contact support.
> Thanks and happy building,
> - The CircleCI team
Next step: determine what if any action is needed in order for our CI jobs to continue uninterrupted.
@rwedge @gsheni FYI this may affect CI for featuretools/compose/woodwork/others as well
|
non_build
|
circleci auth needed for docker hub as of nov message which jeremyliweishih posted in slack today thanks jeremy hi there on november docker hub will begin limiting anonymous image pulls we want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow adding docker authentication to your pipeline config is the easiest way to avoid any service disruptions if you use the docker executor or pull docker images when using the machine executor on circleci we encourage you to authenticate because the anonymous api rate limits are based on ip addresses they will impact circleci cloud customers authenticated users get higher per user rate limits regardless of ip we are currently working on a partnership with docker to minimize the impact of this change for our users and will share more details as we get them for more information or to leave a question for us please head over to discuss or contact support thanks and happy building the circleci team next step determine what if any action is needed in order for our ci jobs to continue uninterrupted rwedge gsheni fyi this may affect ci for featuretools compose woodwork others as well
| 0
|
45,461
| 11,672,816,364
|
IssuesEvent
|
2020-03-04 07:38:38
|
Polymer/tools
|
https://api.github.com/repos/Polymer/tools
|
closed
|
Add glob support for fragments
|
Package: build wontfix
|
It would be nice to pass fragments into `resolveGlob` as well as `sourceGlobs` are processed:
```
"fragments": "/src/pages/**/*"
```
|
1.0
|
Add glob support for fragments - It would be nice to pass fragments into `resolveGlob` as well as `sourceGlobs` are processed:
```
"fragments": "/src/pages/**/*"
```
|
build
|
add glob support for fragments it would be nice to pass fragments into resolveglob as well as sourceglobs are processed fragments src pages
| 1
|
54,520
| 13,383,522,510
|
IssuesEvent
|
2020-09-02 10:27:56
|
lkeegan/spatial-model-editor
|
https://api.github.com/repos/lkeegan/spatial-model-editor
|
closed
|
Pixel OpenMP support
|
CLI build system enhancement
|
Currently the Pixel simulator supports multithreading using tbb, which is statically linked in the distributed binaries. It has the advantage of being very flexible & has a nice c++ interface. But for parallelisation of simple loops, which is all that we are currently using it for, its implementation seems to have significant overhead costs compared to e.g. openMP.
It would be nice to add OpenMP as a compile-time alternative option for users who compile the code themselves, e.g. for running the CLI version on a cluster.
|
1.0
|
Pixel OpenMP support - Currently the Pixel simulator supports multithreading using tbb, which is statically linked in the distributed binaries. It has the advantage of being very flexible & has a nice c++ interface. But for parallelisation of simple loops, which is all that we are currently using it for, its implementation seems to have significant overhead costs compared to e.g. openMP.
It would be nice to add OpenMP as a compile-time alternative option for users who compile the code themselves, e.g. for running the CLI version on a cluster.
|
build
|
pixel openmp support currently the pixel simulator supports multithreading using tbb which is statically linked in the distributed binaries it has the advantage of being very flexible has a nice c interface but for parallelisation of simple loops which is all that we are currently using it for its implementation seems to have significant overhead costs compared to e g openmp it would be nice to add openmp as a compile time alternative option for users who compile the code themselves e g for running the cli version on a cluster
| 1
|
11,277
| 9,301,759,160
|
IssuesEvent
|
2019-03-24 01:35:49
|
dimigoin/dimigoin-front
|
https://api.github.com/repos/dimigoin/dimigoin-front
|
closed
|
Feature request in teacher afterschool management page
|
feature request service
|
- 엑셀 파일 다운로드 버튼이 없습니다 (학년별로 달라야합니다.)
- ~~세부 관리 페이지가 존재하지 않습니다. 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다~~
#54 를 참고해서 구현해야 합니다.
|
1.0
|
Feature request in teacher afterschool management page - - 엑셀 파일 다운로드 버튼이 없습니다 (학년별로 달라야합니다.)
- ~~세부 관리 페이지가 존재하지 않습니다. 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다~~
#54 를 참고해서 구현해야 합니다.
|
non_build
|
feature request in teacher afterschool management page 엑셀 파일 다운로드 버튼이 없습니다 학년별로 달라야합니다 세부 관리 페이지가 존재하지 않습니다 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다 를 참고해서 구현해야 합니다
| 0
|
169,055
| 26,740,743,190
|
IssuesEvent
|
2023-01-30 12:48:35
|
codestates-seb/seb41_main_028
|
https://api.github.com/repos/codestates-seb/seb41_main_028
|
closed
|
[FE] Feat: certificationModal 데이터 바인딩 작업
|
FE Feat Design
|
## To do List
<!--해당 Issue를 해결하기 위해 수행해야 할 To do list를 작성해주세요.-->
- [x] certification 모달창의 데이터를 받아와서 뿌려주기
## Deadline
<!--작업 완료 기한을 입력해주세요. (e.g. 221219 / 15:00)-->
230129 / 23:59
## Issue Checklist
아래 체크리스트를 확인해주세요.
- [x] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 )
- [x] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
|
1.0
|
[FE] Feat: certificationModal 데이터 바인딩 작업 - ## To do List
<!--해당 Issue를 해결하기 위해 수행해야 할 To do list를 작성해주세요.-->
- [x] certification 모달창의 데이터를 받아와서 뿌려주기
## Deadline
<!--작업 완료 기한을 입력해주세요. (e.g. 221219 / 15:00)-->
230129 / 23:59
## Issue Checklist
아래 체크리스트를 확인해주세요.
- [x] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 )
- [x] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
|
non_build
|
feat certificationmodal 데이터 바인딩 작업 to do list certification 모달창의 데이터를 받아와서 뿌려주기 deadline issue checklist 아래 체크리스트를 확인해주세요 issue title을 다음 예시와 같이 작성했습니다 e g feat 회원가입 기능 구현 우측의 assignees labels projects milestone을 적절하게 선택했습니다
| 0
|
244,526
| 7,876,155,042
|
IssuesEvent
|
2018-06-25 23:21:58
|
bleehu/Compound_X
|
https://api.github.com/repos/bleehu/Compound_X
|
closed
|
Weapon Proposal: Staff-sling
|
Feature Request Greenlit - Awaiting Pull Request Priority - Low
|
It's like a sling, but gigantic. Probably requires very high Dex to use, can load with bullets (which means lead lumps) or grenades.
More normal slings also includeable.
|
1.0
|
Weapon Proposal: Staff-sling - It's like a sling, but gigantic. Probably requires very high Dex to use, can load with bullets (which means lead lumps) or grenades.
More normal slings also includeable.
|
non_build
|
weapon proposal staff sling it s like a sling but gigantic probably requires very high dex to use can load with bullets which means lead lumps or grenades more normal slings also includeable
| 0
|
5,956
| 13,391,049,326
|
IssuesEvent
|
2020-09-02 21:42:09
|
pirate/ArchiveBox
|
https://api.github.com/repos/pirate/ArchiveBox
|
closed
|
Bugfix: django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
|
bug changes: architecture complexity: medium status: wip
|
#### Describe the bug
Y'all helped me with upgrading my super old archive to the django branch before official 0.4.9 release. I recently upgraded to the newest version, so I could start adding links. archivebox said I had to re-init. `archivebox init` gives me following error, and will not let me add new links.
```
django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
```
Full log/error below.
#### Steps to reproduce
1. `git checkout master` to switch from django branch.
2. `git pull origin master` to pull new release.
3. `pip install -e .` (also tried with `pip uninstall archivebox && pip install .`)
4. Navigate to archivebox-output directory.
5. Run `archivebox init`.
6. error.
#### Screenshots or log output
```
[i] [2020-07-31 17:34:44] ArchiveBox v0.4.9: archivebox init
> /.archivebox-output/archive-working
[*] Updating existing ArchiveBox collection in this folder...
/.archivebox-output/archive-working
------------------------------------------------------------------
[*] Verifying archive folder structure...
√ /.archivebox-output/archive-working/sources
√ /.archivebox-output/archive-working/archive
√ /.archivebox-output/archive-working/logs
√ /.archivebox-output/archive-working/ArchiveBox.conf
[*] Verifying main SQL index and running migrations...
√ /.archivebox-output/archive-working/index.sqlite3
Operations to perform:
Apply all migrations: admin, auth, contenttypes, core, sessions
Running migrations:
Applying core.0005_auto_20200728_0326... OK
[*] Collecting links from any existing indexes and archive folders...
√ Loaded 1376 links from existing main index.
√ Added 347 orphaned links from existing archive directories.
! Skipped adding 239 invalid link data directories.
X /* SNIP A BUNCH OF BROKEN ARCHIVES /*
Hint: For more information about the link data directories that were skipped, run:
archivebox status
archivebox list --status=invalid
[*] [2020-07-31 18:01:50] Writing 1723 links to main index...
Traceback (most recent call last):
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 575, in update_or_create
obj = self.select_for_update().get(**kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 417, in get
self.model._meta.object_name
core.models.DoesNotExist: Snapshot matching query does not exist.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute
return Database.Cursor.execute(self, query, params)
sqlite3.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/USERNAME/.local/bin/archivebox", line 33, in <module>
sys.exit(load_entry_point('archivebox', 'console_scripts', 'archivebox')())
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 126, in main
pwd=pwd or OUTPUT_DIR,
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 62, in run_subcommand
module.main(args=subcommand_args, stdin=stdin, pwd=pwd) # type: ignore
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/archivebox_init.py", line 35, in main
out_dir=pwd or OUTPUT_DIR,
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/main.py", line 369, in init
write_main_index(list(all_links.values()), out_dir=out_dir)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/__init__.py", line 235, in write_main_index
write_sql_main_index(links, out_dir=out_dir)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/sql.py", line 42, in write_sql_main_index
Snapshot.objects.update_or_create(url=link.url, defaults=info)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 580, in update_or_create
obj, created = self._create_object_from_params(kwargs, params, lock=True)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 604, in _create_object_from_params
raise e
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 596, in _create_object_from_params
obj = self.create(**params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 433, in create
obj.save(force_insert=True, using=self.db)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 746, in save
force_update=force_update, update_fields=update_fields)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 784, in save_base
force_update, using, update_fields,
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 887, in _save_table
results = self._do_insert(cls._base_manager, using, fields, returning_fields, raw)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 926, in _do_insert
using=using, raw=raw,
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 1204, in _insert
return query.get_compiler(using=using).execute_sql(returning_fields)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/sql/compiler.py", line 1392, in execute_sql
cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 68, in execute
return self._execute_with_wrappers(sql, params, many=False, executor=self._execute)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 77, in _execute_with_wrappers
return executor(sql, params, many, context)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/utils.py", line 90, in __exit__
raise dj_exc_value.with_traceback(traceback) from exc_value
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute
return Database.Cursor.execute(self, query, params)
django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
```
#### Software versions
- OS: Ubuntu 18.04
- ArchiveBox version: 0.4.9 (0ac4e12)
- Python version: Python 3.7.8
|
1.0
|
Bugfix: django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp -
#### Describe the bug
Y'all helped me with upgrading my super old archive to the django branch before official 0.4.9 release. I recently upgraded to the newest version, so I could start adding links. archivebox said I had to re-init. `archivebox init` gives me following error, and will not let me add new links.
```
django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
```
Full log/error below.
#### Steps to reproduce
1. `git checkout master` to switch from django branch.
2. `git pull origin master` to pull new release.
3. `pip install -e .` (also tried with `pip uninstall archivebox && pip install .`)
4. Navigate to archivebox-output directory.
5. Run `archivebox init`.
6. error.
#### Screenshots or log output
```
[i] [2020-07-31 17:34:44] ArchiveBox v0.4.9: archivebox init
> /.archivebox-output/archive-working
[*] Updating existing ArchiveBox collection in this folder...
/.archivebox-output/archive-working
------------------------------------------------------------------
[*] Verifying archive folder structure...
√ /.archivebox-output/archive-working/sources
√ /.archivebox-output/archive-working/archive
√ /.archivebox-output/archive-working/logs
√ /.archivebox-output/archive-working/ArchiveBox.conf
[*] Verifying main SQL index and running migrations...
√ /.archivebox-output/archive-working/index.sqlite3
Operations to perform:
Apply all migrations: admin, auth, contenttypes, core, sessions
Running migrations:
Applying core.0005_auto_20200728_0326... OK
[*] Collecting links from any existing indexes and archive folders...
√ Loaded 1376 links from existing main index.
√ Added 347 orphaned links from existing archive directories.
! Skipped adding 239 invalid link data directories.
X /* SNIP A BUNCH OF BROKEN ARCHIVES /*
Hint: For more information about the link data directories that were skipped, run:
archivebox status
archivebox list --status=invalid
[*] [2020-07-31 18:01:50] Writing 1723 links to main index...
Traceback (most recent call last):
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 575, in update_or_create
obj = self.select_for_update().get(**kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 417, in get
self.model._meta.object_name
core.models.DoesNotExist: Snapshot matching query does not exist.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute
return Database.Cursor.execute(self, query, params)
sqlite3.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/USERNAME/.local/bin/archivebox", line 33, in <module>
sys.exit(load_entry_point('archivebox', 'console_scripts', 'archivebox')())
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 126, in main
pwd=pwd or OUTPUT_DIR,
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 62, in run_subcommand
module.main(args=subcommand_args, stdin=stdin, pwd=pwd) # type: ignore
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/archivebox_init.py", line 35, in main
out_dir=pwd or OUTPUT_DIR,
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/main.py", line 369, in init
write_main_index(list(all_links.values()), out_dir=out_dir)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/__init__.py", line 235, in write_main_index
write_sql_main_index(links, out_dir=out_dir)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function
return func(*args, **kwargs)
File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/sql.py", line 42, in write_sql_main_index
Snapshot.objects.update_or_create(url=link.url, defaults=info)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 580, in update_or_create
obj, created = self._create_object_from_params(kwargs, params, lock=True)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 604, in _create_object_from_params
raise e
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 596, in _create_object_from_params
obj = self.create(**params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 433, in create
obj.save(force_insert=True, using=self.db)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 746, in save
force_update=force_update, update_fields=update_fields)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 784, in save_base
force_update, using, update_fields,
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 887, in _save_table
results = self._do_insert(cls._base_manager, using, fields, returning_fields, raw)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 926, in _do_insert
using=using, raw=raw,
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 1204, in _insert
return query.get_compiler(using=using).execute_sql(returning_fields)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/sql/compiler.py", line 1392, in execute_sql
cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 68, in execute
return self._execute_with_wrappers(sql, params, many=False, executor=self._execute)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 77, in _execute_with_wrappers
return executor(sql, params, many, context)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/utils.py", line 90, in __exit__
raise dj_exc_value.with_traceback(traceback) from exc_value
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute
return self.cursor.execute(sql, params)
File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute
return Database.Cursor.execute(self, query, params)
django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
```
#### Software versions
- OS: Ubuntu 18.04
- ArchiveBox version: 0.4.9 (0ac4e12)
- Python version: Python 3.7.8
|
non_build
|
bugfix django db utils integrityerror unique constraint failed core snapshot timestamp describe the bug y all helped me with upgrading my super old archive to the django branch before official release i recently upgraded to the newest version so i could start adding links archivebox said i had to re init archivebox init gives me following error and will not let me add new links django db utils integrityerror unique constraint failed core snapshot timestamp full log error below steps to reproduce git checkout master to switch from django branch git pull origin master to pull new release pip install e also tried with pip uninstall archivebox pip install navigate to archivebox output directory run archivebox init error screenshots or log output archivebox archivebox init archivebox output archive working updating existing archivebox collection in this folder archivebox output archive working verifying archive folder structure √ archivebox output archive working sources √ archivebox output archive working archive √ archivebox output archive working logs √ archivebox output archive working archivebox conf verifying main sql index and running migrations √ archivebox output archive working index operations to perform apply all migrations admin auth contenttypes core sessions running migrations applying core auto ok collecting links from any existing indexes and archive folders √ loaded links from existing main index √ added orphaned links from existing archive directories skipped adding invalid link data directories x snip a bunch of broken archives hint for more information about the link data directories that were skipped run archivebox status archivebox list status invalid writing links to main index traceback most recent call last file home username local lib site packages django db models query py line in update or create obj self select for update get kwargs file home username local lib site packages django db models query py line in get self model meta object name core models doesnotexist snapshot matching query does not exist during handling of the above exception another exception occurred traceback most recent call last file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db backends base py line in execute return database cursor execute self query params integrityerror unique constraint failed core snapshot timestamp the above exception was the direct cause of the following exception traceback most recent call last file home username local bin archivebox line in sys exit load entry point archivebox console scripts archivebox file home username datahoard archivebox archivebox cli init py line in main pwd pwd or output dir file home username datahoard archivebox archivebox cli init py line in run subcommand module main args subcommand args stdin stdin pwd pwd type ignore file home username datahoard archivebox archivebox cli archivebox init py line in main out dir pwd or output dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox main py line in init write main index list all links values out dir out dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox index init py line in write main index write sql main index links out dir out dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox index sql py line in write sql main index snapshot objects update or create url link url defaults info file home username local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file home username local lib site packages django db models query py line in update or create obj created self create object from params kwargs params lock true file home username local lib site packages django db models query py line in create object from params raise e file home username local lib site packages django db models query py line in create object from params obj self create params file home username local lib site packages django db models query py line in create obj save force insert true using self db file home username local lib site packages django db models base py line in save force update force update update fields update fields file home username local lib site packages django db models base py line in save base force update using update fields file home username local lib site packages django db models base py line in save table results self do insert cls base manager using fields returning fields raw file home username local lib site packages django db models base py line in do insert using using raw raw file home username local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file home username local lib site packages django db models query py line in insert return query get compiler using using execute sql returning fields file home username local lib site packages django db models sql compiler py line in execute sql cursor execute sql params file home username local lib site packages django db backends utils py line in execute return self execute with wrappers sql params many false executor self execute file home username local lib site packages django db backends utils py line in execute with wrappers return executor sql params many context file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db utils py line in exit raise dj exc value with traceback traceback from exc value file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db backends base py line in execute return database cursor execute self query params django db utils integrityerror unique constraint failed core snapshot timestamp software versions os ubuntu archivebox version python version python
| 0
|
80,798
| 15,589,001,517
|
IssuesEvent
|
2021-03-18 07:23:03
|
soumya132/pomscan
|
https://api.github.com/repos/soumya132/pomscan
|
closed
|
CVE-2017-5664 (High) detected in tomcat-embed-core-8.5.4.jar
|
security vulnerability
|
## CVE-2017-5664 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: pomscan/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.4.0.RELEASE.jar
- :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.
<p>Publish Date: 2017-06-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p>
<p>Release Date: 2017-06-06</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-5664 (High) detected in tomcat-embed-core-8.5.4.jar - ## CVE-2017-5664 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: pomscan/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.4.0.RELEASE.jar
- :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.
<p>Publish Date: 2017-06-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p>
<p>Release Date: 2017-06-06</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file pomscan pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter jersey release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details the error page mechanism of the java servlet specification requires that when an error occurs and an error page is configured for the error that occurred the original request and response are forwarded to the error page this means that the request is presented to the error page with the original http method if the error page is a static file expected behaviour is to serve content of the file as if processing a get request regardless of the actual http method the default servlet in apache tomcat to to to and to did not do this depending on the original request this could lead to unexpected and undesirable results for static error pages including if the defaultservlet is configured to permit writes the replacement or removal of the custom error page notes for other user provided error pages unless explicitly coded otherwise jsps ignore the http method jsps used as error pages must must ensure that they handle any error dispatch as a get request regardless of the actual method by default the response generated by a servlet does depend on the http method custom servlets used as error pages must ensure that they handle any error dispatch as a get request regardless of the actual method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
| 0
|
38,537
| 10,216,222,842
|
IssuesEvent
|
2019-08-15 09:55:58
|
jenkins-x/jx
|
https://api.github.com/repos/jenkins-x/jx
|
closed
|
GitHub commit status should offer a link to build logs
|
area/build kind/enhancement lifecycle/rotten priority/backlog
|
### Summary
For a PR built by Jenkins X Serverless
Checks show
"
serverless-jenkins — succeeded
"
As `PipelineActivitySpec` do define attribute `BuildLogsURL` I would expect this check gives me a link so I can read build log from gitHub web UI - expecialy in case of build failures
### Steps to reproduce the behavior
### Expected behavior
Check to provide a link to read build log
### Actual behavior
Need to use CLI
### Jx version
```
➜ (master ✗) demo jx version
NAME VERSION
jx 1.3.972
jenkins x platform 0.0.3402
Kubernetes cluster v1.11.7-gke.4
kubectl v1.13.3
helm client v2.12.3+geecf22f
helm server v2.12.2+g7d2b0c7
git git version 2.20.1
Operating System Mac OS X 10.14.1 build 18B75
```
### Jenkins type
<!--
Select which Jenkins installation type are you using.
-->
- [ ] Classic Jenkins
- [X] Serverless Jenkins
|
1.0
|
GitHub commit status should offer a link to build logs - ### Summary
For a PR built by Jenkins X Serverless
Checks show
"
serverless-jenkins — succeeded
"
As `PipelineActivitySpec` do define attribute `BuildLogsURL` I would expect this check gives me a link so I can read build log from gitHub web UI - expecialy in case of build failures
### Steps to reproduce the behavior
### Expected behavior
Check to provide a link to read build log
### Actual behavior
Need to use CLI
### Jx version
```
➜ (master ✗) demo jx version
NAME VERSION
jx 1.3.972
jenkins x platform 0.0.3402
Kubernetes cluster v1.11.7-gke.4
kubectl v1.13.3
helm client v2.12.3+geecf22f
helm server v2.12.2+g7d2b0c7
git git version 2.20.1
Operating System Mac OS X 10.14.1 build 18B75
```
### Jenkins type
<!--
Select which Jenkins installation type are you using.
-->
- [ ] Classic Jenkins
- [X] Serverless Jenkins
|
build
|
github commit status should offer a link to build logs summary for a pr built by jenkins x serverless checks show serverless jenkins — succeeded as pipelineactivityspec do define attribute buildlogsurl i would expect this check gives me a link so i can read build log from github web ui expecialy in case of build failures steps to reproduce the behavior expected behavior check to provide a link to read build log actual behavior need to use cli jx version ➜ master ✗ demo jx version name version jx jenkins x platform kubernetes cluster gke kubectl helm client helm server git git version operating system mac os x build jenkins type select which jenkins installation type are you using classic jenkins serverless jenkins
| 1
|
43,950
| 17,774,494,094
|
IssuesEvent
|
2021-08-30 17:23:11
|
Azure/azure-sdk-for-net
|
https://api.github.com/repos/Azure/azure-sdk-for-net
|
closed
|
[Flaky test] Should_retry_when_throttled_and_no_ambient_transaction_is_detected
|
Service Bus Client
|
Error message
```
Microsoft.Azure.ServiceBus.ServerBusyException : Rico KABOOM!
```
Stack trace
```
at Microsoft.Azure.ServiceBus.UnitTests.RetryPolicyTests.<>c__DisplayClass0_0.<Should_retry_when_throttled_and_no_ambient_transaction_is_detected>b__0() in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\tests\RetryPolicyTests.cs:line 28
at Microsoft.Azure.ServiceBus.RetryPolicy.RunOperation(Func`1 operation, TimeSpan operationTimeout) in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\src\RetryPolicy.cs:line 82
at Microsoft.Azure.ServiceBus.RetryPolicy.RunOperation(Func`1 operation, TimeSpan operationTimeout) in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\src\RetryPolicy.cs:line 107
at Microsoft.Azure.ServiceBus.UnitTests.RetryPolicyTests.Should_retry_when_throttled_and_no_ambient_transaction_is_detected() in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\tests\RetryPolicyTests.cs:line 18
--- End of stack trace from previous location where exception was thrown ---
```
|
1.0
|
[Flaky test] Should_retry_when_throttled_and_no_ambient_transaction_is_detected -
Error message
```
Microsoft.Azure.ServiceBus.ServerBusyException : Rico KABOOM!
```
Stack trace
```
at Microsoft.Azure.ServiceBus.UnitTests.RetryPolicyTests.<>c__DisplayClass0_0.<Should_retry_when_throttled_and_no_ambient_transaction_is_detected>b__0() in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\tests\RetryPolicyTests.cs:line 28
at Microsoft.Azure.ServiceBus.RetryPolicy.RunOperation(Func`1 operation, TimeSpan operationTimeout) in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\src\RetryPolicy.cs:line 82
at Microsoft.Azure.ServiceBus.RetryPolicy.RunOperation(Func`1 operation, TimeSpan operationTimeout) in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\src\RetryPolicy.cs:line 107
at Microsoft.Azure.ServiceBus.UnitTests.RetryPolicyTests.Should_retry_when_throttled_and_no_ambient_transaction_is_detected() in D:\a\1\s\sdk\servicebus\Microsoft.Azure.ServiceBus\tests\RetryPolicyTests.cs:line 18
--- End of stack trace from previous location where exception was thrown ---
```
|
non_build
|
should retry when throttled and no ambient transaction is detected error message microsoft azure servicebus serverbusyexception rico kaboom stack trace at microsoft azure servicebus unittests retrypolicytests c b in d a s sdk servicebus microsoft azure servicebus tests retrypolicytests cs line at microsoft azure servicebus retrypolicy runoperation func operation timespan operationtimeout in d a s sdk servicebus microsoft azure servicebus src retrypolicy cs line at microsoft azure servicebus retrypolicy runoperation func operation timespan operationtimeout in d a s sdk servicebus microsoft azure servicebus src retrypolicy cs line at microsoft azure servicebus unittests retrypolicytests should retry when throttled and no ambient transaction is detected in d a s sdk servicebus microsoft azure servicebus tests retrypolicytests cs line end of stack trace from previous location where exception was thrown
| 0
|
18,691
| 6,628,929,116
|
IssuesEvent
|
2017-09-24 01:15:44
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Ruby failure: "ClientStub #request_response via a call operation behaves like request response should send a request when configured using an override channel" on MacOS
|
infra/BUILDPONY lang/ruby platform/macOS
|
https://grpc-testing.appspot.com/job/gRPC_master_macos/2378/testReport/junit/(root)/ruby_macos_dbg_native/tools_run_tests_helper_scripts_run_ruby_sh/
```
1) ClientStub #request_response via a call operation behaves like request response should send a request when configured using an override channel
Got 0 failures and 2 other errors:
Shared Example Group: "request response" called from ./src/ruby/spec/generic/client_stub_spec.rb:232
1.1) Failure/Error: result = op.execute
GRPC::DeadlineExceeded:
4:Deadline Exceeded
# ./src/ruby/lib/grpc/generic/active_call.rb:46:in `check_status'
# ./src/ruby/lib/grpc/generic/active_call.rb:178:in `attach_status_results_and_complete_call'
# ./src/ruby/lib/grpc/generic/active_call.rb:340:in `request_response'
# ./src/ruby/lib/grpc/generic/client_stub.rb:174:in `block in request_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:227:in `get_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:158:in `block (4 levels) in <top (required)>'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:254:in `instance_exec'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:254:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:500:in `block in with_around_and_singleton_context_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:457:in `block in with_around_example_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:464:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:602:in `run_around_example_hooks_for'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:464:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:457:in `with_around_example_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:500:in `with_around_and_singleton_context_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:251:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:627:in `block in run_examples'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:623:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:623:in `run_examples'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:589:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `block (3 levels) in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `block (2 levels) in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/configuration.rb:1894:in `with_suite_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:113:in `block in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/reporter.rb:79:in `report'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:112:in `run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:87:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:71:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:45:in `invoke'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/exe/rspec:4:in `<main>'
1.2) Failure/Error: recvd_rpc = @server.request_call
GRPC::Core::CallError:
request_call completion failed
# ./src/ruby/spec/generic/client_stub_spec.rb:549:in `request_call'
# ./src/ruby/spec/generic/client_stub_spec.rb:549:in `expect_server_to_be_invoked'
# ./src/ruby/spec/generic/client_stub_spec.rb:522:in `block in run_request_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:37:in `block in wakey_thread'
# /usr/local/rvm/gems/ruby-2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context'
```
|
1.0
|
Ruby failure: "ClientStub #request_response via a call operation behaves like request response should send a request when configured using an override channel" on MacOS - https://grpc-testing.appspot.com/job/gRPC_master_macos/2378/testReport/junit/(root)/ruby_macos_dbg_native/tools_run_tests_helper_scripts_run_ruby_sh/
```
1) ClientStub #request_response via a call operation behaves like request response should send a request when configured using an override channel
Got 0 failures and 2 other errors:
Shared Example Group: "request response" called from ./src/ruby/spec/generic/client_stub_spec.rb:232
1.1) Failure/Error: result = op.execute
GRPC::DeadlineExceeded:
4:Deadline Exceeded
# ./src/ruby/lib/grpc/generic/active_call.rb:46:in `check_status'
# ./src/ruby/lib/grpc/generic/active_call.rb:178:in `attach_status_results_and_complete_call'
# ./src/ruby/lib/grpc/generic/active_call.rb:340:in `request_response'
# ./src/ruby/lib/grpc/generic/client_stub.rb:174:in `block in request_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:227:in `get_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:158:in `block (4 levels) in <top (required)>'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:254:in `instance_exec'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:254:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:500:in `block in with_around_and_singleton_context_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:457:in `block in with_around_example_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:464:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:602:in `run_around_example_hooks_for'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/hooks.rb:464:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:457:in `with_around_example_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:500:in `with_around_and_singleton_context_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example.rb:251:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:627:in `block in run_examples'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:623:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:623:in `run_examples'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:589:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `block in run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/example_group.rb:590:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `block (3 levels) in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `map'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:118:in `block (2 levels) in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/configuration.rb:1894:in `with_suite_hooks'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:113:in `block in run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/reporter.rb:79:in `report'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:112:in `run_specs'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:87:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:71:in `run'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/lib/rspec/core/runner.rb:45:in `invoke'
# /usr/local/rvm/gems/ruby-2.4.0/gems/rspec-core-3.6.0/exe/rspec:4:in `<main>'
1.2) Failure/Error: recvd_rpc = @server.request_call
GRPC::Core::CallError:
request_call completion failed
# ./src/ruby/spec/generic/client_stub_spec.rb:549:in `request_call'
# ./src/ruby/spec/generic/client_stub_spec.rb:549:in `expect_server_to_be_invoked'
# ./src/ruby/spec/generic/client_stub_spec.rb:522:in `block in run_request_response'
# ./src/ruby/spec/generic/client_stub_spec.rb:37:in `block in wakey_thread'
# /usr/local/rvm/gems/ruby-2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context'
```
|
build
|
ruby failure clientstub request response via a call operation behaves like request response should send a request when configured using an override channel on macos clientstub request response via a call operation behaves like request response should send a request when configured using an override channel got failures and other errors shared example group request response called from src ruby spec generic client stub spec rb failure error result op execute grpc deadlineexceeded deadline exceeded src ruby lib grpc generic active call rb in check status src ruby lib grpc generic active call rb in attach status results and complete call src ruby lib grpc generic active call rb in request response src ruby lib grpc generic client stub rb in block in request response src ruby spec generic client stub spec rb in get response src ruby spec generic client stub spec rb in block levels in usr local rvm gems ruby gems rspec core lib rspec core example rb in instance exec usr local rvm gems ruby gems rspec core lib rspec core example rb in block in run usr local rvm gems ruby gems rspec core lib rspec core example rb in block in with around and singleton context hooks usr local rvm gems ruby gems rspec core lib rspec core example rb in block in with around example hooks usr local rvm gems ruby gems rspec core lib rspec core hooks rb in block in run usr local rvm gems ruby gems rspec core lib rspec core hooks rb in run around example hooks for usr local rvm gems ruby gems rspec core lib rspec core hooks rb in run usr local rvm gems ruby gems rspec core lib rspec core example rb in with around example hooks usr local rvm gems ruby gems rspec core lib rspec core example rb in with around and singleton context hooks usr local rvm gems ruby gems rspec core lib rspec core example rb in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in block in run examples usr local rvm gems ruby gems rspec core lib rspec core example group rb in map usr local rvm gems ruby gems rspec core lib rspec core example group rb in run examples usr local rvm gems ruby gems rspec core lib rspec core example group rb in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in block in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in map usr local rvm gems ruby gems rspec core lib rspec core example group rb in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in block in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in map usr local rvm gems ruby gems rspec core lib rspec core example group rb in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in block in run usr local rvm gems ruby gems rspec core lib rspec core example group rb in map usr local rvm gems ruby gems rspec core lib rspec core example group rb in run usr local rvm gems ruby gems rspec core lib rspec core runner rb in block levels in run specs usr local rvm gems ruby gems rspec core lib rspec core runner rb in map usr local rvm gems ruby gems rspec core lib rspec core runner rb in block levels in run specs usr local rvm gems ruby gems rspec core lib rspec core configuration rb in with suite hooks usr local rvm gems ruby gems rspec core lib rspec core runner rb in block in run specs usr local rvm gems ruby gems rspec core lib rspec core reporter rb in report usr local rvm gems ruby gems rspec core lib rspec core runner rb in run specs usr local rvm gems ruby gems rspec core lib rspec core runner rb in run usr local rvm gems ruby gems rspec core lib rspec core runner rb in run usr local rvm gems ruby gems rspec core lib rspec core runner rb in invoke usr local rvm gems ruby gems rspec core exe rspec in failure error recvd rpc server request call grpc core callerror request call completion failed src ruby spec generic client stub spec rb in request call src ruby spec generic client stub spec rb in expect server to be invoked src ruby spec generic client stub spec rb in block in run request response src ruby spec generic client stub spec rb in block in wakey thread usr local rvm gems ruby gems logging lib logging diagnostic context rb in block in create with logging context
| 1
|
1,459
| 2,760,898,803
|
IssuesEvent
|
2015-04-28 14:42:16
|
orbeon/orbeon-forms
|
https://api.github.com/repos/orbeon/orbeon-forms
|
closed
|
Removing schema with used types "crashes" Form Builder
|
Crasher Form Builder
|
To reproduce, create a form, attach this [schema](https://github.com/orbeon/orbeon-forms/blob/master/src/main/resources/org/orbeon/oxf/xforms/xforms-types.xsd), set the input field to be of type `rating`, click on Change Schema, and click on Delete Existing XML Schema. This results on the error dialog being shown, Form Builder becoming unusable, and the error `Invalid schema type 'rating'` shows in the log.
|
1.0
|
Removing schema with used types "crashes" Form Builder - To reproduce, create a form, attach this [schema](https://github.com/orbeon/orbeon-forms/blob/master/src/main/resources/org/orbeon/oxf/xforms/xforms-types.xsd), set the input field to be of type `rating`, click on Change Schema, and click on Delete Existing XML Schema. This results on the error dialog being shown, Form Builder becoming unusable, and the error `Invalid schema type 'rating'` shows in the log.
|
build
|
removing schema with used types crashes form builder to reproduce create a form attach this set the input field to be of type rating click on change schema and click on delete existing xml schema this results on the error dialog being shown form builder becoming unusable and the error invalid schema type rating shows in the log
| 1
|
214,716
| 24,101,256,981
|
IssuesEvent
|
2022-09-20 01:04:19
|
jrshutske/unit-conversion-api
|
https://api.github.com/repos/jrshutske/unit-conversion-api
|
opened
|
CVE-2022-38752 (Medium) detected in snakeyaml-1.23.jar
|
security vulnerability
|
## CVE-2022-38752 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.23.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-starter-2.1.2.RELEASE.jar
- :x: **snakeyaml-1.23.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack-overflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38752>CVE-2022-38752</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9w3m-gqgf-c4p9">https://github.com/advisories/GHSA-9w3m-gqgf-c4p9</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution: org.yaml:snakeyaml:1.32
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-38752 (Medium) detected in snakeyaml-1.23.jar - ## CVE-2022-38752 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.23.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-starter-2.1.2.RELEASE.jar
- :x: **snakeyaml-1.23.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack-overflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38752>CVE-2022-38752</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9w3m-gqgf-c4p9">https://github.com/advisories/GHSA-9w3m-gqgf-c4p9</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution: org.yaml:snakeyaml:1.32
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file unit conversion api pom xml path to vulnerable library root repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter actuator release jar root library spring boot starter release jar x snakeyaml jar vulnerable library vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stack overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
| 0
|
85,839
| 24,701,507,185
|
IssuesEvent
|
2022-10-19 15:36:00
|
joinmarket-webui/jam-docker
|
https://api.github.com/repos/joinmarket-webui/jam-docker
|
opened
|
fix github workflow warnings
|
dependencies build
|
Several warning when running the github workflow, e.g. see https://github.com/joinmarket-webui/jam-docker/actions/runs/3279592234
> Node.js 12 actions are deprecated. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/. Please update the following actions to use Node.js 16: actions/checkout, docker/setup-qemu-action, docker/setup-buildx-action, docker/login-action, docker/metadata-action, docker/build-push-action, docker/build-push-action, docker/login-action, docker/setup-buildx-action, actions/checkout
> The `save-state` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
> The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
|
1.0
|
fix github workflow warnings - Several warning when running the github workflow, e.g. see https://github.com/joinmarket-webui/jam-docker/actions/runs/3279592234
> Node.js 12 actions are deprecated. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/. Please update the following actions to use Node.js 16: actions/checkout, docker/setup-qemu-action, docker/setup-buildx-action, docker/login-action, docker/metadata-action, docker/build-push-action, docker/build-push-action, docker/login-action, docker/setup-buildx-action, actions/checkout
> The `save-state` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
> The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
|
build
|
fix github workflow warnings several warning when running the github workflow e g see node js actions are deprecated for more information see please update the following actions to use node js actions checkout docker setup qemu action docker setup buildx action docker login action docker metadata action docker build push action docker build push action docker login action docker setup buildx action actions checkout the save state command is deprecated and will be disabled soon please upgrade to using environment files for more information see the set output command is deprecated and will be disabled soon please upgrade to using environment files for more information see
| 1
|
271,994
| 23,645,227,943
|
IssuesEvent
|
2022-08-25 21:17:01
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
opened
|
Regen tasks should only overwrite existing files if successful
|
T-enhancement good first issue help wanted C-test-infra S-small
|
We have a bunch of regen-style tasks (testdata, diplomat-gen, fingerprint, ffi-coverage) where we generate a bunch of files on top of checked-in files (and typically, check the difference on CI)
They typically work by deleting the existing files, running the task, and then overwriting the file. This means that if the task fails, the files will be in a weird state that often needs to be reset from git. Furthermore, for testdata, `git status` is basically useless during the long-running testdata regeneration.
I think these tasks should run the generation to some temporary folder _first_, and only if the generation step succeeds, delete the old files and copy over new ones.
|
1.0
|
Regen tasks should only overwrite existing files if successful - We have a bunch of regen-style tasks (testdata, diplomat-gen, fingerprint, ffi-coverage) where we generate a bunch of files on top of checked-in files (and typically, check the difference on CI)
They typically work by deleting the existing files, running the task, and then overwriting the file. This means that if the task fails, the files will be in a weird state that often needs to be reset from git. Furthermore, for testdata, `git status` is basically useless during the long-running testdata regeneration.
I think these tasks should run the generation to some temporary folder _first_, and only if the generation step succeeds, delete the old files and copy over new ones.
|
non_build
|
regen tasks should only overwrite existing files if successful we have a bunch of regen style tasks testdata diplomat gen fingerprint ffi coverage where we generate a bunch of files on top of checked in files and typically check the difference on ci they typically work by deleting the existing files running the task and then overwriting the file this means that if the task fails the files will be in a weird state that often needs to be reset from git furthermore for testdata git status is basically useless during the long running testdata regeneration i think these tasks should run the generation to some temporary folder first and only if the generation step succeeds delete the old files and copy over new ones
| 0
|
67,564
| 17,014,694,953
|
IssuesEvent
|
2021-07-02 10:16:08
|
normade/modernism
|
https://api.github.com/repos/normade/modernism
|
opened
|
Link to City / Country
|
Building Detail Page enhancement
|
Find a way to link in detail content to a certain city or country. The link forwards on click to the building index page and sets the filter accordingly.
|
1.0
|
Link to City / Country - Find a way to link in detail content to a certain city or country. The link forwards on click to the building index page and sets the filter accordingly.
|
build
|
link to city country find a way to link in detail content to a certain city or country the link forwards on click to the building index page and sets the filter accordingly
| 1
|
96,553
| 27,883,387,839
|
IssuesEvent
|
2023-03-21 21:17:08
|
LLNL/DiHydrogen
|
https://api.github.com/repos/LLNL/DiHydrogen
|
closed
|
[H2Core] Logger flags require MPI linkage
|
bug build
|
The [logic for using MPI in the Logger flags](https://github.com/LLNL/DiHydrogen/blob/develop/src/utils/Logger.cpp#L15-L18) is actually incorrect. Finding the header is not sufficient as MPI linkage is not typically included "for free" with most compilers on most systems.
We can actually just tie right in to the DistConv MPI logic for now. Eventually H2 will require MPI, at least indirectly.
|
1.0
|
[H2Core] Logger flags require MPI linkage - The [logic for using MPI in the Logger flags](https://github.com/LLNL/DiHydrogen/blob/develop/src/utils/Logger.cpp#L15-L18) is actually incorrect. Finding the header is not sufficient as MPI linkage is not typically included "for free" with most compilers on most systems.
We can actually just tie right in to the DistConv MPI logic for now. Eventually H2 will require MPI, at least indirectly.
|
build
|
logger flags require mpi linkage the is actually incorrect finding the header is not sufficient as mpi linkage is not typically included for free with most compilers on most systems we can actually just tie right in to the distconv mpi logic for now eventually will require mpi at least indirectly
| 1
|
98,864
| 30,204,917,607
|
IssuesEvent
|
2023-07-05 08:47:22
|
BlueBrain/nmodl
|
https://api.github.com/repos/BlueBrain/nmodl
|
opened
|
Fix docker image deployed to DockerHub
|
good first issue build-system getting_started necessary-feature CI
|
Seems like the NMODL docker image uploaded to DockerHub is not working properly.
Following the guide in https://github.com/BlueBrain/nmodl/tree/master/packaging check whether it works and edit https://github.com/BlueBrain/nmodl/blob/master/packaging/Dockerfile to fix it
|
1.0
|
Fix docker image deployed to DockerHub - Seems like the NMODL docker image uploaded to DockerHub is not working properly.
Following the guide in https://github.com/BlueBrain/nmodl/tree/master/packaging check whether it works and edit https://github.com/BlueBrain/nmodl/blob/master/packaging/Dockerfile to fix it
|
build
|
fix docker image deployed to dockerhub seems like the nmodl docker image uploaded to dockerhub is not working properly following the guide in check whether it works and edit to fix it
| 1
|
201,207
| 15,180,567,591
|
IssuesEvent
|
2021-02-15 00:26:55
|
backend-br/vagas
|
https://api.github.com/repos/backend-br/vagas
|
closed
|
[São Paulo] Datahacker - Desenvolvedor Python Sênior @ Wivo
|
CI Docker Git Kubernetes MongoDB PJ Python Remoto SQL Stale TDD Testes automatizados startup
|
## Descrição da vaga
Se você é um programador:
1- Hands-on
2- Que adora programar
3- Que odeia ficar perdendo tempo com burocracias/reuniões
4- Que acredita que fazer código de qualidade é uma arte
Aqui é o seu lugar!
A Wivo é uma startup de tecnologia que possuí uma plataforma SaaS de inteligência de dados em Marketing e Vendas para empresas B2B.
Procuramos um(a) datahacker para fazer parte do nosso time de Tech que irá trabalhar com nosso:
1- Data engine (Crawlers, workers, parsers....)
2- Backend (API/Microserviços internos)
3- APIs e integrações externas
Buscamos por pessoas que tragam boas práticas de engenharia de software, como:
- Clean coding
- Design patterns
- Testes automatizados (TDD)
- Continuous Integration and Deployment (CI/CD)
Trabalhamos 100% remoto, com horários flexíveis e autonomia
## Local
São Paulo - São Paulo
## Requisitos
- Docker
- Kubernetes
- Git
- Microservices
- Python
- Postgre SQL
- MongoDB
## Contratação
PJ
## Como se candidatar
Se inscreva [clicando aqui](https://www.pyjobs.com.br/job/1881)
## Labels
|
1.0
|
[São Paulo] Datahacker - Desenvolvedor Python Sênior @ Wivo - ## Descrição da vaga
Se você é um programador:
1- Hands-on
2- Que adora programar
3- Que odeia ficar perdendo tempo com burocracias/reuniões
4- Que acredita que fazer código de qualidade é uma arte
Aqui é o seu lugar!
A Wivo é uma startup de tecnologia que possuí uma plataforma SaaS de inteligência de dados em Marketing e Vendas para empresas B2B.
Procuramos um(a) datahacker para fazer parte do nosso time de Tech que irá trabalhar com nosso:
1- Data engine (Crawlers, workers, parsers....)
2- Backend (API/Microserviços internos)
3- APIs e integrações externas
Buscamos por pessoas que tragam boas práticas de engenharia de software, como:
- Clean coding
- Design patterns
- Testes automatizados (TDD)
- Continuous Integration and Deployment (CI/CD)
Trabalhamos 100% remoto, com horários flexíveis e autonomia
## Local
São Paulo - São Paulo
## Requisitos
- Docker
- Kubernetes
- Git
- Microservices
- Python
- Postgre SQL
- MongoDB
## Contratação
PJ
## Como se candidatar
Se inscreva [clicando aqui](https://www.pyjobs.com.br/job/1881)
## Labels
|
non_build
|
datahacker desenvolvedor python sênior wivo descrição da vaga se você é um programador hands on que adora programar que odeia ficar perdendo tempo com burocracias reuniões que acredita que fazer código de qualidade é uma arte aqui é o seu lugar a wivo é uma startup de tecnologia que possuí uma plataforma saas de inteligência de dados em marketing e vendas para empresas procuramos um a datahacker para fazer parte do nosso time de tech que irá trabalhar com nosso data engine crawlers workers parsers backend api microserviços internos apis e integrações externas buscamos por pessoas que tragam boas práticas de engenharia de software como clean coding design patterns testes automatizados tdd continuous integration and deployment ci cd trabalhamos remoto com horários flexíveis e autonomia local são paulo são paulo requisitos docker kubernetes git microservices python postgre sql mongodb contratação pj como se candidatar se inscreva labels
| 0
|
55,666
| 6,912,730,579
|
IssuesEvent
|
2017-11-28 13:05:40
|
yoX64/dia-ui
|
https://api.github.com/repos/yoX64/dia-ui
|
closed
|
Splash screen design
|
design
|
We should have a minimalistic design for the splash screen. Also, you should implement.
|
1.0
|
Splash screen design - We should have a minimalistic design for the splash screen. Also, you should implement.
|
non_build
|
splash screen design we should have a minimalistic design for the splash screen also you should implement
| 0
|
14,781
| 4,996,064,436
|
IssuesEvent
|
2016-12-09 12:30:51
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Logging in using cookie (ie via remember me) doesnt set user id in #__session table
|
No Code Attached Yet
|
### Steps to reproduce the issue
Create clean 3.6.2 install
Login to back end and create user (eg DAVE)
Exit Browser
Delete all entries in Session table
Enter URL for front end of website, logon with created user ensure remember me ticked
Browse session table - guest = 0, userid and username set correctly in session table
Exit browser
Delete all entries in Session table
Enter URL for front end of table - no need to logon as user id remembered using cookie
Browse session table - guest = 1 , userid = 0, username blank. THIS IS INCORRECT
### Expected result
All authenticated users should have session record set with Guest = 0, userid and username set correctly
### Actual result
Authenticated user logged in using cookie has session record where Guest = 1 and userid and username not set
### System information (as much as possible)
### Additional comments
This happened on my website after upgrading from 3.6.0 to 3.6.2. To prove it wasnt my website I did a clean install of 3.6.2 on my PC using Xampp and proved error occured
|
1.0
|
Logging in using cookie (ie via remember me) doesnt set user id in #__session table - ### Steps to reproduce the issue
Create clean 3.6.2 install
Login to back end and create user (eg DAVE)
Exit Browser
Delete all entries in Session table
Enter URL for front end of website, logon with created user ensure remember me ticked
Browse session table - guest = 0, userid and username set correctly in session table
Exit browser
Delete all entries in Session table
Enter URL for front end of table - no need to logon as user id remembered using cookie
Browse session table - guest = 1 , userid = 0, username blank. THIS IS INCORRECT
### Expected result
All authenticated users should have session record set with Guest = 0, userid and username set correctly
### Actual result
Authenticated user logged in using cookie has session record where Guest = 1 and userid and username not set
### System information (as much as possible)
### Additional comments
This happened on my website after upgrading from 3.6.0 to 3.6.2. To prove it wasnt my website I did a clean install of 3.6.2 on my PC using Xampp and proved error occured
|
non_build
|
logging in using cookie ie via remember me doesnt set user id in session table steps to reproduce the issue create clean install login to back end and create user eg dave exit browser delete all entries in session table enter url for front end of website logon with created user ensure remember me ticked browse session table guest userid and username set correctly in session table exit browser delete all entries in session table enter url for front end of table no need to logon as user id remembered using cookie browse session table guest userid username blank this is incorrect expected result all authenticated users should have session record set with guest userid and username set correctly actual result authenticated user logged in using cookie has session record where guest and userid and username not set system information as much as possible additional comments this happened on my website after upgrading from to to prove it wasnt my website i did a clean install of on my pc using xampp and proved error occured
| 0
|
28,428
| 8,140,705,810
|
IssuesEvent
|
2018-08-20 22:06:23
|
JDRF/spirit
|
https://api.github.com/repos/JDRF/spirit
|
closed
|
Container
|
2 BUILD Components
|
### Deliver
* Kitchen sink file showing all variations in all states
* Macro to create all variations
* Update tokens file with necessary style decisions
|
1.0
|
Container - ### Deliver
* Kitchen sink file showing all variations in all states
* Macro to create all variations
* Update tokens file with necessary style decisions
|
build
|
container deliver kitchen sink file showing all variations in all states macro to create all variations update tokens file with necessary style decisions
| 1
|
16,753
| 6,278,268,670
|
IssuesEvent
|
2017-07-18 14:04:04
|
Tirocupidus/TheExiledRPOverhaul
|
https://api.github.com/repos/Tirocupidus/TheExiledRPOverhaul
|
closed
|
T3 Native gear and The Dregs
|
enhancement feat balance ready for build
|
Make T3 Native gear craftable without killing the boss in The Dregs. Also ensure the boss is worth killing for some other reason.
|
1.0
|
T3 Native gear and The Dregs - Make T3 Native gear craftable without killing the boss in The Dregs. Also ensure the boss is worth killing for some other reason.
|
build
|
native gear and the dregs make native gear craftable without killing the boss in the dregs also ensure the boss is worth killing for some other reason
| 1
|
91,085
| 26,271,497,511
|
IssuesEvent
|
2023-01-06 17:25:47
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
build: `dev build` fails with `Error: Cannot find module 'micromatch'`
|
C-bug A-build-system regression T-dev-inf
|
This is a regression from last week.
```
kena@kenax ....com/cockroachdb/cockroach % dev build [30/9284]
TMPDIR set to /tmp/tmp.B9nYBzFz
$ bazel build //pkg/cmd/cockroach:cockroach --config=with_ui
INFO: Invocation ID: 613e4a49-26f7-4bf1-bc86-a5cca8037e90
INFO: Analyzed target //pkg/cmd/cockroach:cockroach (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /data/home/kena/src/go/src/github.com/cockroachdb/cockroach/pkg/ui/workspaces/cluster-ui/BUILD.bazel:174:8: NpmPackageBin pkg/ui/workspaces/cluster-ui/d
ist/js/main.js failed: (Exit 1): webpack-cli.sh failed: error executing command bazel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin
/webpack-cli.sh pkg/ui/workspaces/cluster-ui/src/index.ts --config pkg/ui/workspaces/cluster-ui/webpack.config.js ... (remaining 7 arguments skipped)
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
Error: Cannot find module 'micromatch'
Require stack:
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/optimize/SideEff
ectsFlagPlugin.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/WebpackOptionsAp
ply.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/webpack.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/utils/valida
te-options.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/utils/conver
t-argv.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/cli.js
at Function.Module._resolveFilename (node:internal/modules/cjs/loader:995:15)
at Function.Module._load (node:internal/modules/cjs/loader:841:27)
at Module.require (node:internal/modules/cjs/loader:1067:19)
at require (node_modules/v8-compile-cache/v8-compile-cache.js:159:20)
at Object.<anonymous> (node_modules/webpack/lib/optimize/SideEffectsFlagPlugin.js:7:12)
at Module._compile (node_modules/v8-compile-cache/v8-compile-cache.js:192:30)
at Object.Module._extensions..js (node:internal/modules/cjs/loader:1219:10)
at Module.load (node:internal/modules/cjs/loader:1043:32)
at Function.Module._load (node:internal/modules/cjs/loader:878:12)
at Module.require (node:internal/modules/cjs/loader:1067:19)
at require (node_modules/v8-compile-cache/v8-compile-cache.js:159:20)
at WebpackOptionsApply.process (node_modules/webpack/lib/WebpackOptionsApply.js:345:34)
at webpack (node_modules/webpack/lib/webpack.js:57:48)
at processOptions (node_modules/webpack-cli/bin/cli.js:272:16)
at node_modules/webpack-cli/bin/cli.js:364:3
at Object.parse (node_modules/webpack-cli/node_modules/yargs/yargs.js:576:18)
at node_modules/webpack-cli/bin/cli.js:49:8
at Object.<anonymous> (node_modules/webpack-cli/bin/cli.js:366:3)
at Module._compile (node:internal/modules/cjs/loader:1165:14)
at Object.Module._extensions..js (node:internal/modules/cjs/loader:1219:10)
at Module.load (node:internal/modules/cjs/loader:1043:32)
at Function.Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:22:47
Target //pkg/cmd/cockroach:cockroach failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 14.143s, Critical Path: 4.83s
INFO: 3 processes: 3 internal.
INFO: Build Event Protocol files produced successfully.
FAILED: Build did NOT complete successfully
ERROR: exit status 1
```
cc @rickystewart @sjbarag for triage.
Jira issue: CRDB-23168
|
1.0
|
build: `dev build` fails with `Error: Cannot find module 'micromatch'` - This is a regression from last week.
```
kena@kenax ....com/cockroachdb/cockroach % dev build [30/9284]
TMPDIR set to /tmp/tmp.B9nYBzFz
$ bazel build //pkg/cmd/cockroach:cockroach --config=with_ui
INFO: Invocation ID: 613e4a49-26f7-4bf1-bc86-a5cca8037e90
INFO: Analyzed target //pkg/cmd/cockroach:cockroach (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /data/home/kena/src/go/src/github.com/cockroachdb/cockroach/pkg/ui/workspaces/cluster-ui/BUILD.bazel:174:8: NpmPackageBin pkg/ui/workspaces/cluster-ui/d
ist/js/main.js failed: (Exit 1): webpack-cli.sh failed: error executing command bazel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin
/webpack-cli.sh pkg/ui/workspaces/cluster-ui/src/index.ts --config pkg/ui/workspaces/cluster-ui/webpack.config.js ... (remaining 7 arguments skipped)
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
Error: Cannot find module 'micromatch'
Require stack:
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/optimize/SideEff
ectsFlagPlugin.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/WebpackOptionsAp
ply.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack/lib/webpack.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/utils/valida
te-options.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/utils/conver
t-argv.js
- /data/kena-cache/kena-cache/bazel/_bazel_kena/9b9f53b7334495d6a871cbb472c466bb/sandbox/processwrapper-sandbox/601/execroot/com_github_cockroachdb_cockroach/b
azel-out/freebsd-opt-exec-2B5CBBC6/bin/external/npm_cluster_ui/webpack-cli/bin/webpack-cli.sh.runfiles/npm_cluster_ui/node_modules/webpack-cli/bin/cli.js
at Function.Module._resolveFilename (node:internal/modules/cjs/loader:995:15)
at Function.Module._load (node:internal/modules/cjs/loader:841:27)
at Module.require (node:internal/modules/cjs/loader:1067:19)
at require (node_modules/v8-compile-cache/v8-compile-cache.js:159:20)
at Object.<anonymous> (node_modules/webpack/lib/optimize/SideEffectsFlagPlugin.js:7:12)
at Module._compile (node_modules/v8-compile-cache/v8-compile-cache.js:192:30)
at Object.Module._extensions..js (node:internal/modules/cjs/loader:1219:10)
at Module.load (node:internal/modules/cjs/loader:1043:32)
at Function.Module._load (node:internal/modules/cjs/loader:878:12)
at Module.require (node:internal/modules/cjs/loader:1067:19)
at require (node_modules/v8-compile-cache/v8-compile-cache.js:159:20)
at WebpackOptionsApply.process (node_modules/webpack/lib/WebpackOptionsApply.js:345:34)
at webpack (node_modules/webpack/lib/webpack.js:57:48)
at processOptions (node_modules/webpack-cli/bin/cli.js:272:16)
at node_modules/webpack-cli/bin/cli.js:364:3
at Object.parse (node_modules/webpack-cli/node_modules/yargs/yargs.js:576:18)
at node_modules/webpack-cli/bin/cli.js:49:8
at Object.<anonymous> (node_modules/webpack-cli/bin/cli.js:366:3)
at Module._compile (node:internal/modules/cjs/loader:1165:14)
at Object.Module._extensions..js (node:internal/modules/cjs/loader:1219:10)
at Module.load (node:internal/modules/cjs/loader:1043:32)
at Function.Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:22:47
Target //pkg/cmd/cockroach:cockroach failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 14.143s, Critical Path: 4.83s
INFO: 3 processes: 3 internal.
INFO: Build Event Protocol files produced successfully.
FAILED: Build did NOT complete successfully
ERROR: exit status 1
```
cc @rickystewart @sjbarag for triage.
Jira issue: CRDB-23168
|
build
|
build dev build fails with error cannot find module micromatch this is a regression from last week kena kenax com cockroachdb cockroach dev build tmpdir set to tmp tmp bazel build pkg cmd cockroach cockroach config with ui info invocation id info analyzed target pkg cmd cockroach cockroach packages loaded targets configured info found target error data home kena src go src github com cockroachdb cockroach pkg ui workspaces cluster ui build bazel npmpackagebin pkg ui workspaces cluster ui d ist js main js failed exit webpack cli sh failed error executing command bazel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh pkg ui workspaces cluster ui src index ts config pkg ui workspaces cluster ui webpack config js remaining arguments skipped use sandbox debug to see verbose messages from the sandbox and retain the sandbox build root for debugging error cannot find module micromatch require stack data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack lib optimize sideeff ectsflagplugin js data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack lib webpackoptionsap ply js data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack lib webpack js data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack cli bin utils valida te options js data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack cli bin utils conver t argv js data kena cache kena cache bazel bazel kena sandbox processwrapper sandbox execroot com github cockroachdb cockroach b azel out freebsd opt exec bin external npm cluster ui webpack cli bin webpack cli sh runfiles npm cluster ui node modules webpack cli bin cli js at function module resolvefilename node internal modules cjs loader at function module load node internal modules cjs loader at module require node internal modules cjs loader at require node modules compile cache compile cache js at object node modules webpack lib optimize sideeffectsflagplugin js at module compile node modules compile cache compile cache js at object module extensions js node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader at module require node internal modules cjs loader at require node modules compile cache compile cache js at webpackoptionsapply process node modules webpack lib webpackoptionsapply js at webpack node modules webpack lib webpack js at processoptions node modules webpack cli bin cli js at node modules webpack cli bin cli js at object parse node modules webpack cli node modules yargs yargs js at node modules webpack cli bin cli js at object node modules webpack cli bin cli js at module compile node internal modules cjs loader at object module extensions js node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader at function executeuserentrypoint node internal modules run main at node internal main run main module target pkg cmd cockroach cockroach failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal info build event protocol files produced successfully failed build did not complete successfully error exit status cc rickystewart sjbarag for triage jira issue crdb
| 1
|
53,529
| 13,175,284,730
|
IssuesEvent
|
2020-08-12 01:07:40
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
windows build issues(finally made it)
|
TF 2.2 stalled stat:awaiting response subtype:windows type:build/install
|
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): windows 10 x64
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version: 2.4.0
- Python version: 3.8.3 x64
- Installed using virtualenv? pip? conda?: N/A
- Bazel version (if compiling from source): 3.4.1
- GCC/Compiler version (if compiling from source): visual studio 2019
- CUDA/cuDNN version: 11.0/8.0.1
- GPU model and memory: RTX2070 GDDR6 8GB
**Describe the problem**
many problems with windows build, and I made a few fixes
**Provide the exact sequence of commands / steps that you executed before running into the problem**
1. cudnn_stub dependency problem ( https://github.com/tensorflow/tensorflow/issues/41057 )
I simply add dependency in that BUILD
2. then thrust version unmatch error
I delete thrust version - cub version comparison codes from
bazel-out\x64_windows-opt\bin\external\local_config_cuda\cuda\cuda\include\thrust\system\cuda\config.h
3. finally, big old numpy problem
this has been very long and unsolvable error, I made same fix as before( simply pip uninstall numpy ( uninstall conda numpy ) and pip install numpy )
and still there are too many "ignoring unknown options" from cl
anyway now I succeeded build and have my wheel file
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
1.0
|
windows build issues(finally made it) - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): windows 10 x64
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version: 2.4.0
- Python version: 3.8.3 x64
- Installed using virtualenv? pip? conda?: N/A
- Bazel version (if compiling from source): 3.4.1
- GCC/Compiler version (if compiling from source): visual studio 2019
- CUDA/cuDNN version: 11.0/8.0.1
- GPU model and memory: RTX2070 GDDR6 8GB
**Describe the problem**
many problems with windows build, and I made a few fixes
**Provide the exact sequence of commands / steps that you executed before running into the problem**
1. cudnn_stub dependency problem ( https://github.com/tensorflow/tensorflow/issues/41057 )
I simply add dependency in that BUILD
2. then thrust version unmatch error
I delete thrust version - cub version comparison codes from
bazel-out\x64_windows-opt\bin\external\local_config_cuda\cuda\cuda\include\thrust\system\cuda\config.h
3. finally, big old numpy problem
this has been very long and unsolvable error, I made same fix as before( simply pip uninstall numpy ( uninstall conda numpy ) and pip install numpy )
and still there are too many "ignoring unknown options" from cl
anyway now I succeeded build and have my wheel file
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
build
|
windows build issues finally made it please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary source tensorflow version python version installed using virtualenv pip conda n a bazel version if compiling from source gcc compiler version if compiling from source visual studio cuda cudnn version gpu model and memory describe the problem many problems with windows build and i made a few fixes provide the exact sequence of commands steps that you executed before running into the problem cudnn stub dependency problem i simply add dependency in that build then thrust version unmatch error i delete thrust version cub version comparison codes from bazel out windows opt bin external local config cuda cuda cuda include thrust system cuda config h finally big old numpy problem this has been very long and unsolvable error i made same fix as before simply pip uninstall numpy uninstall conda numpy and pip install numpy and still there are too many ignoring unknown options from cl anyway now i succeeded build and have my wheel file any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 1
|
127,580
| 27,079,617,830
|
IssuesEvent
|
2023-02-14 13:08:36
|
computationalgeography/lue
|
https://api.github.com/repos/computationalgeography/lue
|
opened
|
Add support for passing numpy values to LUE operations accepting scalar values
|
25: kind: enhancement 15: feature: code 05: target: framework / python 35: medium priority
|
Currently performing this operation results in a warning:
```python
np.float32(5) == my_array
```
DeprecationWarning: elementwise comparison failed; this will raise an error in the future.
|
1.0
|
Add support for passing numpy values to LUE operations accepting scalar values - Currently performing this operation results in a warning:
```python
np.float32(5) == my_array
```
DeprecationWarning: elementwise comparison failed; this will raise an error in the future.
|
non_build
|
add support for passing numpy values to lue operations accepting scalar values currently performing this operation results in a warning python np my array deprecationwarning elementwise comparison failed this will raise an error in the future
| 0
|
69,722
| 17,796,192,051
|
IssuesEvent
|
2021-08-31 22:40:13
|
CommunityToolkit/WindowsCommunityToolkit
|
https://api.github.com/repos/CommunityToolkit/WindowsCommunityToolkit
|
opened
|
Can't directly bind ObservableCollection to MetadataControl Items
|
bug :bug: controls :control_knobs: bugbash :building_construction:
|
## Describe the bug
You can't directly provide an `ObservableCollection` to the `MetadataControl` as it expects a `List` specifically (even though it supports observing).
- [ ] Is this bug a regression in the toolkit? If so, what toolkit version did you last see it work:
## Steps to Reproduce
```xml
<controls:MetadataControl
x:Name="Metadata"
AccessibleSeparator=","
Items="{x:Bind MyItems}"
Separator=" 🦙 " />
```
```cs
ObservableCollection<MetadataItem> MyItems { get; set; } = new();
```
## Expected behavior
Able to bind. Think we need to try using `IList` instead.
## Screenshots

## Environment
NuGet Package(s): 7.1.0-rc1
Package Version(s):
Windows 10 Build Number:
- [ ] Fall Creators Update (16299)
- [ ] April 2018 Update (17134)
- [ ] October 2018 Update (17763)
- [ ] May 2019 Update (18362)
- [ ] May 2020 Update (19041)
- [ ] Insider Build ({build_number})
App min and target version:
- [ ] Fall Creators Update (16299)
- [ ] April 2018 Update (17134)
- [ ] October 2018 Update (17763)
- [ ] May 2019 Update (18362)
- [ ] May 2020 Update (19041)
- [ ] Insider Build ({build_number})
Device form factor:
- [ ] Desktop
- [ ] Xbox
- [ ] Surface Hub
- [ ] IoT
Visual Studio version:
- [ ] 2017 (15.{minor_version})
- [ ] 2019 (16.{minor_version})
- [ ] 2022 (17.{minor_version})
## Additional context
<!-- Add any other context about the problem here. -->
|
1.0
|
Can't directly bind ObservableCollection to MetadataControl Items - ## Describe the bug
You can't directly provide an `ObservableCollection` to the `MetadataControl` as it expects a `List` specifically (even though it supports observing).
- [ ] Is this bug a regression in the toolkit? If so, what toolkit version did you last see it work:
## Steps to Reproduce
```xml
<controls:MetadataControl
x:Name="Metadata"
AccessibleSeparator=","
Items="{x:Bind MyItems}"
Separator=" 🦙 " />
```
```cs
ObservableCollection<MetadataItem> MyItems { get; set; } = new();
```
## Expected behavior
Able to bind. Think we need to try using `IList` instead.
## Screenshots

## Environment
NuGet Package(s): 7.1.0-rc1
Package Version(s):
Windows 10 Build Number:
- [ ] Fall Creators Update (16299)
- [ ] April 2018 Update (17134)
- [ ] October 2018 Update (17763)
- [ ] May 2019 Update (18362)
- [ ] May 2020 Update (19041)
- [ ] Insider Build ({build_number})
App min and target version:
- [ ] Fall Creators Update (16299)
- [ ] April 2018 Update (17134)
- [ ] October 2018 Update (17763)
- [ ] May 2019 Update (18362)
- [ ] May 2020 Update (19041)
- [ ] Insider Build ({build_number})
Device form factor:
- [ ] Desktop
- [ ] Xbox
- [ ] Surface Hub
- [ ] IoT
Visual Studio version:
- [ ] 2017 (15.{minor_version})
- [ ] 2019 (16.{minor_version})
- [ ] 2022 (17.{minor_version})
## Additional context
<!-- Add any other context about the problem here. -->
|
build
|
can t directly bind observablecollection to metadatacontrol items describe the bug you can t directly provide an observablecollection to the metadatacontrol as it expects a list specifically even though it supports observing is this bug a regression in the toolkit if so what toolkit version did you last see it work steps to reproduce xml controls metadatacontrol x name metadata accessibleseparator items x bind myitems separator 🦙 cs observablecollection myitems get set new expected behavior able to bind think we need to try using ilist instead screenshots environment nuget package s package version s windows build number fall creators update april update october update may update may update insider build build number app min and target version fall creators update april update october update may update may update insider build build number device form factor desktop xbox surface hub iot visual studio version minor version minor version minor version additional context
| 1
|
90,538
| 26,134,781,529
|
IssuesEvent
|
2022-12-29 10:37:44
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Compiling node on OpenBSD over arm: error: Fatal javascript OOM in MemoryChunk allocation failed during deserialization
|
build
|
### Version
not relevant; since compiling node v18
### Platform
OpenBSD devel0.my.domain 7.1 GENERIC.MP#0 arm64
### Subsystem
_No response_
### What steps will reproduce the bug?
- download node v18 and untar
- set CC=cc, CXX=c++, then run ./configure --prefix /home/shared/Builds/node_v18 --openssl-no-asm
- doas gmake install
The error occurs when compiling:
```
LD_LIBRARY_PATH=/home/shared/downloads/node/out/Release/lib.host:/home/shared/downloads/node/out/Release/lib.target:$LD_LIBRARY_PATH; export LD_LIBRARY_PATH; cd ../tools/v8_gypfiles; mkdir -p /home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni; "/home/shared/downloads/node/out/Release/mksnapshot" --turbo_instruction_scheduling "--target_os=openbsd" "--target_arch=arm64" --startup_src "/home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni/snapshot.cc" --embedded_variant Default --embedded_src "/home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni/embedded.S" --no-native-code-counter
```
The error:
```
#
# Fatal javascript OOM in MemoryChunk allocation failed during deserialization.
#
Trace/BPT trap (core dumped)
gmake[1]: *** [tools/v8_gypfiles/v8_snapshot.target.mk:17: d82c33e09dcee352fdb851b52dad3bf6424bc495.intermediate] Error 133
rm 5aa268fa9eee95a410a10a3c2e32821ec67bb9ab.intermediate b0a101a4ac720edd9c6a3ff9ae16f9387689aa6c.intermediate d82c33e09dcee352fdb851b52dad3bf6424bc495.intermediate 5103e04f66aefbf194b8c8fb69deb47b3f4b97a4.intermediate
```
The machine has 8GB which I think is plenty for compilation. There is no limit as on the process.
### How often does it reproduce? Is there a required condition?
It reproduces always
### What is the expected behavior?
Should be able to compile.
### What do you see instead?
Compiler exits with error.
### Additional information
There could be some openBSD quirks but I don't know what to check to confirm that.
|
1.0
|
Compiling node on OpenBSD over arm: error: Fatal javascript OOM in MemoryChunk allocation failed during deserialization - ### Version
not relevant; since compiling node v18
### Platform
OpenBSD devel0.my.domain 7.1 GENERIC.MP#0 arm64
### Subsystem
_No response_
### What steps will reproduce the bug?
- download node v18 and untar
- set CC=cc, CXX=c++, then run ./configure --prefix /home/shared/Builds/node_v18 --openssl-no-asm
- doas gmake install
The error occurs when compiling:
```
LD_LIBRARY_PATH=/home/shared/downloads/node/out/Release/lib.host:/home/shared/downloads/node/out/Release/lib.target:$LD_LIBRARY_PATH; export LD_LIBRARY_PATH; cd ../tools/v8_gypfiles; mkdir -p /home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni; "/home/shared/downloads/node/out/Release/mksnapshot" --turbo_instruction_scheduling "--target_os=openbsd" "--target_arch=arm64" --startup_src "/home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni/snapshot.cc" --embedded_variant Default --embedded_src "/home/shared/downloads/node/out/Release/obj.target/v8_snapshot/geni/embedded.S" --no-native-code-counter
```
The error:
```
#
# Fatal javascript OOM in MemoryChunk allocation failed during deserialization.
#
Trace/BPT trap (core dumped)
gmake[1]: *** [tools/v8_gypfiles/v8_snapshot.target.mk:17: d82c33e09dcee352fdb851b52dad3bf6424bc495.intermediate] Error 133
rm 5aa268fa9eee95a410a10a3c2e32821ec67bb9ab.intermediate b0a101a4ac720edd9c6a3ff9ae16f9387689aa6c.intermediate d82c33e09dcee352fdb851b52dad3bf6424bc495.intermediate 5103e04f66aefbf194b8c8fb69deb47b3f4b97a4.intermediate
```
The machine has 8GB which I think is plenty for compilation. There is no limit as on the process.
### How often does it reproduce? Is there a required condition?
It reproduces always
### What is the expected behavior?
Should be able to compile.
### What do you see instead?
Compiler exits with error.
### Additional information
There could be some openBSD quirks but I don't know what to check to confirm that.
|
build
|
compiling node on openbsd over arm error fatal javascript oom in memorychunk allocation failed during deserialization version not relevant since compiling node platform openbsd my domain generic mp subsystem no response what steps will reproduce the bug download node and untar set cc cc cxx c then run configure prefix home shared builds node openssl no asm doas gmake install the error occurs when compiling ld library path home shared downloads node out release lib host home shared downloads node out release lib target ld library path export ld library path cd tools gypfiles mkdir p home shared downloads node out release obj target snapshot geni home shared downloads node out release mksnapshot turbo instruction scheduling target os openbsd target arch startup src home shared downloads node out release obj target snapshot geni snapshot cc embedded variant default embedded src home shared downloads node out release obj target snapshot geni embedded s no native code counter the error fatal javascript oom in memorychunk allocation failed during deserialization trace bpt trap core dumped gmake error rm intermediate intermediate intermediate intermediate the machine has which i think is plenty for compilation there is no limit as on the process how often does it reproduce is there a required condition it reproduces always what is the expected behavior should be able to compile what do you see instead compiler exits with error additional information there could be some openbsd quirks but i don t know what to check to confirm that
| 1
|
19,044
| 6,664,776,254
|
IssuesEvent
|
2017-10-02 21:26:05
|
DestinyItemManager/DIM
|
https://api.github.com/repos/DestinyItemManager/DIM
|
closed
|
Touch Friendly way to and/or perks in Loadout Builder
|
Enhancement Feature: Loadout Builder Need More Details
|
The requirement for keyboard use to enable this feature isn't useful for touch users on tablets. I use a Surface Pro when playing.
|
1.0
|
Touch Friendly way to and/or perks in Loadout Builder - The requirement for keyboard use to enable this feature isn't useful for touch users on tablets. I use a Surface Pro when playing.
|
build
|
touch friendly way to and or perks in loadout builder the requirement for keyboard use to enable this feature isn t useful for touch users on tablets i use a surface pro when playing
| 1
|
28,641
| 8,196,491,366
|
IssuesEvent
|
2018-08-31 10:00:05
|
angular/angular-cli
|
https://api.github.com/repos/angular/angular-cli
|
closed
|
While making a build of angular-6 application am getting this error how i can resolve this?
|
comp: devkit/build-angular need: repro steps
|
F:\practice\ang-node\frontend>ng build --prod
Date: 2018-08-28T04:08:06.648Z
Hash: 61fab25db19a36ad55ca
Time: 89374ms
chunk {0} runtime.a66f828dca56eeb90e02.js (runtime) 1.05 kB [entry] [rendered]
chunk {1} styles.78fc82606a60ff243473.css (styles) 193 kB [initial] [rendered]
chunk {2} polyfills.7a0e6866a34e280f48e7.js (polyfills) 59.6 kB [initial] [rendered]
chunk {3} main.b6f6160fdfad736b2ba4.js (main) 1.18 MB [initial] [rendered]
ERROR in ./node_modules/angular2-jwt/angular2-jwt.js
Module build failed: Error: Debug Failure. False expression.
at getJSDocTags (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13197:22)
at getFirstJSDocTag (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13205:24)
at Object.getJSDocTypeTag (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13154:19)
at checkParenthesizedExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45628:56)
at checkExpressionWorker (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45676:28)
at checkExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45609:42)
at checkBinaryLikeExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45155:29)
at checkBinaryExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45147:20)
at checkExpressionWorker (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45702:28)
at checkExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45609:42)
at checkExpressionCached (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45457:38)
at getTypeOfExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45577:28)
at checkDeclarationInitializer (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45469:24)
at getTypeForVariableLikeDeclaration (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:30738:28)
at getWidenedTypeForVariableLikeDeclaration (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:30961:24)
at getTypeOfVariableOrParameterOrProperty (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:31062:28)
F:\practice\ang-node\frontend>
|
1.0
|
While making a build of angular-6 application am getting this error how i can resolve this? - F:\practice\ang-node\frontend>ng build --prod
Date: 2018-08-28T04:08:06.648Z
Hash: 61fab25db19a36ad55ca
Time: 89374ms
chunk {0} runtime.a66f828dca56eeb90e02.js (runtime) 1.05 kB [entry] [rendered]
chunk {1} styles.78fc82606a60ff243473.css (styles) 193 kB [initial] [rendered]
chunk {2} polyfills.7a0e6866a34e280f48e7.js (polyfills) 59.6 kB [initial] [rendered]
chunk {3} main.b6f6160fdfad736b2ba4.js (main) 1.18 MB [initial] [rendered]
ERROR in ./node_modules/angular2-jwt/angular2-jwt.js
Module build failed: Error: Debug Failure. False expression.
at getJSDocTags (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13197:22)
at getFirstJSDocTag (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13205:24)
at Object.getJSDocTypeTag (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:13154:19)
at checkParenthesizedExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45628:56)
at checkExpressionWorker (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45676:28)
at checkExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45609:42)
at checkBinaryLikeExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45155:29)
at checkBinaryExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45147:20)
at checkExpressionWorker (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45702:28)
at checkExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45609:42)
at checkExpressionCached (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45457:38)
at getTypeOfExpression (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45577:28)
at checkDeclarationInitializer (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:45469:24)
at getTypeForVariableLikeDeclaration (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:30738:28)
at getWidenedTypeForVariableLikeDeclaration (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:30961:24)
at getTypeOfVariableOrParameterOrProperty (F:\practice\ang-node\frontend\node_modules\@angular-devkit\build-optimizer\node_modules\typescript\lib\typescript.js:31062:28)
F:\practice\ang-node\frontend>
|
build
|
while making a build of angular application am getting this error how i can resolve this f practice ang node frontend ng build prod date hash time chunk runtime js runtime kb chunk styles css styles kb chunk polyfills js polyfills kb chunk main js main mb error in node modules jwt jwt js module build failed error debug failure false expression at getjsdoctags f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at getfirstjsdoctag f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at object getjsdoctypetag f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkparenthesizedexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkexpressionworker f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkbinarylikeexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkbinaryexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkexpressionworker f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkexpressioncached f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at gettypeofexpression f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at checkdeclarationinitializer f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at gettypeforvariablelikedeclaration f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at getwidenedtypeforvariablelikedeclaration f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js at gettypeofvariableorparameterorproperty f practice ang node frontend node modules angular devkit build optimizer node modules typescript lib typescript js f practice ang node frontend
| 1
|
97,526
| 28,312,613,004
|
IssuesEvent
|
2023-04-10 16:44:55
|
dotnet/arcade
|
https://api.github.com/repos/dotnet/arcade
|
closed
|
Build failed: Maestro Build Promotion/main #Promoting dotnet-arcade-validation build 20230409.1 (173556) to channel(s) '.NET Eng - Validation' #
|
First Responder Build Failed
|
Build [#Promoting dotnet-arcade-validation build 20230409.1 (173556) to channel(s) '.NET Eng - Validation' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2153964) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Mon, 10 Apr 2023 00:58:02 GMT
**Duration** - 1 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Access Denied: The Personal Access Token used has expired.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Publishing symbols failed :
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Access Denied: The Personal Access Token used has expired.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Publishing symbols failed :
### Changes
|
1.0
|
Build failed: Maestro Build Promotion/main #Promoting dotnet-arcade-validation build 20230409.1 (173556) to channel(s) '.NET Eng - Validation' # - Build [#Promoting dotnet-arcade-validation build 20230409.1 (173556) to channel(s) '.NET Eng - Validation' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2153964) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Mon, 10 Apr 2023 00:58:02 GMT
**Duration** - 1 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Access Denied: The Personal Access Token used has expired.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Publishing symbols failed :
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Access Denied: The Personal Access Token used has expired.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2153964/logs/24) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23205.4\tools\SdkTasks\PublishArtifactsInManifest.proj(138,5): error : Publishing symbols failed :
### Changes
|
build
|
build failed maestro build promotion main promoting dotnet arcade validation build to channel s net eng validation build failed x internal maestro build promotion failed summary finished mon apr gmt duration minutes requested for dotnet bot reason manual details publishing x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error access denied the personal access token used has expired x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error publishing symbols failed x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error access denied the personal access token used has expired x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error publishing symbols failed changes
| 1
|
37,214
| 9,979,386,782
|
IssuesEvent
|
2019-07-09 22:40:54
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
Handle deploy for multiple images(artifacts) in helm values
|
area/build area/deploy deploy/helm help wanted kind/bug priority/p2
|
Hello, I am working on a project that has multiple docker images and I pass those images to helm through chart values. For the first time the `dev` command works well, but when I modify some files that trigger the build for only one artifact the deploy fails with the following warning:
```
WARN[0069] run: deploy step: deploying test: matching build results to chart values: No build present for image1
```
But I was expecting to deploy the new values with the new image. I think that the function [JoinTagsToBuildResult](https://github.com/GoogleCloudPlatform/skaffold/blob/master/pkg/skaffold/deploy/deploy.go#L38) should match the builds with the chart values not the other way.
skaffold.yaml:
```
apiVersion: skaffold/v1
kind: Config
build:
tagPolicy: sha256
artifacts:
- imageName: image1
dockerfilePath: hack/docker/toolbox/Dockerfile
workspace: .
- imageName: image2
workspace: .
local: {}
deploy:
helm:
releases:
- name: test
chartPath: hack/charts/controller
values:
image: image1
conf.toolbox-image: image2
```
skaffold version is `v0.2.0`
Thank you
|
1.0
|
Handle deploy for multiple images(artifacts) in helm values - Hello, I am working on a project that has multiple docker images and I pass those images to helm through chart values. For the first time the `dev` command works well, but when I modify some files that trigger the build for only one artifact the deploy fails with the following warning:
```
WARN[0069] run: deploy step: deploying test: matching build results to chart values: No build present for image1
```
But I was expecting to deploy the new values with the new image. I think that the function [JoinTagsToBuildResult](https://github.com/GoogleCloudPlatform/skaffold/blob/master/pkg/skaffold/deploy/deploy.go#L38) should match the builds with the chart values not the other way.
skaffold.yaml:
```
apiVersion: skaffold/v1
kind: Config
build:
tagPolicy: sha256
artifacts:
- imageName: image1
dockerfilePath: hack/docker/toolbox/Dockerfile
workspace: .
- imageName: image2
workspace: .
local: {}
deploy:
helm:
releases:
- name: test
chartPath: hack/charts/controller
values:
image: image1
conf.toolbox-image: image2
```
skaffold version is `v0.2.0`
Thank you
|
build
|
handle deploy for multiple images artifacts in helm values hello i am working on a project that has multiple docker images and i pass those images to helm through chart values for the first time the dev command works well but when i modify some files that trigger the build for only one artifact the deploy fails with the following warning warn run deploy step deploying test matching build results to chart values no build present for but i was expecting to deploy the new values with the new image i think that the function should match the builds with the chart values not the other way skaffold yaml apiversion skaffold kind config build tagpolicy artifacts imagename dockerfilepath hack docker toolbox dockerfile workspace imagename workspace local deploy helm releases name test chartpath hack charts controller values image conf toolbox image skaffold version is thank you
| 1
|
74,243
| 14,224,120,365
|
IssuesEvent
|
2020-11-17 19:11:40
|
foss-responders/fossresponders.com
|
https://api.github.com/repos/foss-responders/fossresponders.com
|
closed
|
Add jekyll-assets plugin
|
code
|
Right now, the Privacy page was removed partially because the asset path changes for images on secondary pages. We need an asset manager. Jekyll-asset should work. This will be important for the blog, too.
|
1.0
|
Add jekyll-assets plugin - Right now, the Privacy page was removed partially because the asset path changes for images on secondary pages. We need an asset manager. Jekyll-asset should work. This will be important for the blog, too.
|
non_build
|
add jekyll assets plugin right now the privacy page was removed partially because the asset path changes for images on secondary pages we need an asset manager jekyll asset should work this will be important for the blog too
| 0
|
130,264
| 27,637,242,070
|
IssuesEvent
|
2023-03-10 15:18:38
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
opened
|
Code Table Request - Fort Worth Zoo
|
Function-CodeTables
|
## Instructions
This is a template to facilitate communication with the Arctos Code Table Committee. Submit a separate request for each relevant value. This form is appropriate for exploring how data may best be stored, for adding vocabulary, or for updating existing definitions.
Reviewing documentation before proceeding will result in a more enjoyable experience.
* [Issue Documentation](http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html)
* [Code Table Documentation](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html)
* [Video Tutorial - Submit a Code Table Request](https://youtu.be/t2jHbsRA3lk)
------------------------------
## Initial Request
**Goal**: Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step.
All [Fort Worth Zoo](https://arctos.database.museum/info/ctDocumentation.cfm?table=ctcoll_other_id_type#fort_worth_zoo) should be replaced with other ID type = other identifier and issued by agent [Fort Worth Zoo](https://arctos.database.museum/agent/10014654)
**Proposed Value**: Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos.
**Proposed Definition**: Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable.
**Context**: Describe why this new value is necessary and existing values are not.
**Table**: Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure.
**Collection type**: Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm
**Priority**: Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking.
**Available for Public View**: Most data are by default publicly available. Describe any necessary access restrictions.
**Project**: Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184).
**Discussion**: Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary.
## Approval
All of the following must be checked before this may proceed.
The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality).
- [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval
- [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval
- [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code.
- [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention)
## Rejection
If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc.
1. Can a suitable solution be found here? If not, proceed to (2)
2. Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3)
3. Take the discussion to a monthly Arctos Working Group meeting for final resolution.
## Implementation
Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made.
Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel.
Make changes as described above. Ensure the URL of this Issue is included in the definition.
Close this Issue.
**DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result.
## Special Exemptions
In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases.
1. Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step.
2. The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology.
|
1.0
|
Code Table Request - Fort Worth Zoo - ## Instructions
This is a template to facilitate communication with the Arctos Code Table Committee. Submit a separate request for each relevant value. This form is appropriate for exploring how data may best be stored, for adding vocabulary, or for updating existing definitions.
Reviewing documentation before proceeding will result in a more enjoyable experience.
* [Issue Documentation](http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html)
* [Code Table Documentation](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html)
* [Video Tutorial - Submit a Code Table Request](https://youtu.be/t2jHbsRA3lk)
------------------------------
## Initial Request
**Goal**: Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step.
All [Fort Worth Zoo](https://arctos.database.museum/info/ctDocumentation.cfm?table=ctcoll_other_id_type#fort_worth_zoo) should be replaced with other ID type = other identifier and issued by agent [Fort Worth Zoo](https://arctos.database.museum/agent/10014654)
**Proposed Value**: Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos.
**Proposed Definition**: Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable.
**Context**: Describe why this new value is necessary and existing values are not.
**Table**: Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure.
**Collection type**: Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm
**Priority**: Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking.
**Available for Public View**: Most data are by default publicly available. Describe any necessary access restrictions.
**Project**: Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184).
**Discussion**: Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary.
## Approval
All of the following must be checked before this may proceed.
The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality).
- [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval
- [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval
- [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code.
- [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention)
## Rejection
If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc.
1. Can a suitable solution be found here? If not, proceed to (2)
2. Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3)
3. Take the discussion to a monthly Arctos Working Group meeting for final resolution.
## Implementation
Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made.
Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel.
Make changes as described above. Ensure the URL of this Issue is included in the definition.
Close this Issue.
**DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result.
## Special Exemptions
In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases.
1. Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step.
2. The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology.
|
non_build
|
code table request fort worth zoo instructions this is a template to facilitate communication with the arctos code table committee submit a separate request for each relevant value this form is appropriate for exploring how data may best be stored for adding vocabulary or for updating existing definitions reviewing documentation before proceeding will result in a more enjoyable experience initial request goal describe what you re trying to accomplish this is the only necessary step to start this process the committee is available to assist with all other steps please clearly indicate any uncertainty or desired guidance if you proceed beyond this step all should be replaced with other id type other identifier and issued by agent proposed value proposed new value this should be clear and compatible with similar values in the relevant table and across arctos proposed definition clear complete non collection type specific functional definition of the value avoid discipline specific terminology if possible include parenthetically if unavoidable context describe why this new value is necessary and existing values are not table code tables are link to the specific table or value this may involve multiple tables and will control datatype for attributes otherid requests require baseurl and example or explanation please ask for assistance if unsure collection type some code tables contain collection type specific values collection cde may be found from priority please describe the urgency and or choose a priority label to the right you should expect a response within two working days and may utilize if you feel response is lacking available for public view most data are by default publicly available describe any necessary access restrictions project add the issue to the discussion please reach out to anyone who might be affected by this change leave a comment or add this to the committee agenda if you believe more focused conversation is necessary approval all of the following must be checked before this may proceed the should be followed pay particular attention to terminology with emphasis on consistency and documentation with emphasis on functionality code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval dba the request is functionally acceptable the term is not a functional duplicate and is compatible with existing data and code dba appropriate code or handlers are in place as necessary id references media relationships encumbrances etc require particular attention rejection if you believe this request should not proceed explain why here suggest any changes that would make the change acceptable alternate usually existing paths to the same goals etc can a suitable solution be found here if not proceed to can a suitable solution be found by code table committee discussion if not proceed to take the discussion to a monthly arctos working group meeting for final resolution implementation once all of the approval checklist is appropriately checked and there are no rejection comments or in special circumstances by decree of the arctos working group the change may be made review everything one last time ensure the how to has been followed ensure all checks have been made by appropriate personnel make changes as described above ensure the url of this issue is included in the definition close this issue do not modify arctos authorities in any way before all points in this issue have been fully addressed data loss may result special exemptions in very specific cases and by prior approval of the committee the approval process may be skipped and implementation requirements may be slightly altered please note here if you are proceeding under one of these use cases adding an existing term to additional collection types may proceed immediately and without discussion but doing so may also subject users to future cleanup efforts if time allows please review the term and definition as part of this step the committee may grant special access on particular tables to particular users this should be exercised with great caution only after several smooth test cases and generally limited to taxonomy like data such as international commission on stratigraphy terminology
| 0
|
244,662
| 20,682,304,472
|
IssuesEvent
|
2022-03-10 14:56:35
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_filtering·ts - dashboard app using current data dashboard filtering disabling a filter unfilters the data on "before all" hook for "pie charts"
|
triage_needed Team:Presentation failed-test
|
A test failed on a tracked branch
```
TimeoutError: Waiting for element to be located By(css selector, .euiFilterSelectItem)
Wait timed out after 10056ms
at /dev/shm/workspace/parallel/20/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:93:5) {
remoteStacktrace: ''
}
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/13112/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_filtering·ts","test.name":"dashboard app using current data dashboard filtering disabling a filter unfilters the data on \"before all\" hook for \"pie charts\"","test.failCount":1}} -->
|
1.0
|
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_filtering·ts - dashboard app using current data dashboard filtering disabling a filter unfilters the data on "before all" hook for "pie charts" - A test failed on a tracked branch
```
TimeoutError: Waiting for element to be located By(css selector, .euiFilterSelectItem)
Wait timed out after 10056ms
at /dev/shm/workspace/parallel/20/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:93:5) {
remoteStacktrace: ''
}
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/13112/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_filtering·ts","test.name":"dashboard app using current data dashboard filtering disabling a filter unfilters the data on \"before all\" hook for \"pie charts\"","test.failCount":1}} -->
|
non_build
|
failing test chrome ui functional tests test functional apps dashboard dashboard filtering·ts dashboard app using current data dashboard filtering disabling a filter unfilters the data on before all hook for pie charts a test failed on a tracked branch timeouterror waiting for element to be located by css selector euifilterselectitem wait timed out after at dev shm workspace parallel kibana node modules selenium webdriver lib webdriver js at runmicrotasks at processticksandrejections internal process task queues js remotestacktrace first failure
| 0
|
121,119
| 4,805,137,089
|
IssuesEvent
|
2016-11-02 15:20:41
|
isawnyu/pleiades-gazetteer
|
https://api.github.com/repos/isawnyu/pleiades-gazetteer
|
closed
|
404s on missing images
|
cleanup priority: high review
|
Loading a place makes a few requests for images that are not found; we should fix these or remove the references to the images so the server doesn't have to spend time serving these 404s.

|
1.0
|
404s on missing images - Loading a place makes a few requests for images that are not found; we should fix these or remove the references to the images so the server doesn't have to spend time serving these 404s.

|
non_build
|
on missing images loading a place makes a few requests for images that are not found we should fix these or remove the references to the images so the server doesn t have to spend time serving these
| 0
|
93,755
| 27,029,816,477
|
IssuesEvent
|
2023-02-12 03:04:01
|
llvm/llvm-project
|
https://api.github.com/repos/llvm/llvm-project
|
closed
|
ninja check-clang command fails with multiple errors despite no errors during compilation process
|
build-problem platform:windows
|
Running on Windows Version 10.0.19044.2486
Steps to reproduce:
1) Cloned git repo of [Moore HDL](https://github.com/fabianschuiki/moore)
2) Ran `git submodule update --init --recursive`
3) Built and compiled with `cmake -G Ninja ..\llvm -DCMAKE_BUILD_TYPE=DEBUG -DLLVM_ENABLE_ASSERTIONS=ON -DLLVM_ENABLE_PROJECTS="mlir;lld;clang" -DLLVM_INSTALL_UTILS=ON -DLLVM_TARGETS_TO_BUILD="host;RISCV"` followed by `ninja`. Build and compilation both succeeded.
4) Ran `ninja check-mlir` with [this](https://pastebin.com/KF6xeSPx) output. (I have no clue if this is relevant, I included it because it's the last command I ran that I can still view the output of.)
5) Ran `ninja check-clang`. Build subcommand failed, full command line output [here](https://pastebin.com/vfdb7ykr)
|
1.0
|
ninja check-clang command fails with multiple errors despite no errors during compilation process - Running on Windows Version 10.0.19044.2486
Steps to reproduce:
1) Cloned git repo of [Moore HDL](https://github.com/fabianschuiki/moore)
2) Ran `git submodule update --init --recursive`
3) Built and compiled with `cmake -G Ninja ..\llvm -DCMAKE_BUILD_TYPE=DEBUG -DLLVM_ENABLE_ASSERTIONS=ON -DLLVM_ENABLE_PROJECTS="mlir;lld;clang" -DLLVM_INSTALL_UTILS=ON -DLLVM_TARGETS_TO_BUILD="host;RISCV"` followed by `ninja`. Build and compilation both succeeded.
4) Ran `ninja check-mlir` with [this](https://pastebin.com/KF6xeSPx) output. (I have no clue if this is relevant, I included it because it's the last command I ran that I can still view the output of.)
5) Ran `ninja check-clang`. Build subcommand failed, full command line output [here](https://pastebin.com/vfdb7ykr)
|
build
|
ninja check clang command fails with multiple errors despite no errors during compilation process running on windows version steps to reproduce cloned git repo of ran git submodule update init recursive built and compiled with cmake g ninja llvm dcmake build type debug dllvm enable assertions on dllvm enable projects mlir lld clang dllvm install utils on dllvm targets to build host riscv followed by ninja build and compilation both succeeded ran ninja check mlir with output i have no clue if this is relevant i included it because it s the last command i ran that i can still view the output of ran ninja check clang build subcommand failed full command line output
| 1
|
57,967
| 14,247,614,369
|
IssuesEvent
|
2020-11-19 11:43:02
|
googleapis/java-asset
|
https://api.github.com/repos/googleapis/java-asset
|
reopened
|
com.example.asset.QuickStartIT: testExportAssetBigqueryExample failed
|
:rotating_light: api: cloudasset buildcop: flaky buildcop: issue priority: p1 type: bug
|
Note: #311 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: ff04d0f62c5fd42146ea30ef3de2c9d4baba13fe
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/68b318d0-84dd-4d1f-96a5-8c14c12a4817), [Sponge](http://sponge2/68b318d0-84dd-4d1f-96a5-8c14c12a4817)
status: failed
<details><summary>Test output</summary><br><pre>java.util.concurrent.CancellationException: Task was cancelled.
at com.google.common.util.concurrent.AbstractFuture.cancellationExceptionWithCause(AbstractFuture.java:1427)
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:564)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:86)
at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:62)
at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:127)
at com.example.asset.ExportAssetsBigqueryExample.exportBigQuery(ExportAssetsBigqueryExample.java:54)
at com.example.asset.QuickStartIT.testExportAssetBigqueryExample(QuickStartIT.java:100)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details>
|
2.0
|
com.example.asset.QuickStartIT: testExportAssetBigqueryExample failed - Note: #311 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: ff04d0f62c5fd42146ea30ef3de2c9d4baba13fe
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/68b318d0-84dd-4d1f-96a5-8c14c12a4817), [Sponge](http://sponge2/68b318d0-84dd-4d1f-96a5-8c14c12a4817)
status: failed
<details><summary>Test output</summary><br><pre>java.util.concurrent.CancellationException: Task was cancelled.
at com.google.common.util.concurrent.AbstractFuture.cancellationExceptionWithCause(AbstractFuture.java:1427)
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:564)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:86)
at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:62)
at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:127)
at com.example.asset.ExportAssetsBigqueryExample.exportBigQuery(ExportAssetsBigqueryExample.java:54)
at com.example.asset.QuickStartIT.testExportAssetBigqueryExample(QuickStartIT.java:100)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details>
|
build
|
com example asset quickstartit testexportassetbigqueryexample failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output java util concurrent cancellationexception task was cancelled at com google common util concurrent abstractfuture cancellationexceptionwithcause abstractfuture java at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com example asset exportassetsbigqueryexample exportbigquery exportassetsbigqueryexample java at com example asset quickstartit testexportassetbigqueryexample quickstartit java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java
| 1
|
8,690
| 27,172,094,289
|
IssuesEvent
|
2023-02-17 20:27:02
|
OneDrive/onedrive-api-docs
|
https://api.github.com/repos/OneDrive/onedrive-api-docs
|
closed
|
Add documentation for photo albums
|
type:enhancement area:Docs automation:Closed
|
Hi there,
please add documentation for the album function in OneDrive. Because there are some issues for this point.
#468
#579
#909
Thanks to @ksmutny and his issue #909 here I have find a solution that work for me. But this was not an easy way.
I hope we can setup fast this points for the documentation:
- get a list of all Albums
- get media in an album
- create an album
- delete an album
- add media to an album
|
1.0
|
Add documentation for photo albums - Hi there,
please add documentation for the album function in OneDrive. Because there are some issues for this point.
#468
#579
#909
Thanks to @ksmutny and his issue #909 here I have find a solution that work for me. But this was not an easy way.
I hope we can setup fast this points for the documentation:
- get a list of all Albums
- get media in an album
- create an album
- delete an album
- add media to an album
|
non_build
|
add documentation for photo albums hi there please add documentation for the album function in onedrive because there are some issues for this point thanks to ksmutny and his issue here i have find a solution that work for me but this was not an easy way i hope we can setup fast this points for the documentation get a list of all albums get media in an album create an album delete an album add media to an album
| 0
|
93,157
| 26,876,359,871
|
IssuesEvent
|
2023-02-05 03:49:48
|
dotnet/docfx
|
https://api.github.com/repos/dotnet/docfx
|
closed
|
Stack overflow exception while generating documentation
|
bug dotnet: msbuild dotnet: xml-comment
|
**Operating System**: Linux
**DocFX Version Used**: 2.56.7
**Template used**: default and material
**Steps to Reproduce**:
1. Clone [this](https://github.com/zsr2531/Bali) repo.
2. `cd docs`
3. `mono /path/to/docfx.exe docfx.json --serve`
**Expected Behavior**:
Documentation is generated and a local HTTP server is started.
**Actual Behavior**:
Docfx runs into a stack overflow and terminates. However, running it a second time (seemingly) fixes the issue and then it serves the documentation correctly.
|
1.0
|
Stack overflow exception while generating documentation - **Operating System**: Linux
**DocFX Version Used**: 2.56.7
**Template used**: default and material
**Steps to Reproduce**:
1. Clone [this](https://github.com/zsr2531/Bali) repo.
2. `cd docs`
3. `mono /path/to/docfx.exe docfx.json --serve`
**Expected Behavior**:
Documentation is generated and a local HTTP server is started.
**Actual Behavior**:
Docfx runs into a stack overflow and terminates. However, running it a second time (seemingly) fixes the issue and then it serves the documentation correctly.
|
build
|
stack overflow exception while generating documentation operating system linux docfx version used template used default and material steps to reproduce clone repo cd docs mono path to docfx exe docfx json serve expected behavior documentation is generated and a local http server is started actual behavior docfx runs into a stack overflow and terminates however running it a second time seemingly fixes the issue and then it serves the documentation correctly
| 1
|
265,423
| 23,166,982,757
|
IssuesEvent
|
2022-07-30 05:02:43
|
istio/istio
|
https://api.github.com/repos/istio/istio
|
closed
|
Disable Istio default retry on errorcode 503
|
area/test and release lifecycle/stale
|
### Bug Description
I will try to explain the scenario which I am trying. I have 2 services appservice-A and appservice-B, both are in the same namespace "mynamespace" and both have seperate Virtual service called VS-A and VS-B.
In MyAPPLICATION, there is a call from appservice-A to appservice-B , and I want to disable the retry when the ServiceB throwing a 503 error, and the ServiceA should not retry again. So what I enabled is in the VS-B, I added the retry-attempt to 0, by expecting that appservice-A wont retry an attempt if the appservice-B throws a 503 error.. Which is not working for me.
I performed a testing again with the sample for you by making the attempt 0 in the VS-B. but still there is no change is happening to the retries, still showing 2.
```
apiVersion: networking.istio.io/v1beta1
kind: VirtualService
metadata:
labels:
app.kubernetes.io/managed-by: Helm
name: appservice-B
namespace: mynamespace
spec:
gateways:
- istio-system/ingress-gateway
hosts:
- appservice-B.mynamespace.svc.cluster.local
- appservice-B.mycompany.com
http:
- match:
- uri:
prefix: /
retries:
attempts: 0
route:
- destination:
host: appservice-B.mynamespace.svc.cluster.local
port:
number: 8080
subset: v1
```
this is the proxy config routing rule out put which generated from appservice-A
istioctl proxy-config route appservice-A-6dbb74bc88-dffb8 -n mynamespace -o json
```
"routes": [
{
"name": "default",
"match": {
"prefix": "/"
},
"route": {
"cluster": "outbound|8080||appservice-B.mynamespace.svc.cluster.local",
"timeout": "0s",
"retryPolicy": {
"retryOn": "connect-failure,refused-stream,unavailable,cancelled,retriable-status-codes",
"numRetries": 2,
"retryHostPredicate": [
{
"name": "envoy.retry_host_predicates.previous_hosts"
}
],
"hostSelectionRetryMaxAttempts": "5",
"retriableStatusCodes": [
503
]
},
"maxStreamDuration": {
"maxStreamDuration": "0s",
"grpcTimeoutHeaderMax": "0s"
}
},
"decorator": {
"operation": "appservice-B.mynamespace.svc.cluster.local:8080/*"
}
}
],
"includeRequestAttemptCount": true
```
So, suspect that , whether I am trying right things for my scenario, If this also right, any workaround like Is there achange the property of status codes or headers from responses to disable, so that istio wont retry appservice-B from appservice-A if a 503 errorcode got from appservice-B ?
### Version
```prose
istio version: 1.11.2
AKS version- 1.20.9, 1.22.6
```
### Additional Information
_No response_
|
1.0
|
Disable Istio default retry on errorcode 503 - ### Bug Description
I will try to explain the scenario which I am trying. I have 2 services appservice-A and appservice-B, both are in the same namespace "mynamespace" and both have seperate Virtual service called VS-A and VS-B.
In MyAPPLICATION, there is a call from appservice-A to appservice-B , and I want to disable the retry when the ServiceB throwing a 503 error, and the ServiceA should not retry again. So what I enabled is in the VS-B, I added the retry-attempt to 0, by expecting that appservice-A wont retry an attempt if the appservice-B throws a 503 error.. Which is not working for me.
I performed a testing again with the sample for you by making the attempt 0 in the VS-B. but still there is no change is happening to the retries, still showing 2.
```
apiVersion: networking.istio.io/v1beta1
kind: VirtualService
metadata:
labels:
app.kubernetes.io/managed-by: Helm
name: appservice-B
namespace: mynamespace
spec:
gateways:
- istio-system/ingress-gateway
hosts:
- appservice-B.mynamespace.svc.cluster.local
- appservice-B.mycompany.com
http:
- match:
- uri:
prefix: /
retries:
attempts: 0
route:
- destination:
host: appservice-B.mynamespace.svc.cluster.local
port:
number: 8080
subset: v1
```
this is the proxy config routing rule out put which generated from appservice-A
istioctl proxy-config route appservice-A-6dbb74bc88-dffb8 -n mynamespace -o json
```
"routes": [
{
"name": "default",
"match": {
"prefix": "/"
},
"route": {
"cluster": "outbound|8080||appservice-B.mynamespace.svc.cluster.local",
"timeout": "0s",
"retryPolicy": {
"retryOn": "connect-failure,refused-stream,unavailable,cancelled,retriable-status-codes",
"numRetries": 2,
"retryHostPredicate": [
{
"name": "envoy.retry_host_predicates.previous_hosts"
}
],
"hostSelectionRetryMaxAttempts": "5",
"retriableStatusCodes": [
503
]
},
"maxStreamDuration": {
"maxStreamDuration": "0s",
"grpcTimeoutHeaderMax": "0s"
}
},
"decorator": {
"operation": "appservice-B.mynamespace.svc.cluster.local:8080/*"
}
}
],
"includeRequestAttemptCount": true
```
So, suspect that , whether I am trying right things for my scenario, If this also right, any workaround like Is there achange the property of status codes or headers from responses to disable, so that istio wont retry appservice-B from appservice-A if a 503 errorcode got from appservice-B ?
### Version
```prose
istio version: 1.11.2
AKS version- 1.20.9, 1.22.6
```
### Additional Information
_No response_
|
non_build
|
disable istio default retry on errorcode bug description i will try to explain the scenario which i am trying i have services appservice a and appservice b both are in the same namespace mynamespace and both have seperate virtual service called vs a and vs b in myapplication there is a call from appservice a to appservice b and i want to disable the retry when the serviceb throwing a error and the servicea should not retry again so what i enabled is in the vs b i added the retry attempt to by expecting that appservice a wont retry an attempt if the appservice b throws a error which is not working for me i performed a testing again with the sample for you by making the attempt in the vs b but still there is no change is happening to the retries still showing apiversion networking istio io kind virtualservice metadata labels app kubernetes io managed by helm name appservice b namespace mynamespace spec gateways istio system ingress gateway hosts appservice b mynamespace svc cluster local appservice b mycompany com http match uri prefix retries attempts route destination host appservice b mynamespace svc cluster local port number subset this is the proxy config routing rule out put which generated from appservice a istioctl proxy config route appservice a n mynamespace o json routes name default match prefix route cluster outbound appservice b mynamespace svc cluster local timeout retrypolicy retryon connect failure refused stream unavailable cancelled retriable status codes numretries retryhostpredicate name envoy retry host predicates previous hosts hostselectionretrymaxattempts retriablestatuscodes maxstreamduration maxstreamduration grpctimeoutheadermax decorator operation appservice b mynamespace svc cluster local includerequestattemptcount true so suspect that whether i am trying right things for my scenario if this also right any workaround like is there achange the property of status codes or headers from responses to disable so that istio wont retry appservice b from appservice a if a errorcode got from appservice b version prose istio version aks version additional information no response
| 0
|
68,421
| 17,287,510,459
|
IssuesEvent
|
2021-07-24 02:28:48
|
opencv/opencv
|
https://api.github.com/repos/opencv/opencv
|
closed
|
Opencv 4.5.3 on Mac Air M1: unresolved symbols for arm64 (C++)
|
category: build/install invalid question (invalid tracker)
|
##### System information (version)
- OpenCV => 4.5.3
- Operating System / Platform => Mac Air M1 running OS Big Sur (arm64-apple-darwin20.4.0)
- Compiler => terminal / cmake 3.21.0 / g++ (clang version 12.0.5)
##### Detailed description
I compiled OpenCV from source which completes with no problems. I then try to compile a c++ test program using opencv and I get a lot of `Undefined symbols for architecture arm64:` and `ld: symbol(s) not found for architecture arm64`.
I have tried different possible solutions but could not find a way to fix this.
##### Steps to reproduce
1. Compile OpenCV 4.5.3:
```
cmake -DCMAKE_APPLE_SILICON_PROCESSOR=arm64 -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
or
```
cmake -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
or
```
cmake -DCMAKE_SYSTEM_PROCESSOR=arm64 -DCMAKE_OSX_ARCHITECTURES=arm64 -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
and then:
```
make -j8
cmake --build . --target install
```
2. Compile the following test program:
```.cpp
// C++ program to demonstrating ellipse
// over a self-formed background image
#include <iostream>
#include <opencv2/core/core.hpp>
// Drawing shapes
#include <opencv2/imgproc.hpp>
#include <opencv2/highgui/highgui.hpp>
using namespace cv;
using namespace std;
// Driver Code
int main(int argc, char** argv)
{
// Creating a blank image with
// white background
Mat image(500, 500, CV_8UC3,
Scalar(255, 255, 255));
// Check if the image is created
// successfully or not
if (!image.data) {
std::cout << "Could not open or "
<< "find the image\n";
return 0;
}
// Drawing the ellipse
ellipse(image, Point(256, 256),
Size(100, 50), 0, 0,
360, Scalar(0, 255, 255),
-1, LINE_AA);
// Showing image inside a window
imshow("Output", image);
waitKey(0);
return 0;
}
```
by using the following command:
```
g++ -std=c++11 -Wall opencv_test.cpp -I/Users/zzzzzz/Documents/yyyyy/xxx/opencv-4.5.3-build-static/install/include/opencv4 -L/Users/zzzzzz/Documents/yyyyy/xxx/opencv-4.5.3-build-static/install/lib -lopencv_core -lopencv_imgproc -lopencv_highgui -o opencv_test
```
I then get the following error:
<details>
```
Undefined symbols for architecture arm64:
"_CGWindowLevelForKey", referenced from:
cvSetPropTopmost_COCOA(char const*, bool) in libopencv_highgui.a(window_cocoa.mm.o)
"_NSAppKitVersionNumber", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
"_NSDefaultRunLoopMode", referenced from:
_cvWaitKey in libopencv_highgui.a(window_cocoa.mm.o)
"_NSDeviceRGBColorSpace", referenced from:
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
"_NSFullScreenModeSetting", referenced from:
cvSetModeWindow_COCOA(char const*, double) in libopencv_highgui.a(window_cocoa.mm.o)
"_NSZeroRect", referenced from:
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSApplication", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSAutoreleasePool", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSBitmapImageRep", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSDate", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSDictionary", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSEvent", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSImage", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSMutableDictionary", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSNumber", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSScreen", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSSlider", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSString", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSTextField", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSThread", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSView", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSWindow", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSObject", referenced from:
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSView", referenced from:
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSWindow", referenced from:
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
"carotene_o4t::bgrx2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgrx2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::bitwiseAnd(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::and8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bitwiseNot(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::not8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bitwiseXor(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::xor8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, float const*, long, float*, long, float)", referenced from:
cv::hal::recip32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, int const*, long, int*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::rgb2bgr565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgr565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2rgb565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2rgb565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, float, float)", referenced from:
cv::hal::addWeighted8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, float, float)", referenced from:
cv::hal::addWeighted8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, float, float, float)", referenced from:
cv::hal::addWeighted32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, float, float)", referenced from:
cv::hal::addWeighted16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, float, float)", referenced from:
cv::hal::addWeighted16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::rgbx2bgr565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgr565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2rgb565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2rgb565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::yuv420i2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::isSupportedConfiguration()", referenced from:
cv::hal::cvtBGRtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, int, bool) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoBGR5x5(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoGray(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtGraytoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoHSV(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool, bool, bool) in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
cv::hal::cvtBGRtoYUV(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool, bool) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
...
"carotene_o4t::add(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::add32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
cv::hal::div32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::max8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::max32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::max8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::max32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::max16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::max16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::min8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::min32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::min8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::min32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::min16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::min16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
cv::hal::mul32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, double, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::sub32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::phase(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
TegraRowOp_phase_Invoker<float const, float>::operator()(cv::Range const&) const in libopencv_core.a(mathfuncs_core.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::absdiff8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::absdiff32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::absdiff8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::absdiff32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::absdiff16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::absdiff16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bgr2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_bgr2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_bgr2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::rgb2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgr_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_rgb2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_rgb2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::bgr2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgr2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bgrx2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_bgrx2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_bgrx2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::gray2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_gray2rgb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgrx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2rgbx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgr_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_rgbx2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_rgbx2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::rgbx2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2rgb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bgr2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgr2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::bgrx2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgrx2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bitwiseOr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::or8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::gray2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_gray2rgbx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::magnitude(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
TegraRowOp_magnitude_Invoker<float const, float>::operator()(cv::Range const&) const in libopencv_core.a(mathfuncs_core.dispatch.cpp.o)
"carotene_o4t::rgb2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::rgbx2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgrx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"___CFConstantStringClassReference", referenced from:
CFString in libopencv_highgui.a(window_cocoa.mm.o)
"___itt_api_version_ptr__3_0", referenced from:
cv::utils::trace::details::isITTEnabled() in libopencv_core.a(trace.cpp.o)
"___itt_domain_create_ptr__3_0", referenced from:
cv::utils::trace::details::isITTEnabled() in libopencv_core.a(trace.cpp.o)
"___itt_id_create_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::Impl(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::Region*, cv::utils::trace::details::Region&, cv::utils::trace::details::Region::LocationStaticStorage const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::registerRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::Impl(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::Region*, cv::utils::trace::details::Region&, cv::utils::trace::details::Region::LocationStaticStorage const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Region(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
"___itt_id_destroy_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::~Impl() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::~Impl() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::release() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::destroy() in libopencv_core.a(trace.cpp.o)
"___itt_metadata_add_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, int) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, double) in libopencv_core.a(trace.cpp.o)
"___itt_metadata_str_add_ptr__3_0", referenced from:
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, char const*) in libopencv_core.a(trace.cpp.o)
"___itt_region_begin_ptr__3_0", referenced from:
cv::utils::trace::details::TraceManager::TraceManager() in libopencv_core.a(trace.cpp.o)
"___itt_region_end_ptr__3_0", referenced from:
cv::utils::trace::details::TraceManager::~TraceManager() in libopencv_core.a(trace.cpp.o)
"___itt_relation_add_ptr__3_0", referenced from:
cv::utils::trace::details::parallelForAttachNestedRegion(cv::utils::trace::details::Region const&) in libopencv_core.a(trace.cpp.o)
"___itt_string_handle_create_ptr__3_0", referenced from:
cv::utils::trace::details::Region::LocationExtraData::LocationExtraData(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::LocationExtraData::LocationExtraData(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::LocationExtraData::init(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::TraceManager::TraceManager() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::initTraceArg(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::TraceArg const&) in libopencv_core.a(trace.cpp.o)
"___itt_task_begin_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::enterRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
"___itt_task_end_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
"___itt_thread_set_name_ptr__3_0", referenced from:
cv::TLSData<cv::(anonymous namespace)::ThreadID>::createDataInstance() const in libopencv_core.a(system.cpp.o)
"__objc_empty_cache", referenced from:
_OBJC_CLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_clBuildProgram", referenced from:
_clBuildProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clBuildProgram_pfn)
"_clCompileProgram", referenced from:
_clCompileProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCompileProgram_pfn)
"_clCreateBuffer", referenced from:
_clCreateBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateBuffer_pfn)
"_clCreateCommandQueue", referenced from:
_clCreateCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateCommandQueue_pfn)
"_clCreateContext", referenced from:
_clCreateContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateContextFromType_pfn, _clCreateContext_pfn )
"_clCreateContextFromType", referenced from:
_clCreateContextFromType_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateContextFromType_pfn)
"_clCreateImage", referenced from:
_clCreateImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage3D_pfn, _clCreateImage_pfn , _clCreateImage2D_pfn )
"_clCreateImage2D", referenced from:
_clCreateImage2D_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage2D_pfn)
"_clCreateImage3D", referenced from:
_clCreateImage3D_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage3D_pfn)
"_clCreateKernel", referenced from:
_clCreateKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateKernelsInProgram_pfn, _clCreateKernel_pfn )
"_clCreateKernelsInProgram", referenced from:
_clCreateKernelsInProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateKernelsInProgram_pfn)
"_clCreateProgramWithBinary", referenced from:
_clCreateProgramWithBinary_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithBinary_pfn)
"_clCreateProgramWithBuiltInKernels", referenced from:
_clCreateProgramWithBuiltInKernels_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithBuiltInKernels_pfn)
"_clCreateProgramWithSource", referenced from:
_clCreateProgramWithSource_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithSource_pfn)
"_clCreateSampler", referenced from:
_clCreateSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSampler_pfn)
"_clCreateSubBuffer", referenced from:
_clCreateSubBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSubBuffer_pfn)
"_clCreateSubDevices", referenced from:
_clCreateSubDevices_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSubDevices_pfn)
"_clCreateUserEvent", referenced from:
_clCreateUserEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateUserEvent_pfn)
"_clEnqueueBarrier", referenced from:
_clEnqueueBarrier_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueBarrierWithWaitList_pfn, _clEnqueueBarrier_pfn )
"_clEnqueueBarrierWithWaitList", referenced from:
_clEnqueueBarrierWithWaitList_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueBarrierWithWaitList_pfn)
"_clEnqueueCopyBuffer", referenced from:
_clEnqueueCopyBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBuffer_pfn, _clEnqueueCopyBufferRect_pfn , _clEnqueueCopyBufferToImage_pfn )
"_clEnqueueCopyBufferRect", referenced from:
_clEnqueueCopyBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBufferRect_pfn)
"_clEnqueueCopyBufferToImage", referenced from:
_clEnqueueCopyBufferToImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBufferToImage_pfn)
"_clEnqueueCopyImage", referenced from:
_clEnqueueCopyImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyImage_pfn, _clEnqueueCopyImageToBuffer_pfn )
"_clEnqueueCopyImageToBuffer", referenced from:
_clEnqueueCopyImageToBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyImageToBuffer_pfn)
"_clEnqueueFillBuffer", referenced from:
_clEnqueueFillBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueFillBuffer_pfn)
"_clEnqueueFillImage", referenced from:
_clEnqueueFillImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueFillImage_pfn)
"_clEnqueueMapBuffer", referenced from:
_clEnqueueMapBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMapBuffer_pfn)
"_clEnqueueMapImage", referenced from:
_clEnqueueMapImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMapImage_pfn)
"_clEnqueueMarker", referenced from:
_clEnqueueMarker_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMarkerWithWaitList_pfn, _clEnqueueMarker_pfn )
"_clEnqueueMarkerWithWaitList", referenced from:
_clEnqueueMarkerWithWaitList_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMarkerWithWaitList_pfn)
"_clEnqueueMigrateMemObjects", referenced from:
_clEnqueueMigrateMemObjects_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMigrateMemObjects_pfn)
"_clEnqueueNDRangeKernel", referenced from:
_clEnqueueNDRangeKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueNDRangeKernel_pfn)
"_clEnqueueNativeKernel", referenced from:
_clEnqueueNativeKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueNativeKernel_pfn)
"_clEnqueueReadBuffer", referenced from:
_clEnqueueReadBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadBuffer_pfn, _clEnqueueReadBufferRect_pfn )
"_clEnqueueReadBufferRect", referenced from:
_clEnqueueReadBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadBufferRect_pfn)
"_clEnqueueReadImage", referenced from:
_clEnqueueReadImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadImage_pfn)
"_clEnqueueTask", referenced from:
_clEnqueueTask_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueTask_pfn)
"_clEnqueueUnmapMemObject", referenced from:
_clEnqueueUnmapMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueUnmapMemObject_pfn)
"_clEnqueueWaitForEvents", referenced from:
_clEnqueueWaitForEvents_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWaitForEvents_pfn)
"_clEnqueueWriteBuffer", referenced from:
_clEnqueueWriteBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteBuffer_pfn, _clEnqueueWriteBufferRect_pfn )
"_clEnqueueWriteBufferRect", referenced from:
_clEnqueueWriteBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteBufferRect_pfn)
"_clEnqueueWriteImage", referenced from:
_clEnqueueWriteImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteImage_pfn)
"_clFinish", referenced from:
_clFinish_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clFinish_pfn)
"_clFlush", referenced from:
_clFlush_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clFlush_pfn)
"_clGetCommandQueueInfo", referenced from:
_clGetCommandQueueInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetCommandQueueInfo_pfn)
"_clGetContextInfo", referenced from:
_clGetContextInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetContextInfo_pfn)
"_clGetDeviceIDs", referenced from:
_clGetDeviceIDs_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetDeviceIDs_pfn)
"_clGetDeviceInfo", referenced from:
_clGetDeviceInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetDeviceInfo_pfn)
"_clGetEventInfo", referenced from:
_clGetEventInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetEventInfo_pfn)
"_clGetEventProfilingInfo", referenced from:
_clGetEventProfilingInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetEventProfilingInfo_pfn)
"_clGetExtensionFunctionAddress", referenced from:
_clGetExtensionFunctionAddress_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetExtensionFunctionAddress_pfn, _clGetExtensionFunctionAddressForPlatform_pfn )
"_clGetExtensionFunctionAddressForPlatform", referenced from:
_clGetExtensionFunctionAddressForPlatform_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetExtensionFunctionAddressForPlatform_pfn)
"_clGetImageInfo", referenced from:
_clGetImageInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetImageInfo_pfn)
"_clGetKernelArgInfo", referenced from:
_clGetKernelArgInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelArgInfo_pfn)
"_clGetKernelInfo", referenced from:
_clGetKernelInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelInfo_pfn)
"_clGetKernelWorkGroupInfo", referenced from:
_clGetKernelWorkGroupInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelWorkGroupInfo_pfn)
"_clGetMemObjectInfo", referenced from:
_clGetMemObjectInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetMemObjectInfo_pfn)
"_clGetPlatformIDs", referenced from:
_clGetPlatformIDs_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetPlatformIDs_pfn)
"_clGetPlatformInfo", referenced from:
_clGetPlatformInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetPlatformInfo_pfn)
"_clGetProgramBuildInfo", referenced from:
_clGetProgramBuildInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetProgramBuildInfo_pfn)
"_clGetProgramInfo", referenced from:
_clGetProgramInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetProgramInfo_pfn)
"_clGetSamplerInfo", referenced from:
_clGetSamplerInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetSamplerInfo_pfn)
"_clGetSupportedImageFormats", referenced from:
_clGetSupportedImageFormats_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetSupportedImageFormats_pfn)
"_clLinkProgram", referenced from:
_clLinkProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clLinkProgram_pfn)
"_clReleaseCommandQueue", referenced from:
_clReleaseCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseCommandQueue_pfn)
"_clReleaseContext", referenced from:
_clReleaseContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseContext_pfn)
"_clReleaseDevice", referenced from:
_clReleaseDevice_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseDevice_pfn)
"_clReleaseEvent", referenced from:
_clReleaseEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseEvent_pfn)
"_clReleaseKernel", referenced from:
_clReleaseKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseKernel_pfn)
"_clReleaseMemObject", referenced from:
_clReleaseMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseMemObject_pfn)
"_clReleaseProgram", referenced from:
_clReleaseProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseProgram_pfn)
"_clReleaseSampler", referenced from:
_clReleaseSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseSampler_pfn)
"_clRetainCommandQueue", referenced from:
_clRetainCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainCommandQueue_pfn)
"_clRetainContext", referenced from:
_clRetainContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainContext_pfn)
"_clRetainDevice", referenced from:
_clRetainDevice_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainDevice_pfn)
"_clRetainEvent", referenced from:
_clRetainEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainEvent_pfn)
"_clRetainKernel", referenced from:
_clRetainKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainKernel_pfn)
"_clRetainMemObject", referenced from:
_clRetainMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainMemObject_pfn)
"_clRetainProgram", referenced from:
_clRetainProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainProgram_pfn)
"_clRetainSampler", referenced from:
_clRetainSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainSampler_pfn)
"_clSetEventCallback", referenced from:
_clSetEventCallback_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetEventCallback_pfn)
"_clSetKernelArg", referenced from:
_clSetKernelArg_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetKernelArg_pfn)
"_clSetMemObjectDestructorCallback", referenced from:
_clSetMemObjectDestructorCallback_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetMemObjectDestructorCallback_pfn)
"_clSetUserEventStatus", referenced from:
_clSetUserEventStatus_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetUserEventStatus_pfn)
"_clUnloadCompiler", referenced from:
_clUnloadCompiler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clUnloadCompiler_pfn)
"_clUnloadPlatformCompiler", referenced from:
_clUnloadPlatformCompiler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clUnloadPlatformCompiler_pfn)
"_clWaitForEvents", referenced from:
_clWaitForEvents_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clWaitForEvents_pfn)
"_objc_alloc", referenced from:
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider init] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_alloc_init", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvResizeWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvMoveWindow in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_enumerationMutation", referenced from:
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvGetWindowName in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_getProperty", referenced from:
-[CVWindow sliders] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView image] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView imageView] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider slider] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider name] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_msgSend", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvResizeWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvMoveWindow in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_msgSendSuper2", referenced from:
-[CVWindow keyDown:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVWindow contentView] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView init] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider init] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_opt_respondsToSelector", referenced from:
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvCreateTrackbar2 in libopencv_highgui.a(window_cocoa.mm.o)
_cvGetTrackbarPos in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarPos in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarMax in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarMin in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_opt_self", referenced from:
cvSetPropTopmost_COCOA(char const*, bool) in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_release", referenced from:
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_setProperty_atomic", referenced from:
-[CVWindow setSliders:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImage:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageView:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider setSlider:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider setName:] in libopencv_highgui.a(window_cocoa.mm.o)
ld: symbol(s) not found for architecture arm64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
```
</details>
##### Issue submission checklist
- [x] I report the issue, it's not a question
<!--
OpenCV team works with forum.opencv.org, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [x] I checked the problem with documentation, FAQ, open issues,
forum.opencv.org, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenCV documentation: https://docs.opencv.org
* FAQ page: https://github.com/opencv/opencv/wiki/FAQ
* OpenCV forum: https://forum.opencv.org
* OpenCV issue tracker: https://github.com/opencv/opencv/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/opencv
-->
- [x] I updated to latest OpenCV version and the issue is still there
<!--
master branch for OpenCV 4.x and 3.4 branch for OpenCV 3.x releases.
OpenCV team supports only latest release for each branch.
The ticket is closed, if the problem is not reproduced with modern version.
-->
- [x] There is reproducer code and related data files: videos, images, onnx, etc
<!--
The best reproducer -- test case for OpenCV that we can add to the library.
Recommendations for media files and binary files:
* Try to reproduce the issue with images and videos in opencv_extra repository
to reduce attachment size
* Use PNG for images, if you report some CV related bug, but not image reader
issue
* Attach the image as archive to the ticket, if you report some reader issue.
Image hosting services compress images and it breaks the repro code.
* Provide ONNX file for some public model or ONNX file with with random weights,
if you report ONNX parsing or handling issue. Architecture details diagram
from netron tool can be very useful too. See https://lutzroeder.github.io/netron/
-->
|
1.0
|
Opencv 4.5.3 on Mac Air M1: unresolved symbols for arm64 (C++) - ##### System information (version)
- OpenCV => 4.5.3
- Operating System / Platform => Mac Air M1 running OS Big Sur (arm64-apple-darwin20.4.0)
- Compiler => terminal / cmake 3.21.0 / g++ (clang version 12.0.5)
##### Detailed description
I compiled OpenCV from source which completes with no problems. I then try to compile a c++ test program using opencv and I get a lot of `Undefined symbols for architecture arm64:` and `ld: symbol(s) not found for architecture arm64`.
I have tried different possible solutions but could not find a way to fix this.
##### Steps to reproduce
1. Compile OpenCV 4.5.3:
```
cmake -DCMAKE_APPLE_SILICON_PROCESSOR=arm64 -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
or
```
cmake -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
or
```
cmake -DCMAKE_SYSTEM_PROCESSOR=arm64 -DCMAKE_OSX_ARCHITECTURES=arm64 -DCMAKE_INSTALL_PREFIX=install -DCMAKE_BUILD_TYPE=RELEASE -DBUILD_opencv_python3=ON -DBUILD_opencv_python2=OFF -DBUILD_SHARED_LIBS=OFF ../opencv-4.5.3
```
and then:
```
make -j8
cmake --build . --target install
```
2. Compile the following test program:
```.cpp
// C++ program to demonstrating ellipse
// over a self-formed background image
#include <iostream>
#include <opencv2/core/core.hpp>
// Drawing shapes
#include <opencv2/imgproc.hpp>
#include <opencv2/highgui/highgui.hpp>
using namespace cv;
using namespace std;
// Driver Code
int main(int argc, char** argv)
{
// Creating a blank image with
// white background
Mat image(500, 500, CV_8UC3,
Scalar(255, 255, 255));
// Check if the image is created
// successfully or not
if (!image.data) {
std::cout << "Could not open or "
<< "find the image\n";
return 0;
}
// Drawing the ellipse
ellipse(image, Point(256, 256),
Size(100, 50), 0, 0,
360, Scalar(0, 255, 255),
-1, LINE_AA);
// Showing image inside a window
imshow("Output", image);
waitKey(0);
return 0;
}
```
by using the following command:
```
g++ -std=c++11 -Wall opencv_test.cpp -I/Users/zzzzzz/Documents/yyyyy/xxx/opencv-4.5.3-build-static/install/include/opencv4 -L/Users/zzzzzz/Documents/yyyyy/xxx/opencv-4.5.3-build-static/install/lib -lopencv_core -lopencv_imgproc -lopencv_highgui -o opencv_test
```
I then get the following error:
<details>
```
Undefined symbols for architecture arm64:
"_CGWindowLevelForKey", referenced from:
cvSetPropTopmost_COCOA(char const*, bool) in libopencv_highgui.a(window_cocoa.mm.o)
"_NSAppKitVersionNumber", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
"_NSDefaultRunLoopMode", referenced from:
_cvWaitKey in libopencv_highgui.a(window_cocoa.mm.o)
"_NSDeviceRGBColorSpace", referenced from:
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
"_NSFullScreenModeSetting", referenced from:
cvSetModeWindow_COCOA(char const*, double) in libopencv_highgui.a(window_cocoa.mm.o)
"_NSZeroRect", referenced from:
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSApplication", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSAutoreleasePool", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSBitmapImageRep", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSDate", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSDictionary", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSEvent", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSImage", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSMutableDictionary", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSNumber", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSScreen", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSSlider", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSString", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSTextField", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSThread", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSView", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_CLASS_$_NSWindow", referenced from:
objc-class-ref in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSObject", referenced from:
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSView", referenced from:
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_OBJC_METACLASS_$_NSWindow", referenced from:
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
"carotene_o4t::bgrx2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgrx2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::bitwiseAnd(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::and8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bitwiseNot(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::not8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bitwiseXor(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::xor8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, float const*, long, float*, long, float)", referenced from:
cv::hal::recip32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, int const*, long, int*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::reciprocal(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::recip16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::rgb2bgr565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgr565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2rgb565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2rgb565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, float, float)", referenced from:
cv::hal::addWeighted8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, float, float)", referenced from:
cv::hal::addWeighted8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, float, float, float)", referenced from:
cv::hal::addWeighted32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, float, float)", referenced from:
cv::hal::addWeighted16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::addWeighted(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, float, float)", referenced from:
cv::hal::addWeighted16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::rgbx2bgr565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgr565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2rgb565(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2rgb565_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::yuv420i2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420i2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::yuv420sp2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::isSupportedConfiguration()", referenced from:
cv::hal::cvtBGRtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, int, bool) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoBGR5x5(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoGray(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtGraytoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int) in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
cv::hal::cvtBGRtoHSV(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool, bool, bool) in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
cv::hal::cvtBGRtoYUV(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, int, bool, bool) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
cv::hal::cvtTwoPlaneYUVtoBGR(unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, int, bool, int) in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
...
"carotene_o4t::add(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::add32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::add(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::add16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
cv::hal::div32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::div(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::div16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::max8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::max32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::max8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::max32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::max16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::max(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::max16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::min8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::min32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::min8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::min32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::min16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::min(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::min16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
cv::hal::mul32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, double, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::mul(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, float, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::mul16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::sub32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::sub(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long, carotene_o4t::CONVERT_POLICY)", referenced from:
cv::hal::sub16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpEQ(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGE(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpGT(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8s(signed char const*, unsigned long, signed char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, float const*, long, float const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, int const*, long, int const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp32s(int const*, unsigned long, int const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, short const*, long, short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::cmpNE(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned char*, long)", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::phase(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long, float)", referenced from:
TegraRowOp_phase_Invoker<float const, float>::operator()(cv::Range const&) const in libopencv_core.a(mathfuncs_core.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, signed char const*, long, signed char const*, long, signed char*, long)", referenced from:
cv::hal::absdiff8s(signed char const*, unsigned long, signed char const*, unsigned long, signed char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
cv::hal::absdiff32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::absdiff8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, int const*, long, int const*, long, int*, long)", referenced from:
cv::hal::absdiff32s(int const*, unsigned long, int const*, unsigned long, int*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, short const*, long, short const*, long, short*, long)", referenced from:
cv::hal::absdiff16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::absDiff(carotene_o4t::Size2D const&, unsigned short const*, long, unsigned short const*, long, unsigned short*, long)", referenced from:
cv::hal::absdiff16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::bgr2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_bgr2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_bgr2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::rgb2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgr_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_rgb2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_rgb2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::bgr2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgr2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bgrx2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_bgrx2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_bgrx2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::gray2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_gray2rgb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2bgrx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgb2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2rgbx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2bgr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgr_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2hsv(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long, int)", referenced from:
TegraCvtColor_rgbx2hsvf_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
TegraCvtColor_rgbx2hsv_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_hsv.dispatch.cpp.o)
"carotene_o4t::rgbx2rgb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2rgb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bgr2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgr2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::bgrx2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_bgrx2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::bitwiseOr(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char const*, long, unsigned char*, long)", referenced from:
cv::hal::or8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in libopencv_core.a(arithm.dispatch.cpp.o)
"carotene_o4t::gray2rgbx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_gray2rgbx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::magnitude(carotene_o4t::Size2D const&, float const*, long, float const*, long, float*, long)", referenced from:
TegraRowOp_magnitude_Invoker<float const, float>::operator()(cv::Range const&) const in libopencv_core.a(mathfuncs_core.dispatch.cpp.o)
"carotene_o4t::rgb2ycrcb(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgb2ycrcb_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_yuv.dispatch.cpp.o)
"carotene_o4t::rgbx2bgrx(carotene_o4t::Size2D const&, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2bgrx_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"carotene_o4t::rgbx2gray(carotene_o4t::Size2D const&, carotene_o4t::COLOR_SPACE, unsigned char const*, long, unsigned char*, long)", referenced from:
TegraCvtColor_rgbx2gray_Invoker::operator()(cv::Range const&) const in libopencv_imgproc.a(color_rgb.dispatch.cpp.o)
"___CFConstantStringClassReference", referenced from:
CFString in libopencv_highgui.a(window_cocoa.mm.o)
"___itt_api_version_ptr__3_0", referenced from:
cv::utils::trace::details::isITTEnabled() in libopencv_core.a(trace.cpp.o)
"___itt_domain_create_ptr__3_0", referenced from:
cv::utils::trace::details::isITTEnabled() in libopencv_core.a(trace.cpp.o)
"___itt_id_create_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::Impl(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::Region*, cv::utils::trace::details::Region&, cv::utils::trace::details::Region::LocationStaticStorage const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::registerRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::Impl(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::Region*, cv::utils::trace::details::Region&, cv::utils::trace::details::Region::LocationStaticStorage const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Region(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
"___itt_id_destroy_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::~Impl() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::~Impl() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::release() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::destroy() in libopencv_core.a(trace.cpp.o)
"___itt_metadata_add_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, int) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, long long) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, double) in libopencv_core.a(trace.cpp.o)
"___itt_metadata_str_add_ptr__3_0", referenced from:
cv::utils::trace::details::traceArg(cv::utils::trace::details::TraceArg const&, char const*) in libopencv_core.a(trace.cpp.o)
"___itt_region_begin_ptr__3_0", referenced from:
cv::utils::trace::details::TraceManager::TraceManager() in libopencv_core.a(trace.cpp.o)
"___itt_region_end_ptr__3_0", referenced from:
cv::utils::trace::details::TraceManager::~TraceManager() in libopencv_core.a(trace.cpp.o)
"___itt_relation_add_ptr__3_0", referenced from:
cv::utils::trace::details::parallelForAttachNestedRegion(cv::utils::trace::details::Region const&) in libopencv_core.a(trace.cpp.o)
"___itt_string_handle_create_ptr__3_0", referenced from:
cv::utils::trace::details::Region::LocationExtraData::LocationExtraData(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::LocationExtraData::LocationExtraData(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::LocationExtraData::init(cv::utils::trace::details::Region::LocationStaticStorage const&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::TraceManager::TraceManager() in libopencv_core.a(trace.cpp.o)
cv::utils::trace::details::initTraceArg(cv::utils::trace::details::TraceManagerThreadLocal&, cv::utils::trace::details::TraceArg const&) in libopencv_core.a(trace.cpp.o)
"___itt_task_begin_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::enterRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
"___itt_task_end_ptr__3_0", referenced from:
cv::utils::trace::details::Region::Impl::leaveRegion(cv::utils::trace::details::TraceManagerThreadLocal&) in libopencv_core.a(trace.cpp.o)
"___itt_thread_set_name_ptr__3_0", referenced from:
cv::TLSData<cv::(anonymous namespace)::ThreadID>::createDataInstance() const in libopencv_core.a(system.cpp.o)
"__objc_empty_cache", referenced from:
_OBJC_CLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_CLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVWindow in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVView in libopencv_highgui.a(window_cocoa.mm.o)
_OBJC_METACLASS_$_CVSlider in libopencv_highgui.a(window_cocoa.mm.o)
"_clBuildProgram", referenced from:
_clBuildProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clBuildProgram_pfn)
"_clCompileProgram", referenced from:
_clCompileProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCompileProgram_pfn)
"_clCreateBuffer", referenced from:
_clCreateBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateBuffer_pfn)
"_clCreateCommandQueue", referenced from:
_clCreateCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateCommandQueue_pfn)
"_clCreateContext", referenced from:
_clCreateContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateContextFromType_pfn, _clCreateContext_pfn )
"_clCreateContextFromType", referenced from:
_clCreateContextFromType_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateContextFromType_pfn)
"_clCreateImage", referenced from:
_clCreateImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage3D_pfn, _clCreateImage_pfn , _clCreateImage2D_pfn )
"_clCreateImage2D", referenced from:
_clCreateImage2D_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage2D_pfn)
"_clCreateImage3D", referenced from:
_clCreateImage3D_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateImage3D_pfn)
"_clCreateKernel", referenced from:
_clCreateKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateKernelsInProgram_pfn, _clCreateKernel_pfn )
"_clCreateKernelsInProgram", referenced from:
_clCreateKernelsInProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateKernelsInProgram_pfn)
"_clCreateProgramWithBinary", referenced from:
_clCreateProgramWithBinary_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithBinary_pfn)
"_clCreateProgramWithBuiltInKernels", referenced from:
_clCreateProgramWithBuiltInKernels_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithBuiltInKernels_pfn)
"_clCreateProgramWithSource", referenced from:
_clCreateProgramWithSource_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateProgramWithSource_pfn)
"_clCreateSampler", referenced from:
_clCreateSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSampler_pfn)
"_clCreateSubBuffer", referenced from:
_clCreateSubBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSubBuffer_pfn)
"_clCreateSubDevices", referenced from:
_clCreateSubDevices_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateSubDevices_pfn)
"_clCreateUserEvent", referenced from:
_clCreateUserEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clCreateUserEvent_pfn)
"_clEnqueueBarrier", referenced from:
_clEnqueueBarrier_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueBarrierWithWaitList_pfn, _clEnqueueBarrier_pfn )
"_clEnqueueBarrierWithWaitList", referenced from:
_clEnqueueBarrierWithWaitList_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueBarrierWithWaitList_pfn)
"_clEnqueueCopyBuffer", referenced from:
_clEnqueueCopyBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBuffer_pfn, _clEnqueueCopyBufferRect_pfn , _clEnqueueCopyBufferToImage_pfn )
"_clEnqueueCopyBufferRect", referenced from:
_clEnqueueCopyBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBufferRect_pfn)
"_clEnqueueCopyBufferToImage", referenced from:
_clEnqueueCopyBufferToImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyBufferToImage_pfn)
"_clEnqueueCopyImage", referenced from:
_clEnqueueCopyImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyImage_pfn, _clEnqueueCopyImageToBuffer_pfn )
"_clEnqueueCopyImageToBuffer", referenced from:
_clEnqueueCopyImageToBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueCopyImageToBuffer_pfn)
"_clEnqueueFillBuffer", referenced from:
_clEnqueueFillBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueFillBuffer_pfn)
"_clEnqueueFillImage", referenced from:
_clEnqueueFillImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueFillImage_pfn)
"_clEnqueueMapBuffer", referenced from:
_clEnqueueMapBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMapBuffer_pfn)
"_clEnqueueMapImage", referenced from:
_clEnqueueMapImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMapImage_pfn)
"_clEnqueueMarker", referenced from:
_clEnqueueMarker_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMarkerWithWaitList_pfn, _clEnqueueMarker_pfn )
"_clEnqueueMarkerWithWaitList", referenced from:
_clEnqueueMarkerWithWaitList_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMarkerWithWaitList_pfn)
"_clEnqueueMigrateMemObjects", referenced from:
_clEnqueueMigrateMemObjects_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueMigrateMemObjects_pfn)
"_clEnqueueNDRangeKernel", referenced from:
_clEnqueueNDRangeKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueNDRangeKernel_pfn)
"_clEnqueueNativeKernel", referenced from:
_clEnqueueNativeKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueNativeKernel_pfn)
"_clEnqueueReadBuffer", referenced from:
_clEnqueueReadBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadBuffer_pfn, _clEnqueueReadBufferRect_pfn )
"_clEnqueueReadBufferRect", referenced from:
_clEnqueueReadBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadBufferRect_pfn)
"_clEnqueueReadImage", referenced from:
_clEnqueueReadImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueReadImage_pfn)
"_clEnqueueTask", referenced from:
_clEnqueueTask_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueTask_pfn)
"_clEnqueueUnmapMemObject", referenced from:
_clEnqueueUnmapMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueUnmapMemObject_pfn)
"_clEnqueueWaitForEvents", referenced from:
_clEnqueueWaitForEvents_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWaitForEvents_pfn)
"_clEnqueueWriteBuffer", referenced from:
_clEnqueueWriteBuffer_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteBuffer_pfn, _clEnqueueWriteBufferRect_pfn )
"_clEnqueueWriteBufferRect", referenced from:
_clEnqueueWriteBufferRect_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteBufferRect_pfn)
"_clEnqueueWriteImage", referenced from:
_clEnqueueWriteImage_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clEnqueueWriteImage_pfn)
"_clFinish", referenced from:
_clFinish_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clFinish_pfn)
"_clFlush", referenced from:
_clFlush_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clFlush_pfn)
"_clGetCommandQueueInfo", referenced from:
_clGetCommandQueueInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetCommandQueueInfo_pfn)
"_clGetContextInfo", referenced from:
_clGetContextInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetContextInfo_pfn)
"_clGetDeviceIDs", referenced from:
_clGetDeviceIDs_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetDeviceIDs_pfn)
"_clGetDeviceInfo", referenced from:
_clGetDeviceInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetDeviceInfo_pfn)
"_clGetEventInfo", referenced from:
_clGetEventInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetEventInfo_pfn)
"_clGetEventProfilingInfo", referenced from:
_clGetEventProfilingInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetEventProfilingInfo_pfn)
"_clGetExtensionFunctionAddress", referenced from:
_clGetExtensionFunctionAddress_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetExtensionFunctionAddress_pfn, _clGetExtensionFunctionAddressForPlatform_pfn )
"_clGetExtensionFunctionAddressForPlatform", referenced from:
_clGetExtensionFunctionAddressForPlatform_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetExtensionFunctionAddressForPlatform_pfn)
"_clGetImageInfo", referenced from:
_clGetImageInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetImageInfo_pfn)
"_clGetKernelArgInfo", referenced from:
_clGetKernelArgInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelArgInfo_pfn)
"_clGetKernelInfo", referenced from:
_clGetKernelInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelInfo_pfn)
"_clGetKernelWorkGroupInfo", referenced from:
_clGetKernelWorkGroupInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetKernelWorkGroupInfo_pfn)
"_clGetMemObjectInfo", referenced from:
_clGetMemObjectInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetMemObjectInfo_pfn)
"_clGetPlatformIDs", referenced from:
_clGetPlatformIDs_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetPlatformIDs_pfn)
"_clGetPlatformInfo", referenced from:
_clGetPlatformInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetPlatformInfo_pfn)
"_clGetProgramBuildInfo", referenced from:
_clGetProgramBuildInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetProgramBuildInfo_pfn)
"_clGetProgramInfo", referenced from:
_clGetProgramInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetProgramInfo_pfn)
"_clGetSamplerInfo", referenced from:
_clGetSamplerInfo_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetSamplerInfo_pfn)
"_clGetSupportedImageFormats", referenced from:
_clGetSupportedImageFormats_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clGetSupportedImageFormats_pfn)
"_clLinkProgram", referenced from:
_clLinkProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clLinkProgram_pfn)
"_clReleaseCommandQueue", referenced from:
_clReleaseCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseCommandQueue_pfn)
"_clReleaseContext", referenced from:
_clReleaseContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseContext_pfn)
"_clReleaseDevice", referenced from:
_clReleaseDevice_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseDevice_pfn)
"_clReleaseEvent", referenced from:
_clReleaseEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseEvent_pfn)
"_clReleaseKernel", referenced from:
_clReleaseKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseKernel_pfn)
"_clReleaseMemObject", referenced from:
_clReleaseMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseMemObject_pfn)
"_clReleaseProgram", referenced from:
_clReleaseProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseProgram_pfn)
"_clReleaseSampler", referenced from:
_clReleaseSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clReleaseSampler_pfn)
"_clRetainCommandQueue", referenced from:
_clRetainCommandQueue_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainCommandQueue_pfn)
"_clRetainContext", referenced from:
_clRetainContext_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainContext_pfn)
"_clRetainDevice", referenced from:
_clRetainDevice_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainDevice_pfn)
"_clRetainEvent", referenced from:
_clRetainEvent_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainEvent_pfn)
"_clRetainKernel", referenced from:
_clRetainKernel_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainKernel_pfn)
"_clRetainMemObject", referenced from:
_clRetainMemObject_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainMemObject_pfn)
"_clRetainProgram", referenced from:
_clRetainProgram_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainProgram_pfn)
"_clRetainSampler", referenced from:
_clRetainSampler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clRetainSampler_pfn)
"_clSetEventCallback", referenced from:
_clSetEventCallback_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetEventCallback_pfn)
"_clSetKernelArg", referenced from:
_clSetKernelArg_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetKernelArg_pfn)
"_clSetMemObjectDestructorCallback", referenced from:
_clSetMemObjectDestructorCallback_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetMemObjectDestructorCallback_pfn)
"_clSetUserEventStatus", referenced from:
_clSetUserEventStatus_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clSetUserEventStatus_pfn)
"_clUnloadCompiler", referenced from:
_clUnloadCompiler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clUnloadCompiler_pfn)
"_clUnloadPlatformCompiler", referenced from:
_clUnloadPlatformCompiler_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clUnloadPlatformCompiler_pfn)
"_clWaitForEvents", referenced from:
_clWaitForEvents_pfn in libopencv_core.a(opencl_core.cpp.o)
(maybe you meant: _clWaitForEvents_pfn)
"_objc_alloc", referenced from:
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider init] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_alloc_init", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvResizeWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvMoveWindow in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_enumerationMutation", referenced from:
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvGetWindowName in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_getProperty", referenced from:
-[CVWindow sliders] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView image] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView imageView] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider slider] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider name] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_msgSend", referenced from:
_cvInitSystem in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvDestroyAllWindows in libopencv_highgui.a(window_cocoa.mm.o)
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvNamedWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvResizeWindow in libopencv_highgui.a(window_cocoa.mm.o)
_cvMoveWindow in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_msgSendSuper2", referenced from:
-[CVWindow keyDown:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVWindow contentView] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView init] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider init] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_opt_respondsToSelector", referenced from:
_cvShowImage in libopencv_highgui.a(window_cocoa.mm.o)
_cvCreateTrackbar2 in libopencv_highgui.a(window_cocoa.mm.o)
_cvGetTrackbarPos in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarPos in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarMax in libopencv_highgui.a(window_cocoa.mm.o)
_cvSetTrackbarMin in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setFrameSize:] in libopencv_highgui.a(window_cocoa.mm.o)
...
"_objc_opt_self", referenced from:
cvSetPropTopmost_COCOA(char const*, bool) in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_release", referenced from:
-[CVView setImageData:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView drawRect:] in libopencv_highgui.a(window_cocoa.mm.o)
"_objc_setProperty_atomic", referenced from:
-[CVWindow setSliders:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImage:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVView setImageView:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider setSlider:] in libopencv_highgui.a(window_cocoa.mm.o)
-[CVSlider setName:] in libopencv_highgui.a(window_cocoa.mm.o)
ld: symbol(s) not found for architecture arm64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
```
</details>
##### Issue submission checklist
- [x] I report the issue, it's not a question
<!--
OpenCV team works with forum.opencv.org, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [x] I checked the problem with documentation, FAQ, open issues,
forum.opencv.org, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenCV documentation: https://docs.opencv.org
* FAQ page: https://github.com/opencv/opencv/wiki/FAQ
* OpenCV forum: https://forum.opencv.org
* OpenCV issue tracker: https://github.com/opencv/opencv/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/opencv
-->
- [x] I updated to latest OpenCV version and the issue is still there
<!--
master branch for OpenCV 4.x and 3.4 branch for OpenCV 3.x releases.
OpenCV team supports only latest release for each branch.
The ticket is closed, if the problem is not reproduced with modern version.
-->
- [x] There is reproducer code and related data files: videos, images, onnx, etc
<!--
The best reproducer -- test case for OpenCV that we can add to the library.
Recommendations for media files and binary files:
* Try to reproduce the issue with images and videos in opencv_extra repository
to reduce attachment size
* Use PNG for images, if you report some CV related bug, but not image reader
issue
* Attach the image as archive to the ticket, if you report some reader issue.
Image hosting services compress images and it breaks the repro code.
* Provide ONNX file for some public model or ONNX file with with random weights,
if you report ONNX parsing or handling issue. Architecture details diagram
from netron tool can be very useful too. See https://lutzroeder.github.io/netron/
-->
|
build
|
opencv on mac air unresolved symbols for c system information version opencv operating system platform mac air running os big sur apple compiler terminal cmake g clang version detailed description i compiled opencv from source which completes with no problems i then try to compile a c test program using opencv and i get a lot of undefined symbols for architecture and ld symbol s not found for architecture i have tried different possible solutions but could not find a way to fix this steps to reproduce compile opencv cmake dcmake apple silicon processor dcmake install prefix install dcmake build type release dbuild opencv on dbuild opencv off dbuild shared libs off opencv or cmake dcmake install prefix install dcmake build type release dbuild opencv on dbuild opencv off dbuild shared libs off opencv or cmake dcmake system processor dcmake osx architectures dcmake install prefix install dcmake build type release dbuild opencv on dbuild opencv off dbuild shared libs off opencv and then make cmake build target install compile the following test program cpp c program to demonstrating ellipse over a self formed background image include include drawing shapes include include using namespace cv using namespace std driver code int main int argc char argv creating a blank image with white background mat image cv scalar check if the image is created successfully or not if image data std cout could not open or find the image n return drawing the ellipse ellipse image point size scalar line aa showing image inside a window imshow output image waitkey return by using the following command g std c wall opencv test cpp i users zzzzzz documents yyyyy xxx opencv build static install include l users zzzzzz documents yyyyy xxx opencv build static install lib lopencv core lopencv imgproc lopencv highgui o opencv test i then get the following error undefined symbols for architecture cgwindowlevelforkey referenced from cvsetproptopmost cocoa char const bool in libopencv highgui a window cocoa mm o nsappkitversionnumber referenced from cvinitsystem in libopencv highgui a window cocoa mm o cvnamedwindow in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o nsdefaultrunloopmode referenced from cvwaitkey in libopencv highgui a window cocoa mm o nsdevicergbcolorspace referenced from in libopencv highgui a window cocoa mm o nsfullscreenmodesetting referenced from cvsetmodewindow cocoa char const double in libopencv highgui a window cocoa mm o nszerorect referenced from in libopencv highgui a window cocoa mm o objc class nsapplication referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsautoreleasepool referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsbitmapimagerep referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsdate referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsdictionary referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsevent referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsimage referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsmutabledictionary referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsnumber referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsscreen referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsslider referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsstring referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nstextfield referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsthread referenced from objc class ref in libopencv highgui a window cocoa mm o objc class nsview referenced from objc class ref in libopencv highgui a window cocoa mm o objc class cvview in libopencv highgui a window cocoa mm o objc class cvslider in libopencv highgui a window cocoa mm o objc class nswindow referenced from objc class ref in libopencv highgui a window cocoa mm o objc class cvwindow in libopencv highgui a window cocoa mm o objc metaclass nsobject referenced from objc metaclass cvwindow in libopencv highgui a window cocoa mm o objc metaclass cvview in libopencv highgui a window cocoa mm o objc metaclass cvslider in libopencv highgui a window cocoa mm o objc metaclass nsview referenced from objc metaclass cvview in libopencv highgui a window cocoa mm o objc metaclass cvslider in libopencv highgui a window cocoa mm o objc metaclass nswindow referenced from objc metaclass cvwindow in libopencv highgui a window cocoa mm o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color yuv dispatch cpp o carotene bitwiseand carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene bitwisenot carotene const unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene bitwisexor carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const signed char const long signed char long float carotene convert policy referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const float const long float long float referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const unsigned char const long unsigned char long float carotene convert policy referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const int const long int long float carotene convert policy referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const short const long short long float carotene convert policy referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene reciprocal carotene const unsigned short const long unsigned short long float carotene convert policy referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color yuv dispatch cpp o carotene addweighted carotene const signed char const long signed char const long signed char long float float float referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene addweighted carotene const unsigned char const long unsigned char const long unsigned char long float float float referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene addweighted carotene const int const long int const long int long float float float referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene addweighted carotene const short const long short const long short long float float float referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene addweighted carotene const unsigned short const long unsigned short const long unsigned short long float float float referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene issupportedconfiguration referenced from cv hal cvtbgrtobgr unsigned char const unsigned long unsigned char unsigned long int int int int int bool in libopencv imgproc a color rgb dispatch cpp o cv hal unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color rgb dispatch cpp o cv hal cvtbgrtogray unsigned char const unsigned long unsigned char unsigned long int int int int bool in libopencv imgproc a color rgb dispatch cpp o cv hal cvtgraytobgr unsigned char const unsigned long unsigned char unsigned long int int int int in libopencv imgproc a color rgb dispatch cpp o cv hal cvtbgrtohsv unsigned char const unsigned long unsigned char unsigned long int int int int bool bool bool in libopencv imgproc a color hsv dispatch cpp o cv hal cvtbgrtoyuv unsigned char const unsigned long unsigned char unsigned long int int int int bool bool in libopencv imgproc a color yuv dispatch cpp o cv hal cvttwoplaneyuvtobgr unsigned char const unsigned long unsigned char unsigned long int int int bool int in libopencv imgproc a color yuv dispatch cpp o carotene add carotene const signed char const long signed char const long signed char long carotene convert policy referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene add carotene const float const long float const long float long referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene add carotene const unsigned char const long unsigned char const long unsigned char long carotene convert policy referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene add carotene const int const long int const long int long carotene convert policy referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene add carotene const short const long short const long short long carotene convert policy referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene add carotene const unsigned short const long unsigned short const long unsigned short long carotene convert policy referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const signed char const long signed char const long signed char long float carotene convert policy referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const float const long float const long float long float referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const unsigned char const long unsigned char const long unsigned char long float carotene convert policy referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const int const long int const long int long float carotene convert policy referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const short const long short const long short long float carotene convert policy referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene div carotene const unsigned short const long unsigned short const long unsigned short long float carotene convert policy referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const signed char const long signed char const long signed char long referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const float const long float const long float long referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const int const long int const long int long referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const short const long short const long short long referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene max carotene const unsigned short const long unsigned short const long unsigned short long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const signed char const long signed char const long signed char long referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const float const long float const long float long referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const int const long int const long int long referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const short const long short const long short long referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene min carotene const unsigned short const long unsigned short const long unsigned short long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const signed char const long signed char const long signed char long float carotene convert policy referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const float const long float const long float long float referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const unsigned char const long unsigned char const long unsigned char long float carotene convert policy referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const int const long int const long int long double carotene convert policy referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const short const long short const long short long float carotene convert policy referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene mul carotene const unsigned short const long unsigned short const long unsigned short long float carotene convert policy referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const signed char const long signed char const long signed char long carotene convert policy referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const float const long float const long float long referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const unsigned char const long unsigned char const long unsigned char long carotene convert policy referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const int const long int const long int long carotene convert policy referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const short const long short const long short long carotene convert policy referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene sub carotene const unsigned short const long unsigned short const long unsigned short long carotene convert policy referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const signed char const long signed char const long unsigned char long referenced from cv hal signed char const unsigned long signed char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const float const long float const long unsigned char long referenced from cv hal float const unsigned long float const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const int const long int const long unsigned char long referenced from cv hal int const unsigned long int const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const short const long short const long unsigned char long referenced from cv hal short const unsigned long short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpeq carotene const unsigned short const long unsigned short const long unsigned char long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const signed char const long signed char const long unsigned char long referenced from cv hal signed char const unsigned long signed char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const float const long float const long unsigned char long referenced from cv hal float const unsigned long float const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const int const long int const long unsigned char long referenced from cv hal int const unsigned long int const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const short const long short const long unsigned char long referenced from cv hal short const unsigned long short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpge carotene const unsigned short const long unsigned short const long unsigned char long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const signed char const long signed char const long unsigned char long referenced from cv hal signed char const unsigned long signed char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const float const long float const long unsigned char long referenced from cv hal float const unsigned long float const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const int const long int const long unsigned char long referenced from cv hal int const unsigned long int const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const short const long short const long unsigned char long referenced from cv hal short const unsigned long short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpgt carotene const unsigned short const long unsigned short const long unsigned char long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const signed char const long signed char const long unsigned char long referenced from cv hal signed char const unsigned long signed char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const float const long float const long unsigned char long referenced from cv hal float const unsigned long float const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const int const long int const long unsigned char long referenced from cv hal int const unsigned long int const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const short const long short const long unsigned char long referenced from cv hal short const unsigned long short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene cmpne carotene const unsigned short const long unsigned short const long unsigned char long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene phase carotene const float const long float const long float long float referenced from tegrarowop phase invoker operator cv range const const in libopencv core a mathfuncs core dispatch cpp o carotene absdiff carotene const signed char const long signed char const long signed char long referenced from cv hal signed char const unsigned long signed char const unsigned long signed char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene absdiff carotene const float const long float const long float long referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in libopencv core a arithm dispatch cpp o carotene absdiff carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene absdiff carotene const int const long int const long int long referenced from cv hal int const unsigned long int const unsigned long int unsigned long int int void in libopencv core a arithm dispatch cpp o carotene absdiff carotene const short const long short const long short long referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene absdiff carotene const unsigned short const long unsigned short const long unsigned short long referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in libopencv core a arithm dispatch cpp o carotene carotene const unsigned char const long unsigned char long int referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long int referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o carotene carotene const carotene color space unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long int referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const carotene color space unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long int referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o tegracvtcolor invoker operator cv range const const in libopencv imgproc a color hsv dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color yuv dispatch cpp o carotene carotene const carotene color space unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene bitwiseor carotene const unsigned char const long unsigned char const long unsigned char long referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in libopencv core a arithm dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene magnitude carotene const float const long float const long float long referenced from tegrarowop magnitude invoker operator cv range const const in libopencv core a mathfuncs core dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color yuv dispatch cpp o carotene carotene const unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o carotene carotene const carotene color space unsigned char const long unsigned char long referenced from tegracvtcolor invoker operator cv range const const in libopencv imgproc a color rgb dispatch cpp o cfconstantstringclassreference referenced from cfstring in libopencv highgui a window cocoa mm o itt api version ptr referenced from cv utils trace details isittenabled in libopencv core a trace cpp o itt domain create ptr referenced from cv utils trace details isittenabled in libopencv core a trace cpp o itt id create ptr referenced from cv utils trace details region impl impl cv utils trace details tracemanagerthreadlocal cv utils trace details region cv utils trace details region cv utils trace details region locationstaticstorage const long long in libopencv core a trace cpp o cv utils trace details region impl registerregion cv utils trace details tracemanagerthreadlocal in libopencv core a trace cpp o cv utils trace details region impl impl cv utils trace details tracemanagerthreadlocal cv utils trace details region cv utils trace details region cv utils trace details region locationstaticstorage const long long in libopencv core a trace cpp o cv utils trace details region region cv utils trace details region locationstaticstorage const in libopencv core a trace cpp o itt id destroy ptr referenced from cv utils trace details region impl impl in libopencv core a trace cpp o cv utils trace details region impl impl in libopencv core a trace cpp o cv utils trace details region impl release in libopencv core a trace cpp o cv utils trace details region destroy in libopencv core a trace cpp o itt metadata add ptr referenced from cv utils trace details region impl leaveregion cv utils trace details tracemanagerthreadlocal in libopencv core a trace cpp o cv utils trace details tracearg cv utils trace details tracearg const int in libopencv core a trace cpp o cv utils trace details tracearg cv utils trace details tracearg const long long in libopencv core a trace cpp o cv utils trace details tracearg cv utils trace details tracearg const double in libopencv core a trace cpp o itt metadata str add ptr referenced from cv utils trace details tracearg cv utils trace details tracearg const char const in libopencv core a trace cpp o itt region begin ptr referenced from cv utils trace details tracemanager tracemanager in libopencv core a trace cpp o itt region end ptr referenced from cv utils trace details tracemanager tracemanager in libopencv core a trace cpp o itt relation add ptr referenced from cv utils trace details parallelforattachnestedregion cv utils trace details region const in libopencv core a trace cpp o itt string handle create ptr referenced from cv utils trace details region locationextradata locationextradata cv utils trace details region locationstaticstorage const in libopencv core a trace cpp o cv utils trace details region locationextradata locationextradata cv utils trace details region locationstaticstorage const in libopencv core a trace cpp o cv utils trace details region locationextradata init cv utils trace details region locationstaticstorage const in libopencv core a trace cpp o cv utils trace details region impl leaveregion cv utils trace details tracemanagerthreadlocal in libopencv core a trace cpp o cv utils trace details tracemanager tracemanager in libopencv core a trace cpp o cv utils trace details inittracearg cv utils trace details tracemanagerthreadlocal cv utils trace details tracearg const in libopencv core a trace cpp o itt task begin ptr referenced from cv utils trace details region impl enterregion cv utils trace details tracemanagerthreadlocal in libopencv core a trace cpp o itt task end ptr referenced from cv utils trace details region impl leaveregion cv utils trace details tracemanagerthreadlocal in libopencv core a trace cpp o itt thread set name ptr referenced from cv tlsdata createdatainstance const in libopencv core a system cpp o objc empty cache referenced from objc class cvwindow in libopencv highgui a window cocoa mm o objc class cvview in libopencv highgui a window cocoa mm o objc class cvslider in libopencv highgui a window cocoa mm o objc metaclass cvwindow in libopencv highgui a window cocoa mm o objc metaclass cvview in libopencv highgui a window cocoa mm o objc metaclass cvslider in libopencv highgui a window cocoa mm o clbuildprogram referenced from clbuildprogram pfn in libopencv core a opencl core cpp o maybe you meant clbuildprogram pfn clcompileprogram referenced from clcompileprogram pfn in libopencv core a opencl core cpp o maybe you meant clcompileprogram pfn clcreatebuffer referenced from clcreatebuffer pfn in libopencv core a opencl core cpp o maybe you meant clcreatebuffer pfn clcreatecommandqueue referenced from clcreatecommandqueue pfn in libopencv core a opencl core cpp o maybe you meant clcreatecommandqueue pfn clcreatecontext referenced from clcreatecontext pfn in libopencv core a opencl core cpp o maybe you meant clcreatecontextfromtype pfn clcreatecontext pfn clcreatecontextfromtype referenced from clcreatecontextfromtype pfn in libopencv core a opencl core cpp o maybe you meant clcreatecontextfromtype pfn clcreateimage referenced from clcreateimage pfn in libopencv core a opencl core cpp o maybe you meant pfn clcreateimage pfn pfn referenced from pfn in libopencv core a opencl core cpp o maybe you meant pfn referenced from pfn in libopencv core a opencl core cpp o maybe you meant pfn clcreatekernel referenced from clcreatekernel pfn in libopencv core a opencl core cpp o maybe you meant clcreatekernelsinprogram pfn clcreatekernel pfn clcreatekernelsinprogram referenced from clcreatekernelsinprogram pfn in libopencv core a opencl core cpp o maybe you meant clcreatekernelsinprogram pfn clcreateprogramwithbinary referenced from clcreateprogramwithbinary pfn in libopencv core a opencl core cpp o maybe you meant clcreateprogramwithbinary pfn clcreateprogramwithbuiltinkernels referenced from clcreateprogramwithbuiltinkernels pfn in libopencv core a opencl core cpp o maybe you meant clcreateprogramwithbuiltinkernels pfn clcreateprogramwithsource referenced from clcreateprogramwithsource pfn in libopencv core a opencl core cpp o maybe you meant clcreateprogramwithsource pfn clcreatesampler referenced from clcreatesampler pfn in libopencv core a opencl core cpp o maybe you meant clcreatesampler pfn clcreatesubbuffer referenced from clcreatesubbuffer pfn in libopencv core a opencl core cpp o maybe you meant clcreatesubbuffer pfn clcreatesubdevices referenced from clcreatesubdevices pfn in libopencv core a opencl core cpp o maybe you meant clcreatesubdevices pfn clcreateuserevent referenced from clcreateuserevent pfn in libopencv core a opencl core cpp o maybe you meant clcreateuserevent pfn clenqueuebarrier referenced from clenqueuebarrier pfn in libopencv core a opencl core cpp o maybe you meant clenqueuebarrierwithwaitlist pfn clenqueuebarrier pfn clenqueuebarrierwithwaitlist referenced from clenqueuebarrierwithwaitlist pfn in libopencv core a opencl core cpp o maybe you meant clenqueuebarrierwithwaitlist pfn clenqueuecopybuffer referenced from clenqueuecopybuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuecopybuffer pfn clenqueuecopybufferrect pfn clenqueuecopybuffertoimage pfn clenqueuecopybufferrect referenced from clenqueuecopybufferrect pfn in libopencv core a opencl core cpp o maybe you meant clenqueuecopybufferrect pfn clenqueuecopybuffertoimage referenced from clenqueuecopybuffertoimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuecopybuffertoimage pfn clenqueuecopyimage referenced from clenqueuecopyimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuecopyimage pfn clenqueuecopyimagetobuffer pfn clenqueuecopyimagetobuffer referenced from clenqueuecopyimagetobuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuecopyimagetobuffer pfn clenqueuefillbuffer referenced from clenqueuefillbuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuefillbuffer pfn clenqueuefillimage referenced from clenqueuefillimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuefillimage pfn clenqueuemapbuffer referenced from clenqueuemapbuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuemapbuffer pfn clenqueuemapimage referenced from clenqueuemapimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuemapimage pfn clenqueuemarker referenced from clenqueuemarker pfn in libopencv core a opencl core cpp o maybe you meant clenqueuemarkerwithwaitlist pfn clenqueuemarker pfn clenqueuemarkerwithwaitlist referenced from clenqueuemarkerwithwaitlist pfn in libopencv core a opencl core cpp o maybe you meant clenqueuemarkerwithwaitlist pfn clenqueuemigratememobjects referenced from clenqueuemigratememobjects pfn in libopencv core a opencl core cpp o maybe you meant clenqueuemigratememobjects pfn clenqueuendrangekernel referenced from clenqueuendrangekernel pfn in libopencv core a opencl core cpp o maybe you meant clenqueuendrangekernel pfn clenqueuenativekernel referenced from clenqueuenativekernel pfn in libopencv core a opencl core cpp o maybe you meant clenqueuenativekernel pfn clenqueuereadbuffer referenced from clenqueuereadbuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuereadbuffer pfn clenqueuereadbufferrect pfn clenqueuereadbufferrect referenced from clenqueuereadbufferrect pfn in libopencv core a opencl core cpp o maybe you meant clenqueuereadbufferrect pfn clenqueuereadimage referenced from clenqueuereadimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuereadimage pfn clenqueuetask referenced from clenqueuetask pfn in libopencv core a opencl core cpp o maybe you meant clenqueuetask pfn clenqueueunmapmemobject referenced from clenqueueunmapmemobject pfn in libopencv core a opencl core cpp o maybe you meant clenqueueunmapmemobject pfn clenqueuewaitforevents referenced from clenqueuewaitforevents pfn in libopencv core a opencl core cpp o maybe you meant clenqueuewaitforevents pfn clenqueuewritebuffer referenced from clenqueuewritebuffer pfn in libopencv core a opencl core cpp o maybe you meant clenqueuewritebuffer pfn clenqueuewritebufferrect pfn clenqueuewritebufferrect referenced from clenqueuewritebufferrect pfn in libopencv core a opencl core cpp o maybe you meant clenqueuewritebufferrect pfn clenqueuewriteimage referenced from clenqueuewriteimage pfn in libopencv core a opencl core cpp o maybe you meant clenqueuewriteimage pfn clfinish referenced from clfinish pfn in libopencv core a opencl core cpp o maybe you meant clfinish pfn clflush referenced from clflush pfn in libopencv core a opencl core cpp o maybe you meant clflush pfn clgetcommandqueueinfo referenced from clgetcommandqueueinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetcommandqueueinfo pfn clgetcontextinfo referenced from clgetcontextinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetcontextinfo pfn clgetdeviceids referenced from clgetdeviceids pfn in libopencv core a opencl core cpp o maybe you meant clgetdeviceids pfn clgetdeviceinfo referenced from clgetdeviceinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetdeviceinfo pfn clgeteventinfo referenced from clgeteventinfo pfn in libopencv core a opencl core cpp o maybe you meant clgeteventinfo pfn clgeteventprofilinginfo referenced from clgeteventprofilinginfo pfn in libopencv core a opencl core cpp o maybe you meant clgeteventprofilinginfo pfn clgetextensionfunctionaddress referenced from clgetextensionfunctionaddress pfn in libopencv core a opencl core cpp o maybe you meant clgetextensionfunctionaddress pfn clgetextensionfunctionaddressforplatform pfn clgetextensionfunctionaddressforplatform referenced from clgetextensionfunctionaddressforplatform pfn in libopencv core a opencl core cpp o maybe you meant clgetextensionfunctionaddressforplatform pfn clgetimageinfo referenced from clgetimageinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetimageinfo pfn clgetkernelarginfo referenced from clgetkernelarginfo pfn in libopencv core a opencl core cpp o maybe you meant clgetkernelarginfo pfn clgetkernelinfo referenced from clgetkernelinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetkernelinfo pfn clgetkernelworkgroupinfo referenced from clgetkernelworkgroupinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetkernelworkgroupinfo pfn clgetmemobjectinfo referenced from clgetmemobjectinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetmemobjectinfo pfn clgetplatformids referenced from clgetplatformids pfn in libopencv core a opencl core cpp o maybe you meant clgetplatformids pfn clgetplatforminfo referenced from clgetplatforminfo pfn in libopencv core a opencl core cpp o maybe you meant clgetplatforminfo pfn clgetprogrambuildinfo referenced from clgetprogrambuildinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetprogrambuildinfo pfn clgetprograminfo referenced from clgetprograminfo pfn in libopencv core a opencl core cpp o maybe you meant clgetprograminfo pfn clgetsamplerinfo referenced from clgetsamplerinfo pfn in libopencv core a opencl core cpp o maybe you meant clgetsamplerinfo pfn clgetsupportedimageformats referenced from clgetsupportedimageformats pfn in libopencv core a opencl core cpp o maybe you meant clgetsupportedimageformats pfn cllinkprogram referenced from cllinkprogram pfn in libopencv core a opencl core cpp o maybe you meant cllinkprogram pfn clreleasecommandqueue referenced from clreleasecommandqueue pfn in libopencv core a opencl core cpp o maybe you meant clreleasecommandqueue pfn clreleasecontext referenced from clreleasecontext pfn in libopencv core a opencl core cpp o maybe you meant clreleasecontext pfn clreleasedevice referenced from clreleasedevice pfn in libopencv core a opencl core cpp o maybe you meant clreleasedevice pfn clreleaseevent referenced from clreleaseevent pfn in libopencv core a opencl core cpp o maybe you meant clreleaseevent pfn clreleasekernel referenced from clreleasekernel pfn in libopencv core a opencl core cpp o maybe you meant clreleasekernel pfn clreleasememobject referenced from clreleasememobject pfn in libopencv core a opencl core cpp o maybe you meant clreleasememobject pfn clreleaseprogram referenced from clreleaseprogram pfn in libopencv core a opencl core cpp o maybe you meant clreleaseprogram pfn clreleasesampler referenced from clreleasesampler pfn in libopencv core a opencl core cpp o maybe you meant clreleasesampler pfn clretaincommandqueue referenced from clretaincommandqueue pfn in libopencv core a opencl core cpp o maybe you meant clretaincommandqueue pfn clretaincontext referenced from clretaincontext pfn in libopencv core a opencl core cpp o maybe you meant clretaincontext pfn clretaindevice referenced from clretaindevice pfn in libopencv core a opencl core cpp o maybe you meant clretaindevice pfn clretainevent referenced from clretainevent pfn in libopencv core a opencl core cpp o maybe you meant clretainevent pfn clretainkernel referenced from clretainkernel pfn in libopencv core a opencl core cpp o maybe you meant clretainkernel pfn clretainmemobject referenced from clretainmemobject pfn in libopencv core a opencl core cpp o maybe you meant clretainmemobject pfn clretainprogram referenced from clretainprogram pfn in libopencv core a opencl core cpp o maybe you meant clretainprogram pfn clretainsampler referenced from clretainsampler pfn in libopencv core a opencl core cpp o maybe you meant clretainsampler pfn clseteventcallback referenced from clseteventcallback pfn in libopencv core a opencl core cpp o maybe you meant clseteventcallback pfn clsetkernelarg referenced from clsetkernelarg pfn in libopencv core a opencl core cpp o maybe you meant clsetkernelarg pfn clsetmemobjectdestructorcallback referenced from clsetmemobjectdestructorcallback pfn in libopencv core a opencl core cpp o maybe you meant clsetmemobjectdestructorcallback pfn clsetusereventstatus referenced from clsetusereventstatus pfn in libopencv core a opencl core cpp o maybe you meant clsetusereventstatus pfn clunloadcompiler referenced from clunloadcompiler pfn in libopencv core a opencl core cpp o maybe you meant clunloadcompiler pfn clunloadplatformcompiler referenced from clunloadplatformcompiler pfn in libopencv core a opencl core cpp o maybe you meant clunloadplatformcompiler pfn clwaitforevents referenced from clwaitforevents pfn in libopencv core a opencl core cpp o maybe you meant clwaitforevents pfn objc alloc referenced from cvnamedwindow in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc alloc init referenced from cvinitsystem in libopencv highgui a window cocoa mm o cvdestroywindow in libopencv highgui a window cocoa mm o cvdestroyallwindows in libopencv highgui a window cocoa mm o cvshowimage in libopencv highgui a window cocoa mm o cvnamedwindow in libopencv highgui a window cocoa mm o cvresizewindow in libopencv highgui a window cocoa mm o cvmovewindow in libopencv highgui a window cocoa mm o objc enumerationmutation referenced from cvdestroyallwindows in libopencv highgui a window cocoa mm o cvgetwindowname in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc getproperty referenced from in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc msgsend referenced from cvinitsystem in libopencv highgui a window cocoa mm o cvdestroywindow in libopencv highgui a window cocoa mm o cvdestroyallwindows in libopencv highgui a window cocoa mm o cvshowimage in libopencv highgui a window cocoa mm o cvnamedwindow in libopencv highgui a window cocoa mm o cvresizewindow in libopencv highgui a window cocoa mm o cvmovewindow in libopencv highgui a window cocoa mm o objc referenced from in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc opt respondstoselector referenced from cvshowimage in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o cvgettrackbarpos in libopencv highgui a window cocoa mm o cvsettrackbarpos in libopencv highgui a window cocoa mm o cvsettrackbarmax in libopencv highgui a window cocoa mm o cvsettrackbarmin in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc opt self referenced from cvsetproptopmost cocoa char const bool in libopencv highgui a window cocoa mm o objc release referenced from in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o objc setproperty atomic referenced from in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o in libopencv highgui a window cocoa mm o ld symbol s not found for architecture clang error linker command failed with exit code use v to see invocation issue submission checklist i report the issue it s not a question opencv team works with forum opencv org stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues forum opencv org stack overflow etc and have not found solution places to check opencv documentation faq page opencv forum opencv issue tracker stack overflow branch i updated to latest opencv version and the issue is still there master branch for opencv x and branch for opencv x releases opencv team supports only latest release for each branch the ticket is closed if the problem is not reproduced with modern version there is reproducer code and related data files videos images onnx etc the best reproducer test case for opencv that we can add to the library recommendations for media files and binary files try to reproduce the issue with images and videos in opencv extra repository to reduce attachment size use png for images if you report some cv related bug but not image reader issue attach the image as archive to the ticket if you report some reader issue image hosting services compress images and it breaks the repro code provide onnx file for some public model or onnx file with with random weights if you report onnx parsing or handling issue architecture details diagram from netron tool can be very useful too see
| 1
|
78,134
| 22,151,414,234
|
IssuesEvent
|
2022-06-03 17:14:27
|
apache/beam
|
https://api.github.com/repos/apache/beam
|
opened
|
Jenkins cross JDK version test on Windows
|
P3 build-system beam-fixit testing task
|
We can set os variant to choose windows for Jenkins test, which can be combined with JDK version test. So that we can have cross OS / cross JDK version test.
This discussion came from https://github.com/apache/beam/pull/3184#pullrequestreview-39303400
Imported from Jira [BEAM-2339](https://issues.apache.org/jira/browse/BEAM-2339). Original Jira may contain additional context.
Reported by: markflyhigh.
|
1.0
|
Jenkins cross JDK version test on Windows - We can set os variant to choose windows for Jenkins test, which can be combined with JDK version test. So that we can have cross OS / cross JDK version test.
This discussion came from https://github.com/apache/beam/pull/3184#pullrequestreview-39303400
Imported from Jira [BEAM-2339](https://issues.apache.org/jira/browse/BEAM-2339). Original Jira may contain additional context.
Reported by: markflyhigh.
|
build
|
jenkins cross jdk version test on windows we can set os variant to choose windows for jenkins test which can be combined with jdk version test so that we can have cross os cross jdk version test this discussion came from imported from jira original jira may contain additional context reported by markflyhigh
| 1
|
29,648
| 8,387,453,360
|
IssuesEvent
|
2018-10-09 00:29:41
|
DestinyItemManager/DIM
|
https://api.github.com/repos/DestinyItemManager/DIM
|
closed
|
First perk selected in Loadout Builder
|
Feature: Loadout Builder
|
https://twitter.com/efishkin/status/1048699426082967552?s=21
@ThisIsDIM I'm having trouble with the loadout builder (beta,I appreciate it :)), when I select two perks under gauntlet, the builder only looks for the first, is this as intended?
|
1.0
|
First perk selected in Loadout Builder - https://twitter.com/efishkin/status/1048699426082967552?s=21
@ThisIsDIM I'm having trouble with the loadout builder (beta,I appreciate it :)), when I select two perks under gauntlet, the builder only looks for the first, is this as intended?
|
build
|
first perk selected in loadout builder thisisdim i m having trouble with the loadout builder beta i appreciate it when i select two perks under gauntlet the builder only looks for the first is this as intended
| 1
|
21,462
| 6,153,824,382
|
IssuesEvent
|
2017-06-28 10:59:45
|
wangdachangchang/Note
|
https://api.github.com/repos/wangdachangchang/Note
|
opened
|
go-redis-base
|
Code
|
```go
package main
import (
"fmt"
"sync"
"time"
redis "gopkg.in/redis.v4"
)
func main() {
client := createClient()
//stringOperation(client)
//listOperation(client)
//setOperation(client)
//hashOperation(client)
//connectPool(client)
}
// 创建客户端
func createClient() *redis.Client {
client := redis.NewClient(&redis.Options{
Addr: "localhost:6379",
Password: "",
DB: 0, // 数据库名
PoolSize: 5, // 最大连接池数量
})
// 通过client.Ping()检查是否成功连接redis服务器
pong, err := client.Ping().Result()
fmt.Println(pong, err)
return client
}
// string操作
func stringOperation(client *redis.Client) {
// 第三个参数是过期时间 0表示没有过期时间
err := client.Set("name", "wnag", 0).Err()
if err != nil {
panic(err)
}
val, err := client.Get("name").Result()
if err != nil {
panic(err)
}
fmt.Println("name", val)
// 设置过期时间
err = client.Set("age", "20", 1*time.Second).Err()
if err != nil {
panic(err)
}
client.Incr("age")
client.Incr("age")
client.Decr("age")
val, err = client.Get("age").Result()
if err != nil {
panic(err)
}
fmt.Println("age", val)
// 设置的过期时间为1秒 1秒过后key自动被删除
time.Sleep(1 * time.Second)
val, err = client.Get("age").Result()
if err != nil {
// key "age"已经被删除返回nil
fmt.Println("err:%v", err)
}
fmt.Println("age", val)
}
// list操作
func listOperation(client *redis.Client) {
client.RPush("fruit", "apple") // 在名称为fruit的list尾添加一个值为value的元素
client.LPush("fruit", "banana") // 在名称为fruit的list头添加一个值为value的元素
length, err := client.LLen("fruit").Result() // 返回名称为fruit的list的长度
if err != nil {
panic(err)
}
fmt.Println("length:", length)
// 返回并删除名称为fruit的list中的首元素
value, err := client.LPop("fruit").Result()
if err != nil {
panic(err)
}
fmt.Println("fruit", value)
// 返回并删除名称为fruit的list中的尾元素
value, err = client.RPop("fruit").Result()
if err != nil {
panic(err)
}
fmt.Println("fruit", value)
}
// set操作
func setOperation(client *redis.Client) {
client.SAdd("blacklist", "Obama") // 向blacklist中添加元素
client.SAdd("blacklist", "Hillary") // 可以继续添加
client.SAdd("blacklist", "the Elder")
client.SAdd("whitelist", "the Elder")
// 判断元素是否在集合中
isMember, err := client.SIsMember("blacklist", "Bush").Result()
if err != nil {
panic(err)
}
fmt.Println("Is Bush in blacklist:", isMember)
// 求交集,既在black中又在white中
names, err := client.SInter("blacklist", "whitelist").Result()
if err != nil {
panic(err)
}
fmt.Println("Inter result:", names)
// 获取指定集合的所有元素
all, err := client.SMembers("blacklist").Result()
if err != nil {
panic(err)
}
fmt.Println("All member:", all)
}
// hash操作
func hashOperation(client *redis.Client) {
client.HSet("user_wang", "name", "wang") // 向名称为user_wang的hash中添加元素name
client.HSet("user_wang", "age", "18") // 向名称为user_wang的hash中添加元素age
// 批量向user_test的hash中添加元素name和age
client.HMSet("user_test", map[string]string{"name": "test", "age": "20"})
// 批量获取user_test的hash中指定字段的值
fields, err := client.HMGet("user_test", "name", "age").Result()
if err != nil {
panic(err)
}
fmt.Println("fields in user_test:", fields)
// 获取user_wang的hash中的字段个数
length, err := client.HLen("user_wang").Result()
if err != nil {
panic(err)
}
fmt.Println("fields count in user_wang:", length)
// 删除user_test的age字段
client.HDel("user_test", "age")
age, err := client.HGet("user_test", "age").Result()
if err != nil {
fmt.Println("Get user_test age error: %v", err)
} else {
fmt.Println("user_test age is:", age)
}
}
// redis.v4的连接池管理
func connectPool(client *redis.Client) {
wg := sync.WaitGroup{}
wg.Add(10)
for i := 0; i < 10; i++ {
go func() {
defer wg.Done()
for j := 0; j < 100; j++ {
client.Set(fmt.Sprintf("name%d", j), fmt.Sprintf("wang%d", j), 0).Err()
client.Get(fmt.Sprintf("name%d", j)).Result()
}
fmt.Printf("PoolStats, TotalConns: %d, FressConns: %d\n", client.PoolStats().TotalConns, client.PoolStats().FreeConns)
}()
}
wg.Wait()
}
```
|
1.0
|
go-redis-base - ```go
package main
import (
"fmt"
"sync"
"time"
redis "gopkg.in/redis.v4"
)
func main() {
client := createClient()
//stringOperation(client)
//listOperation(client)
//setOperation(client)
//hashOperation(client)
//connectPool(client)
}
// 创建客户端
func createClient() *redis.Client {
client := redis.NewClient(&redis.Options{
Addr: "localhost:6379",
Password: "",
DB: 0, // 数据库名
PoolSize: 5, // 最大连接池数量
})
// 通过client.Ping()检查是否成功连接redis服务器
pong, err := client.Ping().Result()
fmt.Println(pong, err)
return client
}
// string操作
func stringOperation(client *redis.Client) {
// 第三个参数是过期时间 0表示没有过期时间
err := client.Set("name", "wnag", 0).Err()
if err != nil {
panic(err)
}
val, err := client.Get("name").Result()
if err != nil {
panic(err)
}
fmt.Println("name", val)
// 设置过期时间
err = client.Set("age", "20", 1*time.Second).Err()
if err != nil {
panic(err)
}
client.Incr("age")
client.Incr("age")
client.Decr("age")
val, err = client.Get("age").Result()
if err != nil {
panic(err)
}
fmt.Println("age", val)
// 设置的过期时间为1秒 1秒过后key自动被删除
time.Sleep(1 * time.Second)
val, err = client.Get("age").Result()
if err != nil {
// key "age"已经被删除返回nil
fmt.Println("err:%v", err)
}
fmt.Println("age", val)
}
// list操作
func listOperation(client *redis.Client) {
client.RPush("fruit", "apple") // 在名称为fruit的list尾添加一个值为value的元素
client.LPush("fruit", "banana") // 在名称为fruit的list头添加一个值为value的元素
length, err := client.LLen("fruit").Result() // 返回名称为fruit的list的长度
if err != nil {
panic(err)
}
fmt.Println("length:", length)
// 返回并删除名称为fruit的list中的首元素
value, err := client.LPop("fruit").Result()
if err != nil {
panic(err)
}
fmt.Println("fruit", value)
// 返回并删除名称为fruit的list中的尾元素
value, err = client.RPop("fruit").Result()
if err != nil {
panic(err)
}
fmt.Println("fruit", value)
}
// set操作
func setOperation(client *redis.Client) {
client.SAdd("blacklist", "Obama") // 向blacklist中添加元素
client.SAdd("blacklist", "Hillary") // 可以继续添加
client.SAdd("blacklist", "the Elder")
client.SAdd("whitelist", "the Elder")
// 判断元素是否在集合中
isMember, err := client.SIsMember("blacklist", "Bush").Result()
if err != nil {
panic(err)
}
fmt.Println("Is Bush in blacklist:", isMember)
// 求交集,既在black中又在white中
names, err := client.SInter("blacklist", "whitelist").Result()
if err != nil {
panic(err)
}
fmt.Println("Inter result:", names)
// 获取指定集合的所有元素
all, err := client.SMembers("blacklist").Result()
if err != nil {
panic(err)
}
fmt.Println("All member:", all)
}
// hash操作
func hashOperation(client *redis.Client) {
client.HSet("user_wang", "name", "wang") // 向名称为user_wang的hash中添加元素name
client.HSet("user_wang", "age", "18") // 向名称为user_wang的hash中添加元素age
// 批量向user_test的hash中添加元素name和age
client.HMSet("user_test", map[string]string{"name": "test", "age": "20"})
// 批量获取user_test的hash中指定字段的值
fields, err := client.HMGet("user_test", "name", "age").Result()
if err != nil {
panic(err)
}
fmt.Println("fields in user_test:", fields)
// 获取user_wang的hash中的字段个数
length, err := client.HLen("user_wang").Result()
if err != nil {
panic(err)
}
fmt.Println("fields count in user_wang:", length)
// 删除user_test的age字段
client.HDel("user_test", "age")
age, err := client.HGet("user_test", "age").Result()
if err != nil {
fmt.Println("Get user_test age error: %v", err)
} else {
fmt.Println("user_test age is:", age)
}
}
// redis.v4的连接池管理
func connectPool(client *redis.Client) {
wg := sync.WaitGroup{}
wg.Add(10)
for i := 0; i < 10; i++ {
go func() {
defer wg.Done()
for j := 0; j < 100; j++ {
client.Set(fmt.Sprintf("name%d", j), fmt.Sprintf("wang%d", j), 0).Err()
client.Get(fmt.Sprintf("name%d", j)).Result()
}
fmt.Printf("PoolStats, TotalConns: %d, FressConns: %d\n", client.PoolStats().TotalConns, client.PoolStats().FreeConns)
}()
}
wg.Wait()
}
```
|
non_build
|
go redis base go package main import fmt sync time redis gopkg in redis func main client createclient stringoperation client listoperation client setoperation client hashoperation client connectpool client 创建客户端 func createclient redis client client redis newclient redis options addr localhost password db 数据库名 poolsize 最大连接池数量 通过client ping 检查是否成功连接redis服务器 pong err client ping result fmt println pong err return client string操作 func stringoperation client redis client 第三个参数是过期时间 err client set name wnag err if err nil panic err val err client get name result if err nil panic err fmt println name val 设置过期时间 err client set age time second err if err nil panic err client incr age client incr age client decr age val err client get age result if err nil panic err fmt println age val time sleep time second val err client get age result if err nil key age 已经被删除返回nil fmt println err v err fmt println age val list操作 func listoperation client redis client client rpush fruit apple 在名称为fruit的list尾添加一个值为value的元素 client lpush fruit banana 在名称为fruit的list头添加一个值为value的元素 length err client llen fruit result 返回名称为fruit的list的长度 if err nil panic err fmt println length length 返回并删除名称为fruit的list中的首元素 value err client lpop fruit result if err nil panic err fmt println fruit value 返回并删除名称为fruit的list中的尾元素 value err client rpop fruit result if err nil panic err fmt println fruit value set操作 func setoperation client redis client client sadd blacklist obama 向blacklist中添加元素 client sadd blacklist hillary 可以继续添加 client sadd blacklist the elder client sadd whitelist the elder 判断元素是否在集合中 ismember err client sismember blacklist bush result if err nil panic err fmt println is bush in blacklist ismember 求交集,既在black中又在white中 names err client sinter blacklist whitelist result if err nil panic err fmt println inter result names 获取指定集合的所有元素 all err client smembers blacklist result if err nil panic err fmt println all member all hash操作 func hashoperation client redis client client hset user wang name wang 向名称为user wang的hash中添加元素name client hset user wang age 向名称为user wang的hash中添加元素age 批量向user test的hash中添加元素name和age client hmset user test map string name test age 批量获取user test的hash中指定字段的值 fields err client hmget user test name age result if err nil panic err fmt println fields in user test fields 获取user wang的hash中的字段个数 length err client hlen user wang result if err nil panic err fmt println fields count in user wang length 删除user test的age字段 client hdel user test age age err client hget user test age result if err nil fmt println get user test age error v err else fmt println user test age is age redis func connectpool client redis client wg sync waitgroup wg add for i i i go func defer wg done for j j j client set fmt sprintf name d j fmt sprintf wang d j err client get fmt sprintf name d j result fmt printf poolstats totalconns d fressconns d n client poolstats totalconns client poolstats freeconns wg wait
| 0
|
86,644
| 24,909,303,346
|
IssuesEvent
|
2022-10-29 17:04:00
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
LanguageTool builder fails with exit code 141
|
6.topic: nixos 0.kind: build failure
|
### Steps To Reproduce
Steps to reproduce the behavior:
1. Add `languagetool.enable = true;` to your system module's `services`
2. Rebuild your system
3. See the error in log
### Build log
```console
root@Ricky-First-Laptop /e/nixos# nixos-rebuild switch
building the system configuration...
error: builder for '/nix/store/1cvd7xgpdc94kkwxk2vqdbr37snhacic-languagetool.conf.drv' failed with exit code 141
error: 1 dependencies of derivation '/nix/store/dsrvvhrhalr8li1i9vq0r2xmb52ywcnk-unit-languagetool.service.drv' failed to build
error: 1 dependencies of derivation '/nix/store/9w66lswkz6bxh6h06vfaf59clvc9ysmy-system-units.drv' failed to build
error: 1 dependencies of derivation '/nix/store/3z6f7snlqnwv4xc1ribq8kfwhw91sfzj-etc.drv' failed to build
error: 1 dependencies of derivation '/nix/store/b56bdlln9b5s30qyz8qk0a4x4pk51yl4-nixos-system-Ricky-First-Laptop-22.11.20221025.f994293.drv' failed to build
```
### Additional context
The Nix Flake is locked on Nixpkgs Unstable rev. f994293d1eb8812f032e8919e10a594567cf6ef7
### Notify maintainers
@edwtjo
@fedeinthemix
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
root@Ricky-First-Laptop /e/nixos# nix-info -m
- system: `"x86_64-linux"`
- host os: `Linux 5.19.12-zen1, NixOS, 22.11 (Raccoon), 22.11.20221013.ba187fb`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.0`
- channels(root): `"nixos-22.05"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
|
1.0
|
LanguageTool builder fails with exit code 141 - ### Steps To Reproduce
Steps to reproduce the behavior:
1. Add `languagetool.enable = true;` to your system module's `services`
2. Rebuild your system
3. See the error in log
### Build log
```console
root@Ricky-First-Laptop /e/nixos# nixos-rebuild switch
building the system configuration...
error: builder for '/nix/store/1cvd7xgpdc94kkwxk2vqdbr37snhacic-languagetool.conf.drv' failed with exit code 141
error: 1 dependencies of derivation '/nix/store/dsrvvhrhalr8li1i9vq0r2xmb52ywcnk-unit-languagetool.service.drv' failed to build
error: 1 dependencies of derivation '/nix/store/9w66lswkz6bxh6h06vfaf59clvc9ysmy-system-units.drv' failed to build
error: 1 dependencies of derivation '/nix/store/3z6f7snlqnwv4xc1ribq8kfwhw91sfzj-etc.drv' failed to build
error: 1 dependencies of derivation '/nix/store/b56bdlln9b5s30qyz8qk0a4x4pk51yl4-nixos-system-Ricky-First-Laptop-22.11.20221025.f994293.drv' failed to build
```
### Additional context
The Nix Flake is locked on Nixpkgs Unstable rev. f994293d1eb8812f032e8919e10a594567cf6ef7
### Notify maintainers
@edwtjo
@fedeinthemix
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
root@Ricky-First-Laptop /e/nixos# nix-info -m
- system: `"x86_64-linux"`
- host os: `Linux 5.19.12-zen1, NixOS, 22.11 (Raccoon), 22.11.20221013.ba187fb`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.0`
- channels(root): `"nixos-22.05"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
|
build
|
languagetool builder fails with exit code steps to reproduce steps to reproduce the behavior add languagetool enable true to your system module s services rebuild your system see the error in log build log console root ricky first laptop e nixos nixos rebuild switch building the system configuration error builder for nix store languagetool conf drv failed with exit code error dependencies of derivation nix store unit languagetool service drv failed to build error dependencies of derivation nix store system units drv failed to build error dependencies of derivation nix store etc drv failed to build error dependencies of derivation nix store nixos system ricky first laptop drv failed to build additional context the nix flake is locked on nixpkgs unstable rev notify maintainers edwtjo fedeinthemix metadata please run nix shell p nix info run nix info m and paste the result console root ricky first laptop e nixos nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels root nixos nixpkgs nix var nix profiles per user root channels nixos
| 1
|
76,525
| 21,480,272,294
|
IssuesEvent
|
2022-04-26 17:02:06
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
node_module caching for binary system tests is flakey
|
process: build stage: needs review process: flaky test
|
### Current behavior
We generate node_module caches in CircleCI for our binary system tests and share them across multiple jobs. To do this, we hash the contents of all package.json/yarn.lock files within the `system-tests/projects` directory and use that as part of our cache key. However, depending on the executor architecture, the generated hash can differ for the same content, causing a mismatch.
We attempt to correct the mismatch by falling back to a recent cache that matches a subset of our key; however, we do not perform a subsequent install to ensure any missing dependencies are added. This has the potential to create volatility and flake in our build system.
### Desired behavior
We need to make sure that:
* Our cache keys are generated in a deterministic way across architectures
* If we don't have a valid cache, ensure that the node modules are properly installed for a project
* We don't fallback to state keys when we are actually doing partial cache fallbacks
### Test code to reproduce
Here's a recent example showing the mismatch: https://app.circleci.com/pipelines/github/cypress-io/cypress/36440/workflows/b32ce66c-e33a-478b-9888-e0046b815fa7/jobs/1459760?invite=true#step-112-1
But this occurs for any build at the moment. Whether or not it's a problem will depend on what else has built recently and how recently the package/lock contents have changed.
### Cypress Version
N/A
### Other
_No response_
|
1.0
|
node_module caching for binary system tests is flakey - ### Current behavior
We generate node_module caches in CircleCI for our binary system tests and share them across multiple jobs. To do this, we hash the contents of all package.json/yarn.lock files within the `system-tests/projects` directory and use that as part of our cache key. However, depending on the executor architecture, the generated hash can differ for the same content, causing a mismatch.
We attempt to correct the mismatch by falling back to a recent cache that matches a subset of our key; however, we do not perform a subsequent install to ensure any missing dependencies are added. This has the potential to create volatility and flake in our build system.
### Desired behavior
We need to make sure that:
* Our cache keys are generated in a deterministic way across architectures
* If we don't have a valid cache, ensure that the node modules are properly installed for a project
* We don't fallback to state keys when we are actually doing partial cache fallbacks
### Test code to reproduce
Here's a recent example showing the mismatch: https://app.circleci.com/pipelines/github/cypress-io/cypress/36440/workflows/b32ce66c-e33a-478b-9888-e0046b815fa7/jobs/1459760?invite=true#step-112-1
But this occurs for any build at the moment. Whether or not it's a problem will depend on what else has built recently and how recently the package/lock contents have changed.
### Cypress Version
N/A
### Other
_No response_
|
build
|
node module caching for binary system tests is flakey current behavior we generate node module caches in circleci for our binary system tests and share them across multiple jobs to do this we hash the contents of all package json yarn lock files within the system tests projects directory and use that as part of our cache key however depending on the executor architecture the generated hash can differ for the same content causing a mismatch we attempt to correct the mismatch by falling back to a recent cache that matches a subset of our key however we do not perform a subsequent install to ensure any missing dependencies are added this has the potential to create volatility and flake in our build system desired behavior we need to make sure that our cache keys are generated in a deterministic way across architectures if we don t have a valid cache ensure that the node modules are properly installed for a project we don t fallback to state keys when we are actually doing partial cache fallbacks test code to reproduce here s a recent example showing the mismatch but this occurs for any build at the moment whether or not it s a problem will depend on what else has built recently and how recently the package lock contents have changed cypress version n a other no response
| 1
|
52,687
| 6,650,277,985
|
IssuesEvent
|
2017-09-28 15:47:33
|
18F/openFEC-web-app
|
https://api.github.com/repos/18F/openFEC-web-app
|
closed
|
Add process details and role documentation to research<>design process diagram
|
Internal Work: Design
|
Using the process sketch from https://github.com/18F/openFEC-web-app/issues/1352
It could be helpful for new team members, deliverable documentation, and 18F comms/outreach for us to add detail about how each step is carried out: both more detail in how to do it that step, and who is involved/responsible for completing.
Filing this issue now, but putting it in the backlog for a few sprints as we test out some of the process changes.
|
1.0
|
Add process details and role documentation to research<>design process diagram - Using the process sketch from https://github.com/18F/openFEC-web-app/issues/1352
It could be helpful for new team members, deliverable documentation, and 18F comms/outreach for us to add detail about how each step is carried out: both more detail in how to do it that step, and who is involved/responsible for completing.
Filing this issue now, but putting it in the backlog for a few sprints as we test out some of the process changes.
|
non_build
|
add process details and role documentation to research design process diagram using the process sketch from it could be helpful for new team members deliverable documentation and comms outreach for us to add detail about how each step is carried out both more detail in how to do it that step and who is involved responsible for completing filing this issue now but putting it in the backlog for a few sprints as we test out some of the process changes
| 0
|
37,967
| 5,152,906,018
|
IssuesEvent
|
2017-01-14 05:26:29
|
nltk/nltk
|
https://api.github.com/repos/nltk/nltk
|
closed
|
DeprecationWarnings for @memoize for Python3.5
|
python3 tests
|
The `memoize` decorator needs some updating such that it's using the new/better `inspect` from Python3.5
```bash
~/git-stuff/nltk$ python3 -W error -m nltk
Traceback (most recent call last):
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 174, in _run_module_as_main
mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 133, in _get_module_details
return _get_module_details(pkg_main_name, error)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 109, in _get_module_details
__import__(pkg_name)
File "/Users/liling.tan/git-stuff/nltk/nltk/__init__.py", line 115, in <module>
from nltk.decorators import decorator, memoize
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 183, in <module>
@decorator
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 172, in decorator
return update_wrapper(_decorator, caller)
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 83, in update_wrapper
infodict = infodict or getinfo(model)
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 59, in getinfo
regargs, varargs, varkwargs, defaults = inspect.getargspec(func)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/inspect.py", line 1041, in getargspec
stacklevel=2)
DeprecationWarning: inspect.getargspec() is deprecated, use inspect.signature() instead
```
Related to #630
|
1.0
|
DeprecationWarnings for @memoize for Python3.5 - The `memoize` decorator needs some updating such that it's using the new/better `inspect` from Python3.5
```bash
~/git-stuff/nltk$ python3 -W error -m nltk
Traceback (most recent call last):
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 174, in _run_module_as_main
mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 133, in _get_module_details
return _get_module_details(pkg_main_name, error)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 109, in _get_module_details
__import__(pkg_name)
File "/Users/liling.tan/git-stuff/nltk/nltk/__init__.py", line 115, in <module>
from nltk.decorators import decorator, memoize
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 183, in <module>
@decorator
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 172, in decorator
return update_wrapper(_decorator, caller)
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 83, in update_wrapper
infodict = infodict or getinfo(model)
File "/Users/liling.tan/git-stuff/nltk/nltk/decorators.py", line 59, in getinfo
regargs, varargs, varkwargs, defaults = inspect.getargspec(func)
File "/usr/local/Cellar/python3/3.5.2_3/Frameworks/Python.framework/Versions/3.5/lib/python3.5/inspect.py", line 1041, in getargspec
stacklevel=2)
DeprecationWarning: inspect.getargspec() is deprecated, use inspect.signature() instead
```
Related to #630
|
non_build
|
deprecationwarnings for memoize for the memoize decorator needs some updating such that it s using the new better inspect from bash git stuff nltk w error m nltk traceback most recent call last file usr local cellar frameworks python framework versions lib runpy py line in run module as main mod name mod spec code get module details mod name error file usr local cellar frameworks python framework versions lib runpy py line in get module details return get module details pkg main name error file usr local cellar frameworks python framework versions lib runpy py line in get module details import pkg name file users liling tan git stuff nltk nltk init py line in from nltk decorators import decorator memoize file users liling tan git stuff nltk nltk decorators py line in decorator file users liling tan git stuff nltk nltk decorators py line in decorator return update wrapper decorator caller file users liling tan git stuff nltk nltk decorators py line in update wrapper infodict infodict or getinfo model file users liling tan git stuff nltk nltk decorators py line in getinfo regargs varargs varkwargs defaults inspect getargspec func file usr local cellar frameworks python framework versions lib inspect py line in getargspec stacklevel deprecationwarning inspect getargspec is deprecated use inspect signature instead related to
| 0
|
101,426
| 21,682,976,830
|
IssuesEvent
|
2022-05-09 08:34:17
|
MJZwart/motivation-app
|
https://api.github.com/repos/MJZwart/motivation-app
|
closed
|
Reduce store usage for component-wide data
|
front-end code-streamline
|
Currently I use the store for all the back-end stuff, but the store isn't needed for when data is only used within the component
|
1.0
|
Reduce store usage for component-wide data - Currently I use the store for all the back-end stuff, but the store isn't needed for when data is only used within the component
|
non_build
|
reduce store usage for component wide data currently i use the store for all the back end stuff but the store isn t needed for when data is only used within the component
| 0
|
64,686
| 16,014,290,280
|
IssuesEvent
|
2021-04-20 14:20:08
|
hashicorp/packer-plugin-outscale
|
https://api.github.com/repos/hashicorp/packer-plugin-outscale
|
opened
|
Outscale: Support for owner account aliases in source_omi_filter
|
bug builder/outscale
|
_This issue was originally opened by @Hakujou as hashicorp/packer#10915. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._
<hr>
#### Overview of the Issue
source_omi_filter in Outscale builders requires "owners" to be set for security reasons ("* For security reasons, your source AMI filter must declare an owner.").
There's no way to set owner pointing to an Outscale Account alias instead of ID.
Per API documentation (https://docs.outscale.com/api#readimages), account aliases should be set as "AccountAliases", not "AccountIds".
#### Reproduction Steps
Set inside your HCL2 Packer script:
```
source_omi_filter {
filters = {
name = "SourceImageName"
}
owners = ["Outscale"]
}
```
Then execute packer build xxx.pkr.hcl. It will return:
```
No OMI was found matching filters: osc.ReadImagesRequest{DryRun:false, Filters:osc.FiltersImage{AccountAliases:[]string(nil), AccountIds:[]string{"Outscale"}...
```
We clearly see that "Outscale" is passed to AccountIds, not AccountAliases.
### Packer version
Packer v1.7.0
ping @marinsalinas
|
1.0
|
Outscale: Support for owner account aliases in source_omi_filter - _This issue was originally opened by @Hakujou as hashicorp/packer#10915. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._
<hr>
#### Overview of the Issue
source_omi_filter in Outscale builders requires "owners" to be set for security reasons ("* For security reasons, your source AMI filter must declare an owner.").
There's no way to set owner pointing to an Outscale Account alias instead of ID.
Per API documentation (https://docs.outscale.com/api#readimages), account aliases should be set as "AccountAliases", not "AccountIds".
#### Reproduction Steps
Set inside your HCL2 Packer script:
```
source_omi_filter {
filters = {
name = "SourceImageName"
}
owners = ["Outscale"]
}
```
Then execute packer build xxx.pkr.hcl. It will return:
```
No OMI was found matching filters: osc.ReadImagesRequest{DryRun:false, Filters:osc.FiltersImage{AccountAliases:[]string(nil), AccountIds:[]string{"Outscale"}...
```
We clearly see that "Outscale" is passed to AccountIds, not AccountAliases.
### Packer version
Packer v1.7.0
ping @marinsalinas
|
build
|
outscale support for owner account aliases in source omi filter this issue was originally opened by hakujou as hashicorp packer it was migrated here as a result of the the original body of the issue is below overview of the issue source omi filter in outscale builders requires owners to be set for security reasons for security reasons your source ami filter must declare an owner there s no way to set owner pointing to an outscale account alias instead of id per api documentation account aliases should be set as accountaliases not accountids reproduction steps set inside your packer script source omi filter filters name sourceimagename owners then execute packer build xxx pkr hcl it will return no omi was found matching filters osc readimagesrequest dryrun false filters osc filtersimage accountaliases string nil accountids string outscale we clearly see that outscale is passed to accountids not accountaliases packer version packer ping marinsalinas
| 1
|
26,104
| 7,783,091,900
|
IssuesEvent
|
2018-06-06 08:51:04
|
uon-coehm/IQCare
|
https://api.github.com/repos/uon-coehm/IQCare
|
closed
|
Pyschosocial Enrollment form: Add field "Child dwelling information"
|
Form Builder
|
Add a select field `Child dwelling information` with the following options:
- Lives with parents
- Orphanage Children Shelter
- Aunt/uncle
- Grandparents
- Lives in the streets
- With well wishers
- Other (Specify) - This opens a free text field to specify
|
1.0
|
Pyschosocial Enrollment form: Add field "Child dwelling information" - Add a select field `Child dwelling information` with the following options:
- Lives with parents
- Orphanage Children Shelter
- Aunt/uncle
- Grandparents
- Lives in the streets
- With well wishers
- Other (Specify) - This opens a free text field to specify
|
build
|
pyschosocial enrollment form add field child dwelling information add a select field child dwelling information with the following options lives with parents orphanage children shelter aunt uncle grandparents lives in the streets with well wishers other specify this opens a free text field to specify
| 1
|
732,748
| 25,275,133,790
|
IssuesEvent
|
2022-11-16 12:06:20
|
epiphany-platform/epiphany
|
https://api.github.com/repos/epiphany-platform/epiphany
|
closed
|
[SPIKE] Usage of Image Digests in requirements download
|
type/spike priority/low area/repository
|
**Is your spike related to a problem or idea? Please describe.**
Currently in download requirements a way to select specific image is to use **tag** (and them checking downloaded image checksum), like:
```
'quay.io/cephcsi/cephcsi:v3.5.1':
sha1: 51dee9ea8ad76fb95ebd16f951e8ffaaaba95eb6
```
Unfortunately tags can be mutable (we had this problem with ceph and image registry images) which can lead to the situation where Epiphany installation is reporting that image has changed - that happens when image used has been updated and assigned to the same tag.
In order to avoid this kind of problems, we can research usage of **container image digests**.
According to the docs: `Images that use the v2 or later format have a content-addressable identifier called a digest. As long as the input used to generate the image is unchanged, the digest value is predictable.`
Therefore it is a more stable way to specify image version.
**Describe the outcome you'd like**
Verify if:
- `image digests` are better solution that tags (safe, stable etc.)
- `image digests` can be used in Epiphany and if their usage is not breaking any functionality
**What is the reason or source for the spike**
So far we faced problem that images were updated and assigned to the same tag, which led to checksum mismatch.
**Additional context**
https://docs.docker.com/engine/reference/commandline/images/#list-image-digests
https://github.com/opencontainers/image-spec/blob/main/descriptor.md#digests
https://www.mikenewswanger.com/posts/2020/docker-image-digests/
https://www.ibm.com/docs/en/filenet-p8-platform/5.5.x?topic=deployment-choosing-image-tags-digests
---
**DoD checklist**
* [ ] Reader is able to understand the results of spike
* [ ] The results of the spike are presented in a table (to show simply what are compared or researched parameters) / not applicable
* [ ] Each value / cell in the results table is described more deeply below
* [ ] Demo of the spike (automated as much as possible)
* [ ] Design doc updated
|
1.0
|
[SPIKE] Usage of Image Digests in requirements download - **Is your spike related to a problem or idea? Please describe.**
Currently in download requirements a way to select specific image is to use **tag** (and them checking downloaded image checksum), like:
```
'quay.io/cephcsi/cephcsi:v3.5.1':
sha1: 51dee9ea8ad76fb95ebd16f951e8ffaaaba95eb6
```
Unfortunately tags can be mutable (we had this problem with ceph and image registry images) which can lead to the situation where Epiphany installation is reporting that image has changed - that happens when image used has been updated and assigned to the same tag.
In order to avoid this kind of problems, we can research usage of **container image digests**.
According to the docs: `Images that use the v2 or later format have a content-addressable identifier called a digest. As long as the input used to generate the image is unchanged, the digest value is predictable.`
Therefore it is a more stable way to specify image version.
**Describe the outcome you'd like**
Verify if:
- `image digests` are better solution that tags (safe, stable etc.)
- `image digests` can be used in Epiphany and if their usage is not breaking any functionality
**What is the reason or source for the spike**
So far we faced problem that images were updated and assigned to the same tag, which led to checksum mismatch.
**Additional context**
https://docs.docker.com/engine/reference/commandline/images/#list-image-digests
https://github.com/opencontainers/image-spec/blob/main/descriptor.md#digests
https://www.mikenewswanger.com/posts/2020/docker-image-digests/
https://www.ibm.com/docs/en/filenet-p8-platform/5.5.x?topic=deployment-choosing-image-tags-digests
---
**DoD checklist**
* [ ] Reader is able to understand the results of spike
* [ ] The results of the spike are presented in a table (to show simply what are compared or researched parameters) / not applicable
* [ ] Each value / cell in the results table is described more deeply below
* [ ] Demo of the spike (automated as much as possible)
* [ ] Design doc updated
|
non_build
|
usage of image digests in requirements download is your spike related to a problem or idea please describe currently in download requirements a way to select specific image is to use tag and them checking downloaded image checksum like quay io cephcsi cephcsi unfortunately tags can be mutable we had this problem with ceph and image registry images which can lead to the situation where epiphany installation is reporting that image has changed that happens when image used has been updated and assigned to the same tag in order to avoid this kind of problems we can research usage of container image digests according to the docs images that use the or later format have a content addressable identifier called a digest as long as the input used to generate the image is unchanged the digest value is predictable therefore it is a more stable way to specify image version describe the outcome you d like verify if image digests are better solution that tags safe stable etc image digests can be used in epiphany and if their usage is not breaking any functionality what is the reason or source for the spike so far we faced problem that images were updated and assigned to the same tag which led to checksum mismatch additional context dod checklist reader is able to understand the results of spike the results of the spike are presented in a table to show simply what are compared or researched parameters not applicable each value cell in the results table is described more deeply below demo of the spike automated as much as possible design doc updated
| 0
|
56,393
| 11,578,485,850
|
IssuesEvent
|
2020-02-21 16:02:12
|
JuliaLang/julia
|
https://api.github.com/repos/JuliaLang/julia
|
closed
|
LLVM assertion failure in core tests
|
bug codegen
|
I get this on master with LLVM assertions enabled:
```
Test (Worker) | Time (s) | GC (s) | GC % | Alloc (MB) | RSS (MB)
core (1) | started at 2020-02-20T14:14:10.279
julia: /home/jeff/src/julia/deps/srccache/llvm-9.0.1/lib/IR/Constants.cpp:1005: static llvm::Constant* llvm::ConstantArray::getImpl(llvm::ArrayType*, llvm::ArrayRef<llvm::Constant*>): Assertion `V[i]->getType() == Ty->getElementType() && "Wrong type in array element initializer"' failed.
signal (6): Aborted
in expression starting at /home/jeff/src/julia/test/core.jl:4930
gsignal at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
abort at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
unknown function (ip: 0x7f0196091399)
__assert_fail at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
_ZN4llvm13ConstantArray7getImplEPNS_9ArrayTypeENS_8ArrayRefIPNS_8ConstantEEE at /home/jeff/src/julia/usr/bin/../lib/libLLVM-9jl.so (unknown line)
_ZN4llvm13ConstantArray3getEPNS_9ArrayTypeENS_8ArrayRefIPNS_8ConstantEEE at /home/jeff/src/julia/usr/bin/../lib/libLLVM-9jl.so (unknown line)
julia_const_to_llvm at /home/jeff/src/julia/src/intrinsics.cpp:260
emit_unbox at /home/jeff/src/julia/src/intrinsics.cpp:333
julia_to_native at /usr/include/c++/7/bits/stl_bvector.h:81 [inlined]
emit_a_ccall at /home/jeff/src/julia/src/ccall.cpp:1756
emit_ccall at /home/jeff/src/julia/src/ccall.cpp:1702
emit_expr at /home/jeff/src/julia/src/codegen.cpp:4170
emit_ssaval_assign at /home/jeff/src/julia/src/codegen.cpp:3862
emit_stmtpos at /home/jeff/src/julia/src/codegen.cpp:4056 [inlined]
emit_function at /home/jeff/src/julia/src/codegen.cpp:6665
jl_compile_linfo at /home/jeff/src/julia/src/codegen.cpp:1256
```
|
1.0
|
LLVM assertion failure in core tests - I get this on master with LLVM assertions enabled:
```
Test (Worker) | Time (s) | GC (s) | GC % | Alloc (MB) | RSS (MB)
core (1) | started at 2020-02-20T14:14:10.279
julia: /home/jeff/src/julia/deps/srccache/llvm-9.0.1/lib/IR/Constants.cpp:1005: static llvm::Constant* llvm::ConstantArray::getImpl(llvm::ArrayType*, llvm::ArrayRef<llvm::Constant*>): Assertion `V[i]->getType() == Ty->getElementType() && "Wrong type in array element initializer"' failed.
signal (6): Aborted
in expression starting at /home/jeff/src/julia/test/core.jl:4930
gsignal at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
abort at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
unknown function (ip: 0x7f0196091399)
__assert_fail at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
_ZN4llvm13ConstantArray7getImplEPNS_9ArrayTypeENS_8ArrayRefIPNS_8ConstantEEE at /home/jeff/src/julia/usr/bin/../lib/libLLVM-9jl.so (unknown line)
_ZN4llvm13ConstantArray3getEPNS_9ArrayTypeENS_8ArrayRefIPNS_8ConstantEEE at /home/jeff/src/julia/usr/bin/../lib/libLLVM-9jl.so (unknown line)
julia_const_to_llvm at /home/jeff/src/julia/src/intrinsics.cpp:260
emit_unbox at /home/jeff/src/julia/src/intrinsics.cpp:333
julia_to_native at /usr/include/c++/7/bits/stl_bvector.h:81 [inlined]
emit_a_ccall at /home/jeff/src/julia/src/ccall.cpp:1756
emit_ccall at /home/jeff/src/julia/src/ccall.cpp:1702
emit_expr at /home/jeff/src/julia/src/codegen.cpp:4170
emit_ssaval_assign at /home/jeff/src/julia/src/codegen.cpp:3862
emit_stmtpos at /home/jeff/src/julia/src/codegen.cpp:4056 [inlined]
emit_function at /home/jeff/src/julia/src/codegen.cpp:6665
jl_compile_linfo at /home/jeff/src/julia/src/codegen.cpp:1256
```
|
non_build
|
llvm assertion failure in core tests i get this on master with llvm assertions enabled test worker time s gc s gc alloc mb rss mb core started at julia home jeff src julia deps srccache llvm lib ir constants cpp static llvm constant llvm constantarray getimpl llvm arraytype llvm arrayref assertion v gettype ty getelementtype wrong type in array element initializer failed signal aborted in expression starting at home jeff src julia test core jl gsignal at lib linux gnu libc so unknown line abort at lib linux gnu libc so unknown line unknown function ip assert fail at lib linux gnu libc so unknown line at home jeff src julia usr bin lib libllvm so unknown line at home jeff src julia usr bin lib libllvm so unknown line julia const to llvm at home jeff src julia src intrinsics cpp emit unbox at home jeff src julia src intrinsics cpp julia to native at usr include c bits stl bvector h emit a ccall at home jeff src julia src ccall cpp emit ccall at home jeff src julia src ccall cpp emit expr at home jeff src julia src codegen cpp emit ssaval assign at home jeff src julia src codegen cpp emit stmtpos at home jeff src julia src codegen cpp emit function at home jeff src julia src codegen cpp jl compile linfo at home jeff src julia src codegen cpp
| 0
|
58,118
| 14,283,218,032
|
IssuesEvent
|
2020-11-23 10:41:57
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
bazel build libtensorflowlite has issue during compiling
|
comp:lite stalled stat:awaiting response type:build/install
|
tizen5.0_cross_toolchain_vd_kant_target_armv7l_host_x86-64/bin/armv7l-tizen-linux-gnueabi-gcc -MD -MF bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.d '-frandom-seed=bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.o' -iquote external/mkl_dnn -iquote bazel-out/armv7-opt/bin/external/mkl_dnn -isystem external/mkl_dnn/include -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/include -isystem external/mkl_dnn/src -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src -isystem external/mkl_dnn/src/common -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/common -isystem external/mkl_dnn/src/cpu -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu -isystem external/mkl_dnn/src/cpu/gemm -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu/gemm -isystem external/mkl_dnn/src/cpu/xbyak -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu/xbyak -w -DAUTOLOAD_DYNAMIC_KERNELS -fPIC '-std=c++14' -fexceptions '-DMKLDNN_THR=MKLDNN_THR_SEQ' -c external/mkl_dnn/src/cpu/simple_concat.cpp -o bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.o)
Execution platform: @local_execution_config_platform//:platform
In file included from external/mkl_dnn/src/common/mkldnn_thread.hpp:20:0,
from external/mkl_dnn/src/cpu/simple_concat.cpp:17:
external/mkl_dnn/src/common/utils.hpp:45:1: error: static assertion failed: Intel(R) MKL-DNN supports 64 bit only
static_assert(sizeof(void*) == 8, "Intel(R) MKL-DNN supports 64 bit only");
^~~~~~~~~~~~~
Target //tensorflow/lite:libtensorflowlite.so failed to build
|
1.0
|
bazel build libtensorflowlite has issue during compiling - tizen5.0_cross_toolchain_vd_kant_target_armv7l_host_x86-64/bin/armv7l-tizen-linux-gnueabi-gcc -MD -MF bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.d '-frandom-seed=bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.o' -iquote external/mkl_dnn -iquote bazel-out/armv7-opt/bin/external/mkl_dnn -isystem external/mkl_dnn/include -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/include -isystem external/mkl_dnn/src -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src -isystem external/mkl_dnn/src/common -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/common -isystem external/mkl_dnn/src/cpu -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu -isystem external/mkl_dnn/src/cpu/gemm -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu/gemm -isystem external/mkl_dnn/src/cpu/xbyak -isystem bazel-out/armv7-opt/bin/external/mkl_dnn/src/cpu/xbyak -w -DAUTOLOAD_DYNAMIC_KERNELS -fPIC '-std=c++14' -fexceptions '-DMKLDNN_THR=MKLDNN_THR_SEQ' -c external/mkl_dnn/src/cpu/simple_concat.cpp -o bazel-out/armv7-opt/bin/external/mkl_dnn/_objs/mkldnn_single_threaded/simple_concat.o)
Execution platform: @local_execution_config_platform//:platform
In file included from external/mkl_dnn/src/common/mkldnn_thread.hpp:20:0,
from external/mkl_dnn/src/cpu/simple_concat.cpp:17:
external/mkl_dnn/src/common/utils.hpp:45:1: error: static assertion failed: Intel(R) MKL-DNN supports 64 bit only
static_assert(sizeof(void*) == 8, "Intel(R) MKL-DNN supports 64 bit only");
^~~~~~~~~~~~~
Target //tensorflow/lite:libtensorflowlite.so failed to build
|
build
|
bazel build libtensorflowlite has issue during compiling cross toolchain vd kant target host bin tizen linux gnueabi gcc md mf bazel out opt bin external mkl dnn objs mkldnn single threaded simple concat d frandom seed bazel out opt bin external mkl dnn objs mkldnn single threaded simple concat o iquote external mkl dnn iquote bazel out opt bin external mkl dnn isystem external mkl dnn include isystem bazel out opt bin external mkl dnn include isystem external mkl dnn src isystem bazel out opt bin external mkl dnn src isystem external mkl dnn src common isystem bazel out opt bin external mkl dnn src common isystem external mkl dnn src cpu isystem bazel out opt bin external mkl dnn src cpu isystem external mkl dnn src cpu gemm isystem bazel out opt bin external mkl dnn src cpu gemm isystem external mkl dnn src cpu xbyak isystem bazel out opt bin external mkl dnn src cpu xbyak w dautoload dynamic kernels fpic std c fexceptions dmkldnn thr mkldnn thr seq c external mkl dnn src cpu simple concat cpp o bazel out opt bin external mkl dnn objs mkldnn single threaded simple concat o execution platform local execution config platform platform in file included from external mkl dnn src common mkldnn thread hpp from external mkl dnn src cpu simple concat cpp external mkl dnn src common utils hpp error static assertion failed intel r mkl dnn supports bit only static assert sizeof void intel r mkl dnn supports bit only target tensorflow lite libtensorflowlite so failed to build
| 1
|
444,838
| 31,151,275,359
|
IssuesEvent
|
2023-08-16 10:06:49
|
Swarmies/sap-web
|
https://api.github.com/repos/Swarmies/sap-web
|
closed
|
Workflow sematic file documentation
|
documentation
|
# Description
Add a documentation on how to use the sematic `types' on PR
## Acceptable Criteria
- definition of the 'types' explaining their use.
## Related Story
This tickets relate to the #9
|
1.0
|
Workflow sematic file documentation - # Description
Add a documentation on how to use the sematic `types' on PR
## Acceptable Criteria
- definition of the 'types' explaining their use.
## Related Story
This tickets relate to the #9
|
non_build
|
workflow sematic file documentation description add a documentation on how to use the sematic types on pr acceptable criteria definition of the types explaining their use related story this tickets relate to the
| 0
|
63,122
| 15,436,633,757
|
IssuesEvent
|
2021-03-07 13:48:12
|
kubevirt/kubevirt
|
https://api.github.com/repos/kubevirt/kubevirt
|
reopened
|
[rfe_id:3423][crit:high][vendor:cnv-qe@redhat.com][level:component]VmWatch [test_id:3466]Should update vmi status with the proper columns using 'kubectl get vmi -w'
|
kind/bug triage/build-watcher
|
**Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
/triage build-watcher
**What happened**:
https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_kubevirt/4953/pull-kubevirt-e2e-k8s-1.20/1356962067667488768
```
tests/vm_watch_test.go:218
Timed out after 60.000s.
Expected
<bool>: false
to be true
tests/vm_watch_test.go:334
```
**What you expected to happen**:
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- KubeVirt version (use `virtctl version`):
- Kubernetes version (use `kubectl version`):
- VM or VMI specifications:
- Cloud provider or hardware configuration:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
1.0
|
[rfe_id:3423][crit:high][vendor:cnv-qe@redhat.com][level:component]VmWatch [test_id:3466]Should update vmi status with the proper columns using 'kubectl get vmi -w' - **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
/triage build-watcher
**What happened**:
https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_kubevirt/4953/pull-kubevirt-e2e-k8s-1.20/1356962067667488768
```
tests/vm_watch_test.go:218
Timed out after 60.000s.
Expected
<bool>: false
to be true
tests/vm_watch_test.go:334
```
**What you expected to happen**:
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- KubeVirt version (use `virtctl version`):
- Kubernetes version (use `kubectl version`):
- VM or VMI specifications:
- Cloud provider or hardware configuration:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
build
|
vmwatch should update vmi status with the proper columns using kubectl get vmi w is this a bug report or feature request kind bug triage build watcher what happened tests vm watch test go timed out after expected false to be true tests vm watch test go what you expected to happen how to reproduce it as minimally and precisely as possible anything else we need to know environment kubevirt version use virtctl version kubernetes version use kubectl version vm or vmi specifications cloud provider or hardware configuration os e g from etc os release kernel e g uname a install tools others
| 1
|
10,811
| 4,827,613,478
|
IssuesEvent
|
2016-11-07 14:10:20
|
CartoDB/cartodb
|
https://api.github.com/repos/CartoDB/cartodb
|
opened
|
Filters applied by widget persist if the widget is deleted
|
bug Builder
|
### Context
If a widget is used for filtering a layer and then the widget is removed, the filter persists.
### Steps to Reproduce
1. Create a map in the Builder
2. Add a widget that allows filtering (e.g. a Category or Histogram widget)
3. Filter the data with the widget (e.g. click on a category or select a range in the histogram)
4. Remove the widget (click "Delete..." on it's menu)
### Current Result
The map remains filtered (showing only the category or range selected in the widget not longer present)
### Expected result
The filter should be removed and all the data visible.
|
1.0
|
Filters applied by widget persist if the widget is deleted - ### Context
If a widget is used for filtering a layer and then the widget is removed, the filter persists.
### Steps to Reproduce
1. Create a map in the Builder
2. Add a widget that allows filtering (e.g. a Category or Histogram widget)
3. Filter the data with the widget (e.g. click on a category or select a range in the histogram)
4. Remove the widget (click "Delete..." on it's menu)
### Current Result
The map remains filtered (showing only the category or range selected in the widget not longer present)
### Expected result
The filter should be removed and all the data visible.
|
build
|
filters applied by widget persist if the widget is deleted context if a widget is used for filtering a layer and then the widget is removed the filter persists steps to reproduce create a map in the builder add a widget that allows filtering e g a category or histogram widget filter the data with the widget e g click on a category or select a range in the histogram remove the widget click delete on it s menu current result the map remains filtered showing only the category or range selected in the widget not longer present expected result the filter should be removed and all the data visible
| 1
|
12,210
| 5,166,586,431
|
IssuesEvent
|
2017-01-17 16:34:29
|
kyamagu/mexopencv
|
https://api.github.com/repos/kyamagu/mexopencv
|
closed
|
Error when compiling in matlab
|
build/install windows
|
Hello,
I have Matlab 2016a. I followed the OpenCV installation instructions in the wiki exactly as far as I know. Yet, when I ran: `mexopencv.make('opencv_path','C:\dev\build\install', 'opencv_contrib',true)`, I got the following error tree:
```
Error using strcmp
Not enough input arguments.
Error in mexopencv.make>compiler_str (line 363)
if strcmp(cc.Manufacturer, 'Microsoft')
Error in mexopencv.make>pkg_config (line 314)
L_path = fullfile(opts.opencv_path, arch_str(), compiler_str(), 'lib');
Error in mexopencv.make>mex_options (line 256)
[cv_cflags, cv_libs] = pkg_config(opts);
Error in mexopencv.make (line 81)
mex_flags = mex_options(opts);
```
I've tried to debug this and it looks like the `cc` object - generated by calling `mex.getCompilerConfigurations('C++','Selected')` doesn't set `cc.Manufacturer` to return a string for `strcmp` to act on. But I don't know if this is the true problem or how to fix things if it is.
Any help would be appreciated
|
1.0
|
Error when compiling in matlab - Hello,
I have Matlab 2016a. I followed the OpenCV installation instructions in the wiki exactly as far as I know. Yet, when I ran: `mexopencv.make('opencv_path','C:\dev\build\install', 'opencv_contrib',true)`, I got the following error tree:
```
Error using strcmp
Not enough input arguments.
Error in mexopencv.make>compiler_str (line 363)
if strcmp(cc.Manufacturer, 'Microsoft')
Error in mexopencv.make>pkg_config (line 314)
L_path = fullfile(opts.opencv_path, arch_str(), compiler_str(), 'lib');
Error in mexopencv.make>mex_options (line 256)
[cv_cflags, cv_libs] = pkg_config(opts);
Error in mexopencv.make (line 81)
mex_flags = mex_options(opts);
```
I've tried to debug this and it looks like the `cc` object - generated by calling `mex.getCompilerConfigurations('C++','Selected')` doesn't set `cc.Manufacturer` to return a string for `strcmp` to act on. But I don't know if this is the true problem or how to fix things if it is.
Any help would be appreciated
|
build
|
error when compiling in matlab hello i have matlab i followed the opencv installation instructions in the wiki exactly as far as i know yet when i ran mexopencv make opencv path c dev build install opencv contrib true i got the following error tree error using strcmp not enough input arguments error in mexopencv make compiler str line if strcmp cc manufacturer microsoft error in mexopencv make pkg config line l path fullfile opts opencv path arch str compiler str lib error in mexopencv make mex options line pkg config opts error in mexopencv make line mex flags mex options opts i ve tried to debug this and it looks like the cc object generated by calling mex getcompilerconfigurations c selected doesn t set cc manufacturer to return a string for strcmp to act on but i don t know if this is the true problem or how to fix things if it is any help would be appreciated
| 1
|
37,061
| 9,955,876,342
|
IssuesEvent
|
2019-07-05 12:24:11
|
scality/metalk8s
|
https://api.github.com/repos/scality/metalk8s
|
closed
|
fix LocalImage cleanup
|
bug build easy moonshot
|
**Component**:
buildchain
**What happened**:
`doit.sh clean` doesn't remove the locally built images from the build tree.
**What was expected**:
The locally built images are properly deleted during the cleanup.
**Steps to reproduce**
From a clean build tree, run:
- build the images: `./doit.sh -n 4 _image_build`
- clean them up: `./doit.sh clean`
- check that they aren't deleted with `ls _build/root/images/*/*`
**Resolution proposal**:
In `buildchain/buildchain/targets/local_image.py`, there is this code to select the cleanup action.
```python
'clean': True if self.save_on_disk else [self.clean]
```
Simply swap the branches of the condition…
|
1.0
|
fix LocalImage cleanup - **Component**:
buildchain
**What happened**:
`doit.sh clean` doesn't remove the locally built images from the build tree.
**What was expected**:
The locally built images are properly deleted during the cleanup.
**Steps to reproduce**
From a clean build tree, run:
- build the images: `./doit.sh -n 4 _image_build`
- clean them up: `./doit.sh clean`
- check that they aren't deleted with `ls _build/root/images/*/*`
**Resolution proposal**:
In `buildchain/buildchain/targets/local_image.py`, there is this code to select the cleanup action.
```python
'clean': True if self.save_on_disk else [self.clean]
```
Simply swap the branches of the condition…
|
build
|
fix localimage cleanup component buildchain what happened doit sh clean doesn t remove the locally built images from the build tree what was expected the locally built images are properly deleted during the cleanup steps to reproduce from a clean build tree run build the images doit sh n image build clean them up doit sh clean check that they aren t deleted with ls build root images resolution proposal in buildchain buildchain targets local image py there is this code to select the cleanup action python clean true if self save on disk else simply swap the branches of the condition…
| 1
|
112,389
| 9,564,143,384
|
IssuesEvent
|
2019-05-05 00:22:09
|
KingAkeem/TorBotApp
|
https://api.github.com/repos/KingAkeem/TorBotApp
|
opened
|
Design method of testing GetLinks and GetEmails functions
|
enhancement test added test improvement
|
I'm not sure how to accurately test these methods since we have to be able to generate HTML that can be parsed by the [HTML parsing library](https://godoc.org/golang.org/x/net/html) being used.
|
2.0
|
Design method of testing GetLinks and GetEmails functions - I'm not sure how to accurately test these methods since we have to be able to generate HTML that can be parsed by the [HTML parsing library](https://godoc.org/golang.org/x/net/html) being used.
|
non_build
|
design method of testing getlinks and getemails functions i m not sure how to accurately test these methods since we have to be able to generate html that can be parsed by the being used
| 0
|
333,110
| 10,115,689,933
|
IssuesEvent
|
2019-07-30 22:37:42
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.zertifikate.commerzbank.de - site is not usable
|
browser-firefox-mobile engine-gecko priority-important severity-critical sitepatch-applied status-first-contact
|
<!-- @browser: Firefox Mobile 57.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.4.4; Mobile; rv:57.0) Gecko/57.0 Firefox/57.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.zertifikate.commerzbank.de/webforms/mobile/index.html?home/{"site":{"country":1,"language":1,"isoCode":"de-DE"}}
**Browser / Version**: Firefox Mobile 57.0
**Operating System**: Android 4.4.4
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: site not displayed correctly at all
**Steps to Reproduce**:
Direct link, please check with other mobile browsers.
[](https://webcompat.com/uploads/2017/8/925b4de5-704d-41e4-b613-d56c87d00b81.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.zertifikate.commerzbank.de - site is not usable - <!-- @browser: Firefox Mobile 57.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.4.4; Mobile; rv:57.0) Gecko/57.0 Firefox/57.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.zertifikate.commerzbank.de/webforms/mobile/index.html?home/{"site":{"country":1,"language":1,"isoCode":"de-DE"}}
**Browser / Version**: Firefox Mobile 57.0
**Operating System**: Android 4.4.4
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: site not displayed correctly at all
**Steps to Reproduce**:
Direct link, please check with other mobile browsers.
[](https://webcompat.com/uploads/2017/8/925b4de5-704d-41e4-b613-d56c87d00b81.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_build
|
site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description site not displayed correctly at all steps to reproduce direct link please check with other mobile browsers from with ❤️
| 0
|
402,328
| 11,808,239,079
|
IssuesEvent
|
2020-03-19 13:02:31
|
deora-earth/rosebudDAO
|
https://api.github.com/repos/deora-earth/rosebudDAO
|
closed
|
The progress bar is fixed in the proposal list view but has bugs in the proposal detail view
|
03 High Priority bug
|
The progress bar should look the same like in the proposal list view.
This is the current look in the detailed view:

|
1.0
|
The progress bar is fixed in the proposal list view but has bugs in the proposal detail view - The progress bar should look the same like in the proposal list view.
This is the current look in the detailed view:

|
non_build
|
the progress bar is fixed in the proposal list view but has bugs in the proposal detail view the progress bar should look the same like in the proposal list view this is the current look in the detailed view
| 0
|
66,516
| 16,626,715,475
|
IssuesEvent
|
2021-06-03 10:29:57
|
daedaleanai/dbt
|
https://api.github.com/repos/daedaleanai/dbt
|
opened
|
Enforce flag registration
|
build enhancement
|
Currently registering flags is not enforced for all flags. Flags should no longer implicitly registered when calling `.Value()` on them. Instead only `.Register()` should register the flags. `.Register` should then also determine the flag's value and return that directly.
|
1.0
|
Enforce flag registration - Currently registering flags is not enforced for all flags. Flags should no longer implicitly registered when calling `.Value()` on them. Instead only `.Register()` should register the flags. `.Register` should then also determine the flag's value and return that directly.
|
build
|
enforce flag registration currently registering flags is not enforced for all flags flags should no longer implicitly registered when calling value on them instead only register should register the flags register should then also determine the flag s value and return that directly
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.