added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:56.754250
| 2020-06-14T09:03:24
|
638313133
|
{
"authors": [
"aravindr93",
"chinmay-singh"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3813",
"repo": "aravindr93/mjrl",
"url": "https://github.com/aravindr93/mjrl/issues/27"
}
|
gharchive/issue
|
Pickling of _VariableFunctions not compatible with PyTorch 1.5.0
pickling of _VariableFunctions not compatible with PyTorch 1.5.0
It would thus be useful if the version of PyTorch can be specified in the env.yml
Thanks for bringing this up. It seems like a recent change in pytorch is causing the break. Will pin to a specific version for now and will switch to torch.save in upcoming releases.
|
2025-04-01T06:37:56.763612
| 2020-10-19T15:40:29
|
724743221
|
{
"authors": [
"aprokop",
"wjge"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3814",
"repo": "arborx/ArborX",
"url": "https://github.com/arborx/ArborX/pull/414"
}
|
gharchive/pull-request
|
Add a ray class and a test for ray-box intersection
The ray class is based on Damien's original commit.
A function is added for ray-box intersections.
A question not blocking the PR: is there a reason for us to normalize direction in the first place? Certainly, not in the intersects test, but where would you think it would be helpful?
|
2025-04-01T06:37:56.765802
| 2022-09-09T12:43:27
|
1367776669
|
{
"authors": [
"Ikabodo",
"arbron"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3815",
"repo": "arbron/fvtt-summoner",
"url": "https://github.com/arbron/fvtt-summoner/issues/9"
}
|
gharchive/issue
|
[FEATURE] Dynamic damage on attacks
Is your feature request related to a problem? Please describe.
A lot of summons increases the damage to attacks when the spell is upcast. Currently it's not possible to increase the damage on attacks on an actor sheet through the summoning. (to my knowledge, but it might be possible with attributes?)
Describe the solution you'd like
An option to add damage to certain attacks based on level.
The summoning spell's roll data is made available to the summoned actor, so you can add @summoner.item.level into a damage formula to get the spell level that was used for the summoning. Check out the readme for more info.
|
2025-04-01T06:37:56.770178
| 2021-10-02T15:59:56
|
1014092099
|
{
"authors": [
"AnandDhakane01",
"KarishmaVanwari"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3816",
"repo": "arc-pune/arc-front-end",
"url": "https://github.com/arc-pune/arc-front-end/issues/29"
}
|
gharchive/issue
|
Bug in the navbar
Describe the bug
On clicking Adopt or About on the navbar, the bottom border does not shift (to the one that is clicked).
Hey, can I fix this as a part of Hacktoberfest 2021? @AnandDhakane01 @majjikishore007
Hey, can I fix this as a part of Hacktoberfest 2021? @AnandDhakane01 @majjikishore007
Hii, we are glad you are willing to contribute, we have actually decided to change the complete UI/UX of the site to a much better design and theme. You can check out the design here design. Let me know if you are willing to work on any of the unassigned pages.
Hey, can I fix this as a part of Hacktoberfest 2021? @AnandDhakane01 @majjikishore007
Hii, we are glad you are willing to contribute, we have actually decided to change the complete UI/UX of the site to a much better design and theme. You can check out the design here design. Let me know if you are willing to work on any of the unassigned pages.
Yup, sure!
We were thinking of assigning you the adopt us page. you can create the issue. Hop onto our discord server for a quick call. You will find the link in the readme file.
|
2025-04-01T06:37:56.775794
| 2024-07-02T18:58:23
|
2386936275
|
{
"authors": [
"dartpain",
"sarthaklaptop",
"siiddhantt"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3817",
"repo": "arc53/DocsGPT",
"url": "https://github.com/arc53/DocsGPT/pull/1025"
}
|
gharchive/pull-request
|
Solved: Query Bar Overflow with Long Text Input
What kind of change does this PR introduce? (Bug fix, feature, docs update, ...)
Query-bar-overflow-issue solved
Before
After
Why was this change needed? (You can also link to an open issue here)
While typing long sentences the last typed word OR while pasting the text the last sentences were not visible.
Other information:
Hello @siiddhantt made all the changes you mentioned please review !!!
hi @sarthaklaptop you don't need to make changes to the lint rules
just do --no-verify for now while you make the commit. And please resolve the conflicts there are currently
Hello @siiddhantt made changes you mentioned
Sorry for late update.
Hey @sarthaklaptop you will have a conflict on merge, please sync with the main branch,
Thank you!
@dartpain please check.
|
2025-04-01T06:37:56.828969
| 2022-08-23T10:31:00
|
1347706472
|
{
"authors": [
"New-arch",
"Phillipus",
"bratgers",
"evlibra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3818",
"repo": "archimatetool/archi",
"url": "https://github.com/archimatetool/archi/issues/864"
}
|
gharchive/issue
|
Labeling in HTML Reports View Documentation
Feature Requirement:
As HTML report provides us with HTML / Markdown Render , it provides a great way to compliment the viewpoint with content.
However the HTML render from the view Documentation tag does not include any metadata from the model. Therefore there is no reuse of the model artefacts or the view properties within the html content.
Feature requirement is to use view's properties in the html setup similar to what we use in labels on the objects presentation.
A stretch target will be to use an object's properties in the view's render.
Architecture Decision Record Card
ID :${property:ADR_ID}
CRITICALITY : ${property:viewProperty}
DATE : 2022/08/20
STATUS : Green
Version of Archi
Archi Plug-ins
Operating System
Expected Behaviour
Actual Behaviour
All of the code for Label Expressions is designed only to render in the objects' text in a View, or nodes in the Models Tree. I think trying to implement this in the Documentation field so that the expressions rendered in an HTML report would require some serious effort. Perhaps something to think about one day.
Yes . Added a feature requirement to consider for the backlog. We are using the html render more and more to augment the diagrams with content , rather than just diagrams and go to other design tools for content. This feature will improve the usability of the html report as a more hoisting solution offering where we can bring in lists and matrixes to the repository managed by arch file with reusing the models artefacts in properties possibly. This extends the view's usage to explain something to a specific user community .
It turns out that an initial implementation of rendering label expressions in documentation fields in HTML and Jasper Reports was not so complicated:
https://github.com/archimatetool/archi/commit/f987970fbd0a4e9f221e053bad672f7a96665fa1
However:
Is this something that other people want?
Are there side-effects?
What is the impact of this change?
We need to think about this.
I use the html report allot and content to compliment the solution architecture viewpoints instead of running with separate pptx for stakeholders that need more information or different content mediums. Other architects recommended we move to C4Builder that gives you the combination of markdown files and plantuml for diagraming within one solution. I would rather stay with Archi.
It turns out that an initial implementation of rendering label expressions in documentation fields in HTML and Jasper Reports was not so complicated:
f987970
However:
* Is this something that other people want?
* Are there side-effects?
* What is the impact of this change?
We need feedback on this.
For my new assignment I used Archi again, after several years. I like the simplicity and the power. of the tool I think it could be a great advantage if we could harmonize and unify the documentation and properties parts with a Markdown render, for all the output and report forms.
For my previous assignment it was necessary to use BizzDesign and in my opinion BizzDesign has three advances over Archi:
The ability to format text.
The ability to use references (hyperlinks) to elements in the documentation field.
The ability to use vertical oriented text labels.
I use the HTML and Jasper reporting tools and jArchi scripts to produce out in the same look and feel as the documentation of the customer. Te lack of formatting text is annoying. As a first step, I think it it is a big plus if the documentation fields support Markdown. As a second step maybe it is possible to support internal linking to elements and maybe in the future vertical orientation for text labels.
There is probably one issue, Jasper Report support Markup formatting, and own implementation of of text fromating, beside html and rtf. For the use of Markdown in Jasper Report the Markdown needs to be converted to html , rtf or the native Markkup format.
See: https://jasperreports.sourceforge.net/sample.reference/markup/index.html#markup
It turns out that an initial implementation of rendering label expressions in documentation fields in HTML and Jasper Reports was not so complicated:
f987970
However:
* Is this something that other people want?
* Are there side-effects?
* What is the impact of this change?
We need feedback on this.
For my new assignment I used Archi again, after several years. I like the simplicity and the power. of the tool I think it could be a great advantage if we could harmonize and unify the documentation and properties parts with a Markdown render, for all the output and report forms.
For my previous assignment it was necessary to use BizzDesign and in my opinion BizzDesign has three advances over Archi:
The ability to format text.
The ability to use references (hyperlinks) to elements in the documentation field.
The ability to use vertical oriented text labels.
I use the HTML and Jasper reporting tools and jArchi scripts to produce output with the same look and feel as the documentation of the customer. The lack of formatting text is annoying. As a first step, I think it it is a big plus if the documentation fields support Markdown.
As a second step maybe it is possible to support internal linking to elements and maybe in the future vertical orientation for text labels.
There is probably one issue, Jasper Report support Markup formatting (the option styles). Markup is an own implementation of text formatting, beside html and rtf. For the use of Markdown in Jasper Report the Markdown needs to be converted to html , rtf or the native Markkup format. See: https://jasperreports.sourceforge.net/sample.reference/markup/index.html#markup
What about usage of reStructuredText for the report? See https://docutils.sourceforge.io/rst.html
Advantage would be ability to edit pages independently, even directly on the web, e.g. at GitHub. For that also would be good to include element name in the page name at the beginning of name (in addition to the id and element type, or instead of the element type)
|
2025-04-01T06:37:56.886807
| 2020-09-25T05:03:32
|
708642922
|
{
"authors": [
"KimJeonghun91",
"jsamr"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3819",
"repo": "archriss/react-native-render-html",
"url": "https://github.com/archriss/react-native-render-html/issues/412"
}
|
gharchive/issue
|
imagesMaxWidth not working
It does not work after changing the RN version from 0.60.3 to 0.63.2.
"react": "16.13.1",
"react-native": "0.63.2",
"react-native-render-html": "^4.2.3",
"react-native-webview": "^10.9.0"
RN
<HTML
html={detailData.content}
imagesMaxWidth={Dimensions.get('window').width}
imagesInitialDimensions={{ width: Dimensions.get('window').width, height: Dimensions.get('window').width }}
tagsStyles={{
p: { fontFamily: Layout.fsFontNsR, marginTop: 5, lineHeight: 20 },
div: { fontFamily: Layout.fsFontNsR, marginTop: 5, lineHeight: 20 },
img: { width: Dimensions.get('window').width }
}}
/>
HTML
<img id="se_object_1598178785239" src="https://blogfiles.pstatic.net/MjAyMDA4MjNfMjIy/MDAxNTk4MTc4Mzc0NzY3.lP4y6IgYLhsLj98qYypYb1TJCzHGHAArTBsjcafM5ckg.SyBbjTSKLfVT_Tve1bW55GsCgn4NZA74_ZqhxOPWPDYg.JPEG.0610studio/6-%ED%94%BD_01.jpg"
class="__se_object" s_type="attachment" s_subtype="photo"
style=" width:1080px; height:1084px; rwidth:1080px; rheight:1084px;" width="1080" height="1084" imgqe="true"
jsonvalue="%7B%7D" rwidth="1080px" rheight="1084px">
<img id="se_object_1598178852473" src="https://blogfiles.pstatic.net/MjAyMDA4MjNfMTE2/MDAxNTk4MTc4Mzc0OTkz.aTquUUjfOjIJHOIOQOmKl8EM5J2V8iks6Xjg2ngs7wEg.-qG--Ba8CK8u6LHxmUtn4ApJAyuZUiCykGUta2VRzykg.JPEG.0610studio/6-%ED%94%BD_02.jpg"
class="__se_object" s_type="attachment" s_subtype="photo"
style=" width:1080px; height:1076px; rwidth:1080px; rheight:1076px;" width="1080" height="1076" imgqe="true"
jsonvalue="%7B%7D" rwidth="1080px" rheight="1076px">
@KimJeonghun91 Are you sure it's the RN version changes which cause a different behavior? We have proposed a new RFC for image and moreover media width here. There is a pending PR #389 ready to ship for v5.x. I could merge that to master and publish an alpha for you to test this. Are you interested?
I am interested.
Would you please merge it and publish the alpha for me?
I will test it on my project.
@KimJeonghun91 See https://github.com/archriss/react-native-render-html/releases/tag/v5.0.0-alpha.1
@KimJeonghun91 I tried your code with the latest (alpha.2) and it works!
import * as React from 'react';
import {ScrollView, StyleSheet, useWindowDimensions} from 'react-native';
import HTML from 'react-native-render-html';
const html = `
<img id="se_object_1598178785239" src="https://blogfiles.pstatic.net/MjAyMDA4MjNfMjIy/MDAxNTk4MTc4Mzc0NzY3.lP4y6IgYLhsLj98qYypYb1TJCzHGHAArTBsjcafM5ckg.SyBbjTSKLfVT_Tve1bW55GsCgn4NZA74_ZqhxOPWPDYg.JPEG.0610studio/6-%ED%94%BD_01.jpg"
class="__se_object" s_type="attachment" s_subtype="photo"
style=" width:1080px; height:1084px; rwidth:1080px; rheight:1084px;" width="1080" height="1084" imgqe="true"
jsonvalue="%7B%7D" rwidth="1080px" rheight="1084px">
<img id="se_object_1598178852473" src="https://blogfiles.pstatic.net/MjAyMDA4MjNfMTE2/MDAxNTk4MTc4Mzc0OTkz.aTquUUjfOjIJHOIOQOmKl8EM5J2V8iks6Xjg2ngs7wEg.-qG--Ba8CK8u6LHxmUtn4ApJAyuZUiCykGUta2VRzykg.JPEG.0610studio/6-%ED%94%BD_02.jpg"
class="__se_object" s_type="attachment" s_subtype="photo"
style=" width:1080px; height:1076px; rwidth:1080px; rheight:1076px;" width="1080" height="1076" imgqe="true"
jsonvalue="%7B%7D" rwidth="1080px" rheight="1076px">
`;
export default function App() {
const {width} = useWindowDimensions();
return (
<ScrollView contentContainerStyle={styles.container}>
<HTML contentWidth={width} html={html} />
</ScrollView>
);
}
const styles = StyleSheet.create({
container: {
flexGrow: 1,
},
});
Notice that you don't need to set explicit width to your images anymore. The crucial step is to provide contentWidth prop. If you want to have images max width computed from content width, you'll need to use computeImagesMaxWidth prop. A great addition with using useWindowDimensions is that the image will handle screen rotations gracefully.
I'm closing now, but feel free to respond and / or add feedback.
|
2025-04-01T06:37:56.913367
| 2018-06-26T05:18:38
|
335662725
|
{
"authors": [
"Marcello168",
"StephentTom",
"anoopmm",
"bd-arc",
"dugginenisagar",
"haduy97",
"iyoucheng",
"shamca0633",
"shashank19909"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3820",
"repo": "archriss/react-native-snap-carousel",
"url": "https://github.com/archriss/react-native-snap-carousel/issues/342"
}
|
gharchive/issue
|
Clicking/Tapping image or Navigating leading to stop auto play
Is this a bug report or a feature request?
Bug report ( or need help , not sure if this is achievable through existing code )
Have you read the guidelines regarding bug report?
Yes
Have you read the documentation in its entirety?
Yes
Have you made sure that your issue hasn't already been reported/solved?
Yes.(Looked into old issues, didn't find anything with related keywords.)
Is the bug specific to iOS or Android? Or can it be reproduced on both platforms?
Yes, both platform.
Is the bug reproductible in a production environment (not a debug one)?
YES
Have you made sure that it wasn't a React Native bug?
Doesn't seems like to be React native bug.
<!-
Have you been able to reproduce the bug in the provided example?
Yes.(With slight modification, but should happen in example1 code as well)
Environment
Environment:
React Native - 0.50,
React - 16
react-native-snap-carousel: 3.6/3.7
Target Platform:
Android (8.1)
I have tried https://snack.expo.io/rJ1XQrJM7 and able to reproduce this issue. ( IN both IOS/Android)
Steps to Reproduce
Execute the example1(loop one ) from https://github.com/archriss/react-native-snap-carousel/blob/master/example/src/index.js or https://snack.expo.io/rJ1XQrJM7
2.Click on image, auto play stops(Same is happening in case of navigation (with TouchableOpacity) as well).
Expected Behavior
Auto play should continue after user clicks on image(or navigate).
Note : Auto play get restored when user swipes screen manually to next slides, but slides should have moved to next slides automatically without manual intervention.
Actual Behavior
Clicking/tapping on image(or navigating through TouchableOpacity) stops autoplay.
Reproducible Demo
https://snack.expo.io/rJ1XQrJM7
Additional questions:
If we swipe from one slide to another we can see white color margin b/w 2 images , Is it possible to minimize the margin(may be 0). Below one suggests to add margin,I have made that 0 but still see margin while slides to next image.
https://github.com/archriss/react-native-snap-carousel/blob/master/doc/TIPS_AND_TRICKS.md#margin-between-slides.
2.Also, currently we have req. to show image of size =( width = screenWidth, height = screenWidth/2), if I understand correctly https://github.com/archriss/react-native-snap-carousel/blob/master/doc/TIPS_AND_TRICKS.md#viewport-wide-slides--no-preview-effect
should we explore https://github.com/leecade/react-native-swiper#basic-usage as well?
I think , if we get solution to auto play issue then I am happy to use this carousel for our use case(as we might be interested in preview feature in future)
Thanks.
Hi @shamca0633 ,
This is not relevant to your problem, but I have a question in the above demo code you provided.
I am trying to call a function when I click on an Item, onPress of Touchable if I call any method it is giving an error saying not a function.
<TouchableOpacity onPress={this.onPress.bind(this)} >
Error is _this6.onPress is not a function.
I am able to directly alert on Touch like below
<TouchableWithoutFeedback onPress={()=>{ Alert.alert("onPress");}}>
How can I able to call a method onPress of any Item?
Hi @shamca0633,
Regarding your issue, this is a known bug that I haven't addressed yet. I'll take a look at it as soon as possible. FYI, the root of the matter lies here.
To answer your questions:
The "margin" is linked to the inactiveSlideScale prop. You can set it to 1 if needed, but you'll then lose the scale effect.
Yes, I confirm what I've written in the doc.
Thanks for you quick response.
I wanted to check few more things which is blocking us to move forward on this plug-in.
I am observing a delay in initial loading of images. We have small set of images(max 10) only still facing this issue. Also to compare we tested it with some simple code (without auto scroll/loop,etc) and can see images getting load faster.Is it this plug-in adding some delay due to additional feature, if yes what are the recommendation to minimize initial load time?it has anything to do with loopClonesPerSide , is it loading 3 copies and adding to delay?
https://medium.com/the-react-native-log/custom-scrolling-carousel-in-react-native-15ee129e7e68
2.Once images are initially loaded seems like it scrolls a bit(and appears like flicker). Is it known issue?
3.At certain slide( 4th one out of 6 , with SLIDER_1_FIRST_ITEM =1 and loopClonesPerSide={3}) always flicker and in every loop, rest is fine.Not able to get logic on why 4th one.
4.Also, when I swipe fast then it reaches to last slides and refuse to swipe further in same direction, but able to swipe in reverse direction. Also once auto play happens then user can further swipe to next slide.Is there any end of loop from where user can't swipe further ? Is there any parameters there to handle this?
Configuration:
<View>
<Carousel
ref={c => this._slider1Ref = c}
data={ENTRIES1}
renderItem={this._renderItem}
sliderWidth={sliderWidth}
itemWidth={itemWidth}
firstItem={SLIDER_1_FIRST_ITEM}
inactiveSlideScale={1}
activeSlideAlignment={'start'}
inactiveSlideOpacity={1}
inactiveSlideShift={0.6}
containerCustomStyle={styles.slider}
contentContainerCustomStyle={styles.sliderContentContainer}
loop={true}
loopClonesPerSide={3}
autoplay={true}
lockScrollWhileSnapping={true}
enableMomentum ={true}
autoplayDelay={3000}
autoplayInterval={2000}
onSnapToItem={(index) => this.setState({ slider1ActiveSlide: index})
}
/>
<Pagination
dotsLength={ENTRIES1.length}
activeDotIndex={slider1ActiveSlide}
containerStyle={styles.paginationContainer}
dotColor={'rgba(255, 255, 255, 0.92)'}
dotStyle={styles.paginationDot}
inactiveDotColor={colors.black}
inactiveDotOpacity={0.6}
inactiveDotScale={0.60}
carouselRef={this._slider1Ref}
tappableDots={!!this._slider1Ref}
/>
</View>
Note : I have tested above on Android Simulator (8.1) and on Android real device.
I have gone through some existing performance/flickers issues but was not able to relate it with my case( or may be you can point me to , if I missed)
Thanks.
Thanks, setting "inactiveSlideScale" to 1 solved the margin b/w slides issue.
@bd-arc any estimate to when can we excpect this fix?
Also thank you very much for this wonderful library
same issue
are there any updates on this?
same issue
same issue
,how to solve?
Fixed in 3.8.0.
Thanks for you quick response.
I wanted to check few more things which is blocking us to move forward on this plug-in.
I am observing a delay in initial loading of images. We have small set of images(max 10) only still facing this issue. Also to compare we tested it with some simple code (without auto scroll/loop,etc) and can see images getting load faster.Is it this plug-in adding some delay due to additional feature, if yes what are the recommendation to minimize initial load time?it has anything to do with loopClonesPerSide , is it loading 3 copies and adding to delay?
https://medium.com/the-react-native-log/custom-scrolling-carousel-in-react-native-15ee129e7e68
2.Once images are initially loaded seems like it scrolls a bit(and appears like flicker). Is it known issue?
3.At certain slide( 4th one out of 6 , with SLIDER_1_FIRST_ITEM =1 and loopClonesPerSide={3}) always flicker and in every loop, rest is fine.Not able to get logic on why 4th one.
4.Also, when I swipe fast then it reaches to last slides and refuse to swipe further in same direction, but able to swipe in reverse direction. Also once auto play happens then user can further swipe to next slide.Is there any end of loop from where user can't swipe further ? Is there any parameters there to handle this?
Configuration:
<View>
<Carousel
ref={c => this._slider1Ref = c}
data={ENTRIES1}
renderItem={this._renderItem}
sliderWidth={sliderWidth}
itemWidth={itemWidth}
firstItem={SLIDER_1_FIRST_ITEM}
inactiveSlideScale={1}
activeSlideAlignment={'start'}
inactiveSlideOpacity={1}
inactiveSlideShift={0.6}
containerCustomStyle={styles.slider}
contentContainerCustomStyle={styles.sliderContentContainer}
loop={true}
loopClonesPerSide={3}
autoplay={true}
lockScrollWhileSnapping={true}
enableMomentum ={true}
autoplayDelay={3000}
autoplayInterval={2000}
onSnapToItem={(index) => this.setState({ slider1ActiveSlide: index})
}
/>
<Pagination
dotsLength={ENTRIES1.length}
activeDotIndex={slider1ActiveSlide}
containerStyle={styles.paginationContainer}
dotColor={'rgba(255, 255, 255, 0.92)'}
dotStyle={styles.paginationDot}
inactiveDotColor={colors.black}
inactiveDotOpacity={0.6}
inactiveDotScale={0.60}
carouselRef={this._slider1Ref}
tappableDots={!!this._slider1Ref}
/>
</View>
Note : I have tested above on Android Simulator (8.1) and on Android real device.
I have gone through some existing performance/flickers issues but was not able to relate it with my case( or may be you can point me to , if I missed)
Thanks.
Thank you man, you just save my day.
This props:
inactiveSlideScale={1}
inactiveSlideOpacity={1}
inactiveSlideShift={0.6}
Solved the flicker/blink issue when reached to last item and jump into the first item without blinking
|
2025-04-01T06:37:56.929815
| 2024-03-02T14:44:31
|
2164839806
|
{
"authors": [
"laravel-shift"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3821",
"repo": "archtechx/laravel-pages",
"url": "https://github.com/archtechx/laravel-pages/pull/8"
}
|
gharchive/pull-request
|
Laravel 11.x Compatibility
This is an automated pull request from Shift to update your package code and dependencies to be compatible with Laravel 11.x.
Before merging, you need to:
Checkout the l11-compatibility branch
Review all comments for additional changes
Thoroughly test your package
If you do find an issue, please report it by commenting on this PR to help improve future automation.
:warning: Shift detected GitHub Actions which run jobs using a version matrix. Shift attempted to update your configuration for Laravel 11. However, you should review these changes to ensure the desired combination of versions are built for your package.
:alembic: Using this package? If you would like to help test these changes or believe them to be compatible, you may update your project to reference this branch.
To do so, temporarily add Shift's fork to the repositories property of your composer.json:
{
"repositories": [
{
"type": "vcs",
"url": "https://github.com/laravel-shift/laravel-pages.git"
}
]
}
Then update your dependency constraint to reference this branch:
{
"require": {
"archtechx/laravel-pages": "dev-l11-compatibility",
}
}
Finally, run: composer update
|
2025-04-01T06:37:56.974792
| 2020-08-28T17:24:50
|
688235742
|
{
"authors": [
"ardalis",
"fiseni",
"ovation22"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3822",
"repo": "ardalis/Specification",
"url": "https://github.com/ardalis/Specification/issues/47"
}
|
gharchive/issue
|
Expose Top or Take
Looking to select example: 4 items from my specification. I've resorted to using Query.Paginate(0, 4).
@ardalis suggested raising an Issue here to see if it might be a trivial item to implement.
Hey @ovation22 ,
If I understood correctly, you want to expose Take in the specification? You want to use Query.Take(4), instead of Query.Paginate(0, 4) ?
What would be the benefit actually? Not writing the 0?
Implementation wise, not hard to add. But, we have to ensure that Paginate() and Take() won't be used within the same specification, since it will overwrite the values. And that might add to the confusion.
It's a good point about the two being exclusive of one another. @ovation22 and I were discussing this Friday and my first thought was that it would be more intuitive to offer a .Take directly rather than having to "page" to do it. Conceptually to me at least, I don't think about paging as a way to get "Top X Rows/Records". Under the covers I know paging uses Skip/Take but it still seems like a bit of a hack to have to go through the Paging abstraction just to take advantage of this fact.
It shouldn't be too hard for us to support Take or Paginate, and throw if both are used, right?
Not a problem to implement. I was just thinking how to provide a consistent API. Not a huge fan of overlapping functionalities :)
If we think twice, Paginate is just a dummy wrapper and made sense in the legacy API, but now that we have LINQ alike structure, we may be more verbose and expose Top and Skip directly. For now we can support both ways, and decorate the Paginate as obsolete. In the future we'll remove it completely. Makes sense?
But, I have a feeling we misunderstood @ovation22. He might want to get top records regardless of the specification (for any spec). If this is a case, this can be done in your repository implementation.
We provide RepositoryBase generic repository ready to use. It's an abstract one, so anyway you will have to derive from it. You could do something like this in your repo.
public interface IRepository<T> : IRepositoryBase<T> where T : class
{
Task<List<T>> GetTopAsync(ISpecification<T> specification, int take);
Task<List<TResult>> GetTopAsync<TResult>(ISpecification<T, TResult> specification, int take);
}
public class Repository<T> : RepositoryBase<T>, IRepository<T> where T : class
{
protected readonly TestDbContext dbContext;
public Repository(TestDbContext dbContext) : base(dbContext)
{
this.dbContext = dbContext;
}
public async Task<List<T>> GetTopAsync(ISpecification<T> specification, int take)
{
return await ApplySpecification(specification).Take(take).ToListAsync();
}
public async Task<List<TResult>> GetTopAsync<TResult>(ISpecification<T, TResult> specification, int take)
{
return await ApplySpecification(specification).Take(take).ToListAsync();
}
}
PS. I just noticed ApplySpecification is a private method. We'll make that protected, no harm of exposing it.
I was specifically referencing using Top/Take in a Specification, but I appreciate the option with the IRepository.
Giving it more thought over the weekend I was thinking along those lines same lines as you mention @fiseni if it would make sense to expose Take and Skip and deprecating Paginate. Makes sense to me.
One of the key benefits of the Specification pattern IMO is that it helps prevent the need for repository implementations to have tons of different methods. If we can avoid adding more methods to repository by allowing fewer methods to do the same thing using a specification, I think we should do that. I also don't want to require devs to have to use our base repository just because they want to use specifications (as that will hurt adoption).
So, I think your original idea works:
Not a problem to implement. I was just thinking how to provide a consistent API. Not a huge fan of overlapping functionalities :)
If we think twice, Paginate is just a dummy wrapper and made sense in the legacy API, but now that we have LINQ alike structure, we may be more verbose and expose Top and Skip directly. For now we can support both ways, and decorate the Paginate as obsolete. In the future we'll remove it completely. Makes sense?
I'm still not sure I want to drop Paginate, since I like the naming of it (it's a higher level abstraction that covers up the low level skip/take details) but I agree it's not great design to have 2 ways to do the same thing.
Giving it more thought over the weekend I was thinking along those lines same lines as you mention @fiseni if it would make sense to expose Take and Skip and deprecating Paginate. Makes sense to me.
Sounds like we all agree. Add Take/Skip to specification and mark Paginate obsolete.
I do share your opinion regarding the repositories. I didn't mean to provide the functionality through the repo, or enforce it in any way. It's just an example how users can implement such thing in their code base, in their repo (cos that's the only way you can get such general Top functionality).
I'm free for the next 30 mins, I'll do the changes now.
This is implemented. If no one has additional comments, we may close the issue.
Ok so all I need to do if it looks good is publish a new point release, yes?
Yes, I think so.
I don't think we have some breaking changes, but you might want to go through the history and cross check. There is a comment on the last commit too.
Other than that, tests should be refactored. It's in my backlog too, if I find time in the next period, I'll try to go through that.
|
2025-04-01T06:37:56.978039
| 2022-11-16T13:59:25
|
1451659416
|
{
"authors": [
"jspaaks"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3823",
"repo": "ardc-fair-checklist/ssg",
"url": "https://github.com/ardc-fair-checklist/ssg/issues/25"
}
|
gharchive/issue
|
index pages have hardcoded version
consider getting the latest version from the version store
e.g. https://github.com/ardc-fair-checklist/ssg/blob/4f7d5cf8e056c538d92d351bd7a627b1596edff4/pages/software/v0.2/index.page.vue#L79
partly fixed by fda6670, still some references in BadgeSnippets:
https://github.com/ardc-fair-checklist/ssg/blob/fda66701d35e54ba2b08e5b80820224be1c5bccb/pages/software/v0.2/BadgeSnippets.vue#L41-L50
no more hardcoded version strings in code base, reactivity problems (#36) seem to have been solved by 49dee0d
|
2025-04-01T06:37:57.071315
| 2022-06-04T00:14:07
|
1260539884
|
{
"authors": [
"jaggzh"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3824",
"repo": "arduino/library-registry",
"url": "https://github.com/arduino/library-registry/pull/1464"
}
|
gharchive/pull-request
|
Update repositories.txt
Added mini-ppm-info (2nd attempt)
I have a library.properties file already in the repository.
https://github.com/jaggzh/mini-ppm-info/blob/main/library.properties
@ArduinoBot, I incremented the version with some improvements.
|
2025-04-01T06:37:57.078654
| 2019-08-14T15:21:43
|
480741792
|
{
"authors": [
"CLAassistant",
"nikhiltri"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3826",
"repo": "area17/twill",
"url": "https://github.com/area17/twill/pull/346"
}
|
gharchive/pull-request
|
Add noindex and nofollow header to CMS pages
In instances where the same codebase serves the frontend as well
as the CMS, it's challenging to use a robots.txt file to block
search engine spiders from indexing admin pages. Adding this
header help to ensure that CMS pages done find their way into
search engines.
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T06:37:57.085834
| 2017-10-09T22:26:53
|
264044376
|
{
"authors": [
"broskoski",
"dzucconi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3827",
"repo": "aredotna/case",
"url": "https://github.com/aredotna/case/pull/198"
}
|
gharchive/pull-request
|
Add title to comments screen
Closes #148
I didn't do this because adding comments feels really weird. Let's just merge it and deal with it later. I was hoping to do it with a count state, but really I think we need to just dispatch a title update (we can refactor the goofy title visibility toggle to do this)
By feels really weird I mean you expect the title to update when you add one, and it doesn't.
Ah right, good call. I'm fine to close this, was just trying to get some easy issues closed. Your call.
Lol
|
2025-04-01T06:37:57.089980
| 2024-05-13T14:53:19
|
2293069240
|
{
"authors": [
"MrHalless",
"tigrulya-exe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3828",
"repo": "arenadata/SSM",
"url": "https://github.com/arenadata/SSM/pull/38"
}
|
gharchive/pull-request
|
ADH-4390: [UI] FormField
https://tracker.yandex.ru/ADH-4390
Please, rebase on the master and run the following command to add license headers to the new source files. Otherwise, subsequent CI builds will fail.
|
2025-04-01T06:37:57.102918
| 2021-03-09T09:55:53
|
825643821
|
{
"authors": [
"jgwest",
"rumstead",
"sachincool"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3829",
"repo": "argoproj-labs/applicationset",
"url": "https://github.com/argoproj-labs/applicationset/issues/149"
}
|
gharchive/issue
|
non-existent docker image referenced in manifest
Hi!
I tried to install the applicationset CRD into my cluster and found that the docker images for controller deployment doesn't exist.
manifests
docker pull argoprojlabs/
argocd-applicationset:latest
Error response from daemon: pull access denied for argoprojlabs/argocd-applicationset, repository does not exist or may require 'docker login': denied: requested access to the resource is denied
can you point me to the correct image if exists? makefile also points to this image.
@jgwest, why not publish a "latest" image on every commit to master?
@rumstead Working on it :smile:.
|
2025-04-01T06:37:57.109457
| 2022-07-25T19:59:20
|
1317301952
|
{
"authors": [
"egkelly"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3830",
"repo": "argoproj-labs/argocd-image-updater",
"url": "https://github.com/argoproj-labs/argocd-image-updater/issues/461"
}
|
gharchive/issue
|
Helm Image is Not Updated
Describe the bug
I'm not sure if this is a legitimate bug or if I'm missing a key piece of documentation/understanding, but I have a helm chart successfully deploying to my cluster via ArgoCD, and I'm trying to use the image updater to keep my image up to date based on the latest update strategy using git write-back and manual sync. When a new image is published to my repository, image-updater successfully updates my .argo-source-cmd-webapp.yaml file with the image tag, but on the ArgoCD end, it doesn't recognize that the image tag I have defined in my values.yaml file is different from the image tag in the argo source file and therefore does not update to my latest image update.
To Reproduce
My relevant charts:
deployment.yaml
kind: Deployment
metadata:
name: {{ include "cmd-webapp.fullname" . }}
labels:
{{- include "cmd-webapp.labels" . | nindent 4 }}
{{- with .Values.webappDeploymentAnnotations }}
annotations:
{{- toYaml . | nindent 4 }}
{{- end }}
namespace: {{ .Values.webapp.namespace }}
...
...
...
template:
...
...
spec:
containers:
- name: {{ .Chart.Name }}
image: {{ .Values.image.repository }}:{{ .Values.image.tag | default "latest" }}
imagePullPolicy: {{ .Values.image.pullPolicy }}
...
...
values.yaml
image:
repository: XXXXXXXX.amazonaws.com/webapp
pullPolicy: Always
tag: 42feaa43
...
...
.argocd-source-cmd-webapp.yaml
helm:
parameters:
- name: image.name
value: XXXXXXXX.amazonaws.com/webapp
forcestring: true
- name: image.tag
value: 0ca9d43f
forcestring: true
Application annotations
argocd-image-updater.argoproj.io/force-update: true
argocd-image-updater.argoproj.io/helm.image-name: image.repository
argocd-image-updater.argoproj.io/helm.image-tag: image.tag
argocd-image-updater.argoproj.io/ignore-tags: latest
argocd-image-updater.argoproj.io/image-list: XXXXXXXXX.amazonaws.com/webapp
argocd-image-updater.argoproj.io/update-strategy: latest
argocd-image-updater.argoproj.io/write-back-method: git
Expected behavior
I would think ArgoCD would recognize the change to the helm parameters and show the application is Out of Sync, but it doesn't register a change
Version
v0.12.0+aee153d
I did a little messing around and figured out that while argocd-image-updaer was writing to argocd-source-cmd-webapp.yaml, the helm override parameter will only kick in when the file is named argocd-source.yaml. I'm still not sure where it's configured that the git writeback writes to a file called argocd-source-cmd-webapp.yaml and continues to do so even though I've renamed the file to argocd-source.yaml
Notice argocd is writing only to argocd-source-cmd-webapp.yaml instead of argocd-source.yaml
I think this issue belongs in the argo-cd repository as argocd-image-updater seems to be doing it's part just fine. Closing this issue and opening adjacent issue in argo-cd: https://github.com/argoproj/argo-cd/issues/10177
|
2025-04-01T06:37:57.164908
| 2022-02-11T15:37:43
|
1132766221
|
{
"authors": [
"blkperl",
"cnu1812",
"terrytangyuan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3831",
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/7857"
}
|
gharchive/issue
|
I would like a mentor
Summary
What change needs making?
Use Cases
When would you use this?
Message from the maintainers:
Love this enhancement proposal? Give it a 👍. We prioritise the proposals with the most 👍.
I would like to work on #7814 and #7815
@cnu1812 the first issue is closed and the second one already has an interested person on the issue. Is there another issue you are interested in? Let me know if you want have a call or slack conversation to help get you started on an issue.
Have you read the mentoring guide? Please only submit this request for general mentoring (not GSoC).
Answer: Yes.
What is your background? Any experience with Go, Kubernetes, React, Typescript, etc.?
Answer: I have a bit experience with kubernetes and go.
Is there any particular issue you'd like to work on? You may want to check out the list of good first issues.
Answer:
Would like to work on #6924
@cnu1812 Sounds good, let me know if you would like to talk about it or look at a draft PR.
I added a comment in https://github.com/argoproj/argo-workflows/issues/6924#issuecomment-1048893567 since GitHub does not allow me to assign to you.
@cnu1812 Are you still working on that issue? If not, I will close this and unassign that issue from you.
I am working on this. give me some time
|
2025-04-01T06:37:57.182866
| 2022-10-10T01:12:53
|
1402467618
|
{
"authors": [
"alexec",
"sanqiuli",
"sarabala1979"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3832",
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/9780"
}
|
gharchive/issue
|
pod deleted error when running multiple workflows concurrently
Pre-requisites
[X] I have double-checked my configuration
[X] I can confirm the issues exists when I tested with :latest
[ ] I'd like to contribute the fix myself (see contributing guide)
What happened/what you expected to happen?
I wrote a workflow about biological data processing. This workflow may require a lot of cpu and memory during the running process, but I did not specify the amount of resources required by the template in the workflow. This workflow I run individually is It is 100% successful, but the pod deleted error will appear when running more than one at the same time, but because the pod has been deleted, I cannot see the logs of the init, wait, and main containers in the pod.
Version
v3.4.1
Paste a small workflow that reproduces the issue. We must be able to run the workflow; don't enter a workflows that uses private images.
none
Logs from the workflow controller
kubectl logs -n argo deploy/workflow-controller | grep sky2-pipeline-jwtf8.feature-classify.feature-classify-1
time="2022-10-09T12:05:06.044Z" level=info msg="All of node sky2-pipeline-jwtf8.feature-classify.feature-classify-1 dependencies [] completed" namespace=argo workflow=sky2-pipeline-jwtf8
time="2022-10-09T12:05:06.054Z" level=info msg="Created pod: sky2-pipeline-jwtf8.feature-classify.feature-classify-1 (sky2-pipeline-jwtf8-qiime2-picrust2-2951325495)" namespace=argo workflow=sky2-pipeline-jwtf8
time="2022-10-09T12:08:32.500Z" level=info msg="Workflow pod is missing" namespace=argo nodeName=sky2-pipeline-jwtf8.feature-classify.feature-classify-1 nodePhase=Running recentlyStarted=false workflow=sky2-pipeline-jwtf8
Logs from in your workflow's wait container
none
I think pod deletion because of node auto scale or node rotation. You can prevent this with two options.
Add retry strategy for every step
Enable the PDB on workflow
Sent from my iPhone
On Oct 9, 2022, at 6:13 PM, sanqiuli @.***> wrote:
Pre-requisites
I have double-checked my configuration
I can confirm the issues exists when I tested with :latest
I'd like to contribute the fix myself (see contributing guide)
What happened/what you expected to happen?
I wrote a workflow about biological data processing. This workflow may require a lot of cpu and memory during the running process, but I did not specify the amount of resources required by the template in the workflow. This workflow I run individually is It is 100% successful, but the pod deleted error will appear when running more than one at the same time, but because the pod has been deleted, I cannot see the logs of the init, wait, and main containers in the pod.
Version
v3.4.1
Paste a small workflow that reproduces the issue. We must be able to run the workflow; don't enter a workflows that uses private images.
none
Logs from the workflow controller
kubectl logs -n argo deploy/workflow-controller | grep sky2-pipeline-jwtf8.feature-classify.feature-classify-1
time="2022-10-09T12:05:06.044Z" level=info msg="All of node sky2-pipeline-jwtf8.feature-classify.feature-classify-1 dependencies [] completed" namespace=argo workflow=sky2-pipeline-jwtf8
time="2022-10-09T12:05:06.054Z" level=info msg="Created pod: sky2-pipeline-jwtf8.feature-classify.feature-classify-1 (sky2-pipeline-jwtf8-qiime2-picrust2-2951325495)" namespace=argo workflow=sky2-pipeline-jwtf8
time="2022-10-09T12:08:32.500Z" level=info msg="Workflow pod is missing" namespace=argo nodeName=sky2-pipeline-jwtf8.feature-classify.feature-classify-1 nodePhase=Running recentlyStarted=false workflow=sky2-pipeline-jwtf8
Logs from in your workflow's wait container
none
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you are subscribed to this thread.
What is PDB? There are many steps in the workflow, and Add retry strategy for every step is very troublesome.
https://github.com/argoproj/argo-workflows/blob/master/examples/default-pdb-support.yaml
Sent from my iPhone
On Oct 9, 2022, at 6:36 PM, sanqiuli @.***> wrote:
What is PDB? There are many steps in the workflow, and Add retry strategy for every step is very troublesome.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you commented.
https://github.com/argoproj/argo-workflows/blob/master/examples/default-pdb-support.yaml
…
Sent from my iPhone
On Oct 9, 2022, at 6:36 PM, sanqiuli @.***> wrote: What is PDB? There are many steps in the workflow, and Add retry strategy for every step is very troublesome. — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.
it didn't work
@sanqiuli can you provide the pod describe? Can you gett more details why the POD is getting deleted in kubernetes?
@sanqiuli can you provide the pod describe? Can you gett more details why the POD is getting deleted in kubernetes?
I didn't set argo's gc, but my pods are garbage collected
I only query the log of the pod being gc
~$ kubectl logs -n kube-system -f kube-controller-manager-picb2-master-1 | grep sky2-pipeline-ff5gq
I1010 11:16:53.226435 1 gc_controller.go:78] PodGC is force deleting Pod: argo/sky2-pipeline-ff5gq-script-2925226625
I1010 11:16:53.226680 1 gc_controller.go:78] PodGC is force deleting Pod: argo/sky2-pipeline-ff5gq-script-1056479775
I1010 11:17:53.271529 1 gc_controller.go:78] PodGC is force deleting Pod: argo/sky2-pipeline-ff5gq-script-3730124130
I1010 11:19:33.319019 1 gc_controller.go:78] PodGC is force deleting Pod: argo/sky2-pipeline-ff5gq-qiime2-picrust2-1605826948
```I didn't set argo's gc, but my pods are garbage collected
I only query the log of the pod being gc
I found an article that is somewhat familiar to my problem [link](https://blog.csdn.net/alex_yangchuansheng/article/details/119156912)
see #9058
I succeeded after adding both retry and pdb in yaml
见#9058
I found that when my pod is waiting to be scheduled, it is garbage collected, which is related to the --terminated-pod-gc-threshold (default 12500, I set it to 100) set by the kube-controller-manager of k8s. When the number of terminated pods exceeds When this value is used, a pod deleted error will occur
|
2025-04-01T06:37:57.184629
| 2023-06-08T09:34:01
|
1747485224
|
{
"authors": [
"isubasinghe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3833",
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/pull/11189"
}
|
gharchive/pull-request
|
fix: use only relevant files in build
This optimises the nix build time by depending on the required files, changing the docs for example shouldn't force a rebuild.
@terrytangyuan can you please approve this?
It fixes some directory related issues since the nix files are in /dev/nix.
Thanks.
|
2025-04-01T06:37:57.187466
| 2022-03-31T04:41:19
|
1187489109
|
{
"authors": [
"gdsoumya",
"tiwarisanjay"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3834",
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/pull/8280"
}
|
gharchive/pull-request
|
fix: install.yaml missing crb subject ns
This PR fixes the issue with missing ns in crb subjects inside the install.yaml file, introduced in PR #8266
ref: https://github.com/argoproj/argo-workflows/issues/8250#issuecomment-1084046919
@gdsoumya Why exactly we are hardcoding the namespace for cluster install ? whoever refers the cluster install can mention the namespace in their own kustomization.yaml
@alexec @jessesuen
@gdsoumya Yes, Apologies for install.yaml bug I introduced, as it was showing diff when I was fixing kustomize. But isn't kustomize supposed to be for generic use instead of hard-coding a namespace.
Also with kustomize you can not override namespace if its hardcoded. There is a open ticket for the same. :
https://github.com/kubernetes-sigs/kustomize/issues/880
|
2025-04-01T06:37:57.190884
| 2020-07-13T16:41:17
|
655990474
|
{
"authors": [
"saosebastiao",
"simster7"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3835",
"repo": "argoproj/argo",
"url": "https://github.com/argoproj/argo/issues/3463"
}
|
gharchive/issue
|
Windows CLI is gone
Summary
What do you want to know about this project?
What happened to the Windows CLI app?
Motivation
Why do you need to know this, any examples or use cases you could include?
I have been going through some due diligence to test/adopt this project at T-Mobile, but right before we were about to deploy, the Windows CLI downloads seemed to disappear. This could potentially be a dealbreaker. If support for testing a Windows CLI is necessary, we may be able to assist.
The windows CLI is available in our latest release page:
https://github.com/argoproj/argo/releases/tag/v2.9.2
https://github.com/argoproj/argo/releases/download/v2.9.2/argo-windows-amd64
Am I missing something?
Sorry, forgive me, I'm not sure what I was thinking.
|
2025-04-01T06:37:57.193444
| 2018-02-07T13:40:40
|
295141232
|
{
"authors": [
"LaurentVB",
"abelosorio"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3836",
"repo": "arhs/iban.js",
"url": "https://github.com/arhs/iban.js/issues/48"
}
|
gharchive/issue
|
Valid BBAN is considered invalid
I have a valid BBAN but the method isValidBBAN returns false with it:
iban.toBBAN('CR25010200009074883572');
// '0102<PHONE_NUMBER>3572'
iban.isValidBBAN('0102<PHONE_NUMBER>3572');
// false
iban.isValidBBAN(iban.toBBAN('CR25010200009074883572'));
// false
Hello,
Sorry for the super late reply, going through the issues backlog.
The correct way to call isValidBBAN is by passing the country code as first parameter (which is not properly documented, I must say). So your code becomes:
iban.isValidBBAN('CR', iban.toBBAN('CR25010200009074883572'));
// true
Hi @LaurentVB! It's okay.
I wouldn't close this issue until the documentation is updated, since other people may face it.
Thanks!
|
2025-04-01T06:37:57.237044
| 2020-10-02T05:36:08
|
713359780
|
{
"authors": [
"EJFielding",
"dbekaert",
"emlundell"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3837",
"repo": "aria-tools/ARIA-tools",
"url": "https://github.com/aria-tools/ARIA-tools/issues/231"
}
|
gharchive/issue
|
[BUG] ariaDownload gets first file blocked by JPL network URL-block service
Describe the bug
I ran a download of seven products, but the first one was blocked and the .nc contains only an error message. This is while connected to the JPL full tunnel VPN. I ran it a second time after deleting the bad file and got the same result for the first file. Rest of the files are downloaded correctly (or skipped on second try).
To Reproduce
Steps to reproduce the behavior:
Download
Command used
ariaDownload.py -b '32.8 33.2 -115.8 -115.2' --output download -s 20200301
Error Output
First run:
https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
> attempting to download https://urs.earthdata.nasa.gov/profile
> Could not validate old cookie Jar
No existing URS cookie found, please enter Earthdata username & password:
(Credentials will not be stored, saved or logged anywhere)
> attempting to download https://urs.earthdata.nasa.gov/profile
Username: EJFielding
Password (will not be displayed):
> attempting to download https://urs.earthdata.nasa.gov/profile
Using 1 threads for parallel downloads
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
> 'Temporary' Redirect download @ Remote archive:
> http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAawAAABCCjw6LxwMDD0XFz2sNndb~AAAAECMSKEWn~kEQPYh2LihxyoAAAAA7AAAAO~p3kf18EXgRp~5E7Bc25Cb3p91yfIwFRPVW2HQK7Z0DHlPa_2qwelb9CkwUEdEUTmGOoOdal3~2cL95&url=http://<IP_ADDRESS>%2fdoor%2fdownload%2fS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
> Downloaded 2929 of 2929 bytes (100.00%)
Downloaded 2929b in 2.38secs, Average Rate: 0.00MB/sec
> 'Temporary' Redirect download @ Remote archive:
> https://dlz0dhr6g6ukw.cloudfront.net/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc?userid=ejfielding&Expires=1601608741&Signature=DCwkrT8g4OYj8SQyvjGjvFunpGi7~Xfd3giRHVJ1OcM~BvgLb6FVk4NzNVxHZRvlmitYESKwkev7Se2ZlwnoGC2x-g83d2f~Sr6WvjxLbLOZTYI5DhJNesqf2zU4xGSHR8Q74ESdSOyhi68GEq~FTgMkW4BQERAc9M7RP4WOpj0_&Key-Pair-Id=APKAINVNJF4BDB5SS5QQ
(2/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
> Downloaded 62309232 of 62309232 bytes (100.00%)
Downloaded 62309232b in 138.80secs, Average Rate: 0.43MB/sec
second run:
https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
Using 1 threads for parallel downloads
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
> 'Temporary' Redirect download @ Remote archive:
> http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAawAAABDLlKSx~7KaNYP9TBQUf9moAAAAEGkl43OcAp4LLNXlGwliy3IAAAA7AAAAOzlF0zvgVhZ7GyMSdP~v0_uRKv2a3awjbxwSxZkGh1p3qnEhXbs1dP56CzfSU_LUJVq_pTPDcMoG8Cf1&url=http://<IP_ADDRESS>%2fdoor%2fdownload%2fS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
> Downloaded 2933 of 2933 bytes (100.00%)
Downloaded 2933b in 1.96secs, Average Rate: 0.00MB/sec
> Download file S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc.
Expected behavior
I expected all seven files to be downloaded correctly.
Screenshots
If applicable, add screenshots to help explain your problem.
Desktop (please complete the following information):
ARIA-tools git tag: [e.g. git show --summary]
commit b787e827a799fecbf45d6aa267eeb95940148d83 (HEAD -> dev, origin/dev, origin/HEAD)
Author: BB<EMAIL_ADDRESS>Date: Wed Sep 30 08:01:33 2020 -0700
fix file writing bug for KMZ dl (#228)
OS
MacOS 10.14.6, Conda installation
Additional context
The bad download file contains the attached error text:
blocked-download.txt
Not running behind the fire-wall works to download the file: https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc. Although, need to say i needed to run it twice as it reported at first some files reported errors. So seems some transient on the API client side and not something specific to ariaDownload program.
> attempting to download https://urs.earthdata.nasa.gov/profile
> 'Temporary' Redirect download @ Remote archive:
> https://dlz0dhr6g6ukw.cloudfront.net/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc?userid=dbekaert&Expires=1601657555&Signature=aAvOpnnnAooFwvDN~6Uv3o4DaQNLqWqeDfbckHxXd0~RnQzVwdDo2iPJfiUN8kw4I~ajfRsFOkxUG376Cmq3-p2O9fbbcUC3EAegAX0g56XK-OIEDwwwk~Yr9jc8M8yrTaahoOPna-l2GpltUfFG~w0XnfDCdRJX2DtQMJ5Fif4_&Key-Pair-Id=APKAINVNJF4BDB5SS5QQ
(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
> Downloaded 21905408 of 62552131 bytes (35.02%)
> There was an error reading data.
and the second attempt:
Download Summary
--------------------------------------------------------------------------------
Successes: 2 files, 125077866 bytes
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc 59.65MB
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc 59.63MB
Skipped: 5 files
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc
Average Rate: 8.19MB/sec
@EJFielding If this is specific to the JPL fire-wall, and the recent changes implemented with respect to url defense. Suggest you work with your IT admin as unrelated to ARIA-tools.
I verified that this does not happen from my home ISP without the JPL full tunnel VPN. It does happen running on a Linux machine inside the JPL network, so it must be some kind of strange interaction between the ASF API and the JPL firewall. This is the output from the JPL machine:
ariaDownload.py -b '32.8 33.2 -115.8 -115.2' --output download -s 20200301
https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
> attempting to download https://urs.earthdata.nasa.gov/profile
> Could not validate old cookie Jar
No existing URS cookie found, please enter Earthdata username & password:
(Credentials will not be stored, saved or logged anywhere)
> attempting to download https://urs.earthdata.nasa.gov/profile
Username: EJFielding
Password (will not be displayed):
> attempting to download https://urs.earthdata.nasa.gov/profile
Using 1 threads for parallel downloads
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
> 'Temporary' Redirect download @ Remote archive:
> http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBwpAb7QQR4~szp0hI7hescAAAAED~TtODCx4FNxdBfaxdD0xAAAABFAAAARfnSPom0cnv1MFBgOWnEq64Zz9oVnwbZZW_v4zdETR0h7s9J3zCQbjVHLIBciBlXhEbykANQe4ooGANNdTqQb21PBj0H3A==&url=http://<IP_ADDRESS>%2fdoor%2fdownload%2fS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
> Downloaded 2945 of 2945 bytes (100.00%)
Downloaded 2945b in 1.54secs, Average Rate: 0.00MB/sec
> 'Temporary' Redirect download @ Remote archive:
> https://dlz0dhr6g6ukw.cloudfront.net/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc?userid=ejfielding&Expires=1601843075&Signature=TWL79xHFjRrMhOJe5~Lq-hl-rDBX-~wC-5qFanCMNUmodXgi4TjTmYvbtCBeKQ4pNPTYGoJzYZ-FCRQjp8AdPoom68ZVVnWjy5HFHxt3-SrEzdnwVkRYcXpd5DPahBYstgDNXznXBYEEELeQmmkUb04zspb53N~SYjTGlP9XpT8_&Key-Pair-Id=APKAINVNJF4BDB5SS5QQ
(2/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
> Downloaded 62309232 of 62309232 bytes (100.00%)
Downloaded 62309232b in 3.29secs, Average Rate: 18.04MB/sec
I understand now that this is not likely to be due to a bug in ARIA-tools, but it seems like ARIA-tools should do a better job of detecting that there was a download error. I tried running it a second time, and it correctly determined that the previous download was bad, removed it, and tried again, but then it puts the bad file in the "successful" part of the output list when it is still a bad file.
ariaDownload.py -b '32.8 33.2 -115.8 -115.2' --output download -s 20200301
https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
Using 1 threads for parallel downloads
> attempting to download https://urs.earthdata.nasa.gov/profile
> Re-using previous cookie jar.
> Found S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc but it wasn't fully downloaded. Removing file and downloading again.
> 'Temporary' Redirect download @ Remote archive:
> http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBTiFBxhKo4j87OzC7baDDvAAAAEI3H9N2E7hHh59ydT~9cenUAAABFAAAARTD9SOe2fKSdJnx34~LmkIlnKfjA4WESmIery8u5hctUS2FNTHK1TwaCqM50KUc0Zxfln0phHSMmrEkL3J6M7H4KC_hLFQ==&url=http://<IP_ADDRESS>%2fdoor%2fdownload%2fS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
> Downloaded 2949 of 2949 bytes (100.00%)
Downloaded 2949b in 2.31secs, Average Rate: 0.00MB/sec
> Download file S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc.
> Download file S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc.
> Download file S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc.
> Download file S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc.
> Download file S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc.
> Download file S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc exists!
> Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc.
Download Summary
--------------------------------------------------------------------------------
Successes: 1 files, 2949 bytes
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc 0.00MB
Skipped: 6 files
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc
Average Rate: 0.00MB/sec
--------------------------------------------------------------------------------
ariaDownload.py merely wraps around the DAAC API to provide additional
functionality for subsetting. Checking if files are complete, checksums,
etc, the error download message are all passed through via the existing API
functionality to my knowledge. Suggest you open an issue ticket with ASF,
as all the underlying functionality is leverage from the DAAC API.
On Sun, 4 Oct 2020 at 13:34, Eric Fielding<EMAIL_ADDRESS>wrote:
I verified that this does not happen from my home ISP without the JPL full
tunnel VPN. It does happen running on a Linux machine inside the JPL
network, so it must be some kind of strange interaction between the ASF API
and the JPL firewall. This is the output from the JPL machine:
ariaDownload.py -b '32.8 33.2 -115.8 -115.2' --output download -s 20200301https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
attempting to download https://urs.earthdata.nasa.gov/profile
Could not validate old cookie Jar
No existing URS cookie found, please enter Earthdata username & password:
(Credentials will not be stored, saved or logged anywhere)
attempting to download https://urs.earthdata.nasa.gov/profile
Username: EJFielding
Password (will not be displayed):
attempting to download https://urs.earthdata.nasa.gov/profile
Using 1 threads for parallel downloads
attempting to download https://urs.earthdata.nasa.gov/profile
Re-using previous cookie jar.
'Temporary' Redirect download @ Remote archive:
http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBwpAb7QQR4~szp0hI7hescAAAAED~TtODCx4FNxdBfaxdD0xAAAABFAAAARfnSPom0cnv1MFBgOWnEq64Zz9oVnwbZZW_v4zdETR0h7s9J3zCQbjVHLIBciBlXhEbykANQe4ooGANNdTqQb21PBj0H3A==&url=http://<IP_ADDRESS>%2Fdoor%2Fdownload%2FS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7 http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBwpAb7QQR4~szp0hI7hescAAAAED~TtODCx4FNxdBfaxdD0xAAAABFAAAARfnSPom0cnv1MFBgOWnEq64Zz9oVnwbZZW_v4zdETR0h7s9J3zCQbjVHLIBciBlXhEbykANQe4ooGANNdTqQb21PBj0H3A==&url=http://<IP_ADDRESS>%2Fdoor%2Fdownload%2FS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
Downloaded 2945 of 2945 bytes (100.00%)
Downloaded 2945b in 1.54secs, Average Rate: 0.00MB/sec
'Temporary' Redirect download @ Remote archive:
https://dlz0dhr6g6ukw.cloudfront.net/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc?userid=ejfielding&Expires=1601843075&Signature=TWL79xHFjRrMhOJe5~Lq-hl-rDBX-~wC-5qFanCMNUmodXgi4TjTmYvbtCBeKQ4pNPTYGoJzYZ-FCRQjp8AdPoom68ZVVnWjy5HFHxt3-SrEzdnwVkRYcXpd5DPahBYstgDNXznXBYEEELeQmmkUb04zspb53N~SYjTGlP9XpT8_&Key-Pair-Id=APKAINVNJF4BDB5SS5QQ
(2/7 https://dlz0dhr6g6ukw.cloudfront.net/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc?userid=ejfielding&Expires=1601843075&Signature=TWL79xHFjRrMhOJe5~Lq-hl-rDBX-~wC-5qFanCMNUmodXgi4TjTmYvbtCBeKQ4pNPTYGoJzYZ-FCRQjp8AdPoom68ZVVnWjy5HFHxt3-SrEzdnwVkRYcXpd5DPahBYstgDNXznXBYEEELeQmmkUb04zspb53N~SYjTGlP9XpT8_&Key-Pair-Id=APKAINVNJF4BDB5SS5QQ(2/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
Downloaded 62309232 of 62309232 bytes (100.00%)
Downloaded 62309232b in 3.29secs, Average Rate: 18.04MB/sec
I understand now that this is not likely to be due to a bug in ARIA-tools,
but it seems like ARIA-tools should do a better job of detecting that there
was a download error. I tried running it a second time, and it correctly
determined that the previous download was bad, removed it, and tried again,
but then it puts the bad file in the "successful" part of the output list
when it is still a bad file.
ariaDownload.py -b '32.8 33.2 -115.8 -115.2' --output download -s 20200301https://api.daac.asf.alaska.edu/services/search/param?asfplatform=Sentinel-1%20Interferogram%20(BETA)&processingLevel=GUNW_STD&output=JSON&bbox=-115.8,32.8,-115.2,33.2
Attempting to obtaining user/pass from .netrc
WARNING: Could not obtain credentials from existing .netrc
attempting to download https://urs.earthdata.nasa.gov/profile
Re-using previous cookie jar.
Using 1 threads for parallel downloads
attempting to download https://urs.earthdata.nasa.gov/profile
Re-using previous cookie jar.
Found S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc but it wasn't fully downloaded. Removing file and downloading again.
'Temporary' Redirect download @ Remote archive:
http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBTiFBxhKo4j87OzC7baDDvAAAAEI3H9N2E7hHh59ydT~9cenUAAABFAAAARTD9SOe2fKSdJnx34~LmkIlnKfjA4WESmIery8u5hctUS2FNTHK1TwaCqM50KUc0Zxfln0phHSMmrEkL3J6M7H4KC_hLFQ==&url=http://<IP_ADDRESS>%2Fdoor%2Fdownload%2FS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
(1/7 http://<IP_ADDRESS>:6080/php/urlblock.php?args=AAAAdQAAABBTiFBxhKo4j87OzC7baDDvAAAAEI3H9N2E7hHh59ydT~9cenUAAABFAAAARTD9SOe2fKSdJnx34~LmkIlnKfjA4WESmIery8u5hctUS2FNTHK1TwaCqM50KUc0Zxfln0phHSMmrEkL3J6M7H4KC_hLFQ==&url=http://<IP_ADDRESS>%2Fdoor%2Fdownload%2FS1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc(1/7) Downloading https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc
Downloaded 2949 of 2949 bytes (100.00%)
Downloaded 2949b in 2.31secs, Average Rate: 0.00MB/sec
Download file S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc.
Download file S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc.
Download file S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc.
Download file S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc.
Download file S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc.
Download file S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc exists!
Skipping download of https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc.
Download Summary
Successes: 1 files, 2949 bytes
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc 0.00MB
Skipped: 6 files
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134413-33072N_31090N-PP-9984-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200301-134348-34565N_32585N-PP-2e23-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200313_20200307-134348-34565N_32585N-PP-5a99-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-A-R-166-tops-20200313_20200301-014119-34455N_32574N-PP-5234-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134455-33072N_31090N-PP-7fbd-v2_0_2.nc
- https://grfn.asf.alaska.edu/door/download/S1-GUNW-D-R-173-tops-20200307_20200301-134430-34565N_32585N-PP-9eba-v2_0_2.nc
Average Rate: 0.00MB/sec
—
You are receiving this because you were assigned.
Reply to this email directly, view it on GitHub
https://github.com/aria-tools/ARIA-tools/issues/231#issuecomment-703312302,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AESZPSKSVIJYXQNXVYVYRBTSJDL6NANCNFSM4SBI6BNA
.
@emlundell Are you the one to ask about this ASF API problem where it is providing a bad redirect for one of the GRFN products (S1-GUNW-D-R-173-tops-20200313_20200301-134413-33072N_31090N-PP-7694-v2_0_2.nc) that then gets blocked by the JPL firewall?
@EJFielding @glshort is responsible for the ASF Search API up until the GRFN products are requested via grfn.asf.alaska.edu/door/download. After that point @asjohnston-asf is currently responsible. Any bugs or outstanding errors should be redirected to them since they know their systems the best.
I tried this download again from a machine on the JPL network, and it completed correctly, so it seems that whatever was causing the problem was fixed.
|
2025-04-01T06:37:57.245388
| 2021-08-03T02:41:00
|
958662810
|
{
"authors": [
"Ftwrr",
"ariffb25"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3838",
"repo": "ariffb25/stikerinbot",
"url": "https://github.com/ariffb25/stikerinbot/issues/51"
}
|
gharchive/issue
|
sticker.js gini
[UnhandledPromiseRejection: This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason "Maksimal 10 detik!".] {
code: 'ERR_UNHANDLED_REJECTION'
} unhandledRejection
cek kembali
|
2025-04-01T06:37:57.246280
| 2021-04-23T08:52:02
|
865899306
|
{
"authors": [
"ClausHolbechArista"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3839",
"repo": "aristanetworks/ansible-avd",
"url": "https://github.com/aristanetworks/ansible-avd/pull/891"
}
|
gharchive/pull-request
|
Fix references to cvp dirs
Change Summary
Correct CI references to AVD directories instead of CVP.
Not needed
|
2025-04-01T06:37:57.260103
| 2018-05-31T14:05:45
|
328152985
|
{
"authors": [
"aritchie",
"nguyenthanhliemfc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3840",
"repo": "aritchie/geofences",
"url": "https://github.com/aritchie/geofences/issues/12"
}
|
gharchive/issue
|
How long geofence will monitor?
How long geofence will monitor? (1 day, 1 month, forever or one time trigged)
Please check all of the platforms you are having the issue on (if platform is not listed, it is not supported)
[x] iOS
[x] Android
[x] UWP
[x] .NET Standard
Version of OS(s) listed above with issue
Version of Library
Expected Behaviour
Actual Behavior
If applicable, please include entire exception - type, message, stacktrace
Code of Conduct
No Library Update Requests
No Timeline Requests
Please read through the documentation and FAQs
Please be respectful of my time. I put a lot of effort into my libraries. I try to support the issue requests within a day. I'm not psychic, so please take the time to fill out the above form as complete as possible
Until you stop monitoring it. Please take a look through the code yourself in the future. It isn’t that big
Thank you!
|
2025-04-01T06:37:57.278533
| 2014-04-22T05:30:14
|
31946509
|
{
"authors": [
"Vitallium",
"elboletaire",
"phantomhappy101"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3841",
"repo": "ariya/phantomjs",
"url": "https://github.com/ariya/phantomjs/issues/12168"
}
|
gharchive/issue
|
Phantom continually crashing on .abort() or changing the location. (minidump included)
I have about 300 unit tests and about 10% of them are causing phantom to crash.
I minidumped all their .dmp file and they were ALL in the same stack trace (as below)
I removed the line to request.abort() in one of my onResourceRequested events and it immediately fell to about 2% ... so I just had 5-6 .dmp files a this point and these ALSO were crashing at the same didCancel() point.
... and I suspect this is because after the page is loading, some pages have pending resources. But I load about:blank to save memory and that causes the browser to crash since it aborts URLs being loaded (at least that's my theory).
So pretty sure this is a frequent crash issue for others a well.
Operating system: Linux
0.0.0 Linux 3.0.3 #1 SMP Thu Aug 25 19:42:36 UTC 2011 x86_64
CPU: amd64
family 6 model 26 stepping 5
16 CPUs
Crash reason: SIGSEGV
Crash address: 0x0
Thread 0 (crashed)
0 phantomjs!WebCore::ResourceLoader::didCancel [ResourceLoader.cpp : 366 + 0x0]
rbx = 0x00007f8ef239cb00 r12 = 0x00007f8ef2503978
r13 = 0x00007f8ef2399808 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x000000000088573a
rsp = 0x00007ffff7547e28 rbp = 0x00007ffff7547e50
Found by: given as instruction pointer in context
1 0x7f8ef25038ff
rbx = 0x0000000000872e00 r12 = 0x00007f8ef2503978
r13 = 0x00007f8ef2399808 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x00007f8ef2503900
rsp = 0x00007ffff7547e48 rbp = 0x00007f8ef239cb00
Found by: call frame info
2 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 381 + 0x7]
rip = 0x0000000000885527 rsp = 0x00007ffff7547e50
Found by: stack scanning
3 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 371 + 0x1]
rbx = 0x00007f8ef2399800 r12 = 0x00007f8ef2503900
rip = 0x0000000000884e69 rsp = 0x00007ffff7547ea0
rbp = 0x00007f8ef239cb00
Found by: call frame info
4 phantomjs!WebCore::DocumentLoader::stopLoading [DocumentLoader.cpp : 252 + 0x7]
rbx = 0x00007f8ef2399800 r12 = 0x00007f8ef2503900
rip = 0x0000000000843932 rsp = 0x00007ffff7547ef0
rbp = 0x00007f8ef239cb00
Found by: call frame info
5 phantomjs!WebCore::NavigationScheduler::schedule [NavigationScheduler.cpp : 405 + 0x4]
rbx = 0x00007f8ef2503c98 r12 = 0x0000000000000001
r13 = 0x00007f8ef31b1f50 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000882e86
rsp = 0x00007ffff7547fa0 rbp = 0x00007ffff7548080
Found by: call frame info
6 phantomjs!WebCore::NavigationScheduler::scheduleLocationChange [NavigationScheduler.cpp : 333 + 0x4]
rbx = 0x00007f8ef2503c98 r12 = 0x0000000000000001
r13 = 0x00007f8ef31b1f50 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000883fd9
rsp = 0x00007ffff7547fc0 rbp = 0x00007ffff7548180
Found by: call frame info
7 phantomjs!WebCore::SubframeLoader::loadOrRedirectSubframe [SubframeLoader.cpp : 238 + 0x19]
rbx = 0x0000000000000000 r12 = 0x00007ffff75480f0
r13 = 0x00007f8eec06b2a0 r14 = 0x0000000000000000
r15 = 0x00007f8ef2503900 rip = 0x000000000088e41f
rsp = 0x00007ffff75480e0 rbp = 0x00007f8f3aed9d10
Found by: call frame info
8 phantomjs!WebCore::SubframeLoader::requestFrame [SubframeLoader.cpp : 83 + 0x1d]
rbx = 0x0000000000000000 r12 = 0x0000000000000001
r13 = 0x00007f8f3aed9d10 r14 = 0x0000000000000000
r15 = 0x0000000000000000 rip = 0x000000000088efa8
rsp = 0x00007ffff7548160 rbp = 0x00007f8ef22e57e0
Found by: call frame info
9 phantomjs!WebCore::HTMLFrameElementBase::openURL [HTMLFrameElementBase.cpp : 108 + 0x23]
rbx = 0x0000000000000001 r12 = 0x00007f8eec06b210
r13 = 0x00007f8eec06b210 r14 = 0x00007f8eec06b298
r15 = 0x0000000000000000 rip = 0x000000000077b238
rsp = 0x00007ffff75482c0 rbp = 0x00007f8ef3591b60
Found by: call frame info
10 phantomjs!WebCore::HTMLFrameElementBase::parseMappedAttribute [HTMLFrameElementBase.cpp : 116 + 0x16]
rbx = 0x0000000000000001 r12 = 0x00007f8eec06b210
r13 = 0x00007f8eec06b210 r14 = 0x0000000000000000
r15 = 0x00007f8eec06b210 rip = 0x000000000077b5c7
rsp = 0x00007ffff7548320 rbp = 0x00007f8ef3591b60
Found by: call frame info
11 phantomjs!WebCore::HTMLIFrameElement::parseMappedAttribute [HTMLIFrameElement.cpp : 131 + 0x4]
rbx = 0x0000000000000001 r12 = 0x00007f8ef3591b60
r13 = 0x00007f8eec06b210 r14 = 0x0000000000000000
r15 = 0x00007f8eec06b210 rip = 0x00000000010e170f
rsp = 0x00007ffff75483b0 rbp = 0x00007f8ef3591b60
Found by: call frame info
12 phantomjs!WebCore::StyledElement::attributeChanged [StyledElement.cpp : 188 + 0xf]
rbx = 0x0000000000000001 r12 = 0x00007f8ef3591b60
r13 = 0x00007f8eec06b210 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x00000000010734a4
rsp = 0x00007ffff7548410 rbp = 0x0000000000000001
Found by: call frame info
13 phantomjs!WebCore::Element::setAttribute [Element.cpp : 692 + 0x15]
rbx = 0x00007f8ef3591b60 r12 = 0x0000000000000008
r13 = 0x00007f8ef2e9d1d8 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x0000000000634762
rsp = 0x00007ffff75484b0 rbp = 0x0000000000000000
Found by: call frame info
14 phantomjs!WebCore::jsElementPrototypeFunctionSetAttribute [JSElement.cpp : 1614 + 0x1a]
rbx = 0x000000000000000a r12 = 0x00007f8f3b7e11b8
r13 = 0x00007f8ef19117e0 r14 = 0x00007f8f3b9f5fd0
r15 = 0x0000000000000001 rip = 0x0000000000c2def4
rsp = 0x00007ffff7548550 rbp = 0x00007f8f3b9f5fe8
Found by: call frame info
15 0x7f8ef40001e7
rbx = 0x00007f8f3b606e60 r12 = 0x000000000327599a
r13 = 0x00007f8f3b9f5f58 r14 = 0xffff000000000000
r15 = 0xffff000000000002 rip = 0x00007f8ef40001e8
rsp = 0x00007ffff7548620 rbp = 0x00007ffff75486d0
Found by: call frame info
16 phantomjs!JSC::Interpreter::execute [JITCode.h : 77 + 0x7]
rip = 0x0000000000f2c1eb rsp = 0x00007ffff75486e0
Found by: stack scanning
17 phantomjs!JSC::evaluate [Completion.cpp : 64 + 0x1b]
rbx = 0x00007f8f3c62b868 r12 = 0x00007f8ef1aa1d50
r13 = 0x00007ffff75489a8 r14 = 0x00007f8f3efa8940
r15 = 0x00007f8f3b606e60 rip = 0x00000000004f9538
rsp = 0x00007ffff7548760 rbp = 0x00007f8f3efa8958
Found by: call frame info
18 phantomjs!WebCore::ScriptController::evaluateInWorld [JSMainThreadExecState.h : 54 + 0x4]
rbx = 0x0000000002a9ed10 r12 = 0x00007f8f3efa8958
r13 = 0x00007f8f3aed9f18 r14 = 0x00007f8f3aed9b00
r15 = 0x00007f8f3b5e25d0 rip = 0x00000000005291da
rsp = 0x00007ffff7548810 rbp = 0x00007f8f3efa8958
Found by: call frame info
19 phantomjs!WebCore::ScriptController::evaluate [ScriptController.cpp : 167 + 0x15]
rbx = 0x00007ffff75489a0 r12 = 0x00007f8f3aed9f18
r13 = 0x00007ffff7548920 r14 = 0x0000000000000001
r15 = 0x0000000000000001 rip = 0x00000000005294d2
rsp = 0x00007ffff7548900 rbp = 0x00007ffff7548a00
Found by: call frame info
20 phantomjs!WebCore::ScriptController::executeScript [ScriptControllerBase.cpp : 64 + 0xd]
rbx = 0x00007f8f3aed9f18 r12 = 0x00007ffff75489a0
r13 = 0x00007f8f3aed9b00 r14 = 0x0000000000000001
r15 = 0x0000000000000001 rip = 0x000000000051a475
rsp = 0x00007ffff7548920 rbp = 0x00007ffff7548a00
Found by: call frame info
21 phantomjs!QWebFrame::evaluateJavaScript [qwebframe.cpp : 1556 + 0x23]
rbx = 0x00007ffff75489a0 r12 = 0x0000000000000001
r13 = 0x00007f8ef1e22e10 r14 = 0x00007ffff7548aa0
r15 = 0x00007f8f3aed9f18 rip = 0x000000000047fb3c
rsp = 0x00007ffff7548980 rbp = 0x0000000000000001
Found by: call frame info
22 phantomjs!WebPage::evaluateJavaScript [webpage.cpp : 683 + 0x13]
rbx = 0x00007f8eec1e14f0 r12 = 0x00007f8eec1e1500
r13 = 0x00007ffff7548c00 r14 = 0x00007ffff7548ab0
r15 = 0x000000000401ec00 rip = 0x000000000041f1b9
rsp = 0x00007ffff7548aa0 rbp = 0x00007ffff7548b30
Found by: call frame info
23 phantomjs!WebPage::qt_static_metacall [moc_webpage.cpp : 224 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aa42
rsp = 0x00007ffff7548c00 rbp = 0x00007ffff75495c0
Found by: call frame info
24 phantomjs!WebPage::qt_metacall [moc_webpage.cpp : 338 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aebf
rsp = 0x00007ffff7548d30 rbp = 0x00007ffff75495c0
Found by: call frame info
25 phantomjs!JSC::Bindings::QtRuntimeMetaMethod::call [qt_runtime.cpp : 1455 + 0x17]
rbx = 0x00007ffff7549390 r12 = 0x00007ffff75495c0
r13 = 0x00007f8ef1a9cdd0 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x0000000000552f07
rsp = 0x00007ffff7548ec0 rbp = 0x0000000000000002
Found by: call frame info
26 phantomjs!JSC::cti_op_call_NotJSFunction [JITStubs.cpp : 2191 + 0x3]
rbx = 0x0000000000000000 r12 = 0x00007f8f3b9f5b80
r13 = 0x00007f8ef276d260 r14 = 0x00007ffff7549cc8
r15 = 0x00007f8f3b9f5b00 rip = 0x0000000000f5d075
rsp = 0x00007ffff7549c70 rbp = 0x00007ffff7549cd0
Found by: call frame info
27 0x7f8ef473b440
rbx = 0x00007f8f3efbcb10 r12 = 0x0000000000000178
r13 = 0x00007f8f3b9f5b00 r14 = 0xffff000000000000
r15 = 0xffff000000000002 rip = 0x00007f8ef473b441
rsp = 0x00007ffff7549cd0 rbp = 0x00007ffff7549d70
Found by: call frame info
28 libc-2.11.3.so + 0x76c6f
rip = 0x00007f8f3d9bbc70 rsp = 0x00007ffff7549d10
rbp = 0x00007ffff7549d70
Found by: stack scanning
29 phantomjs!JSC::Interpreter::executeCall [JITCode.h : 77 + 0x16]
rip = 0x0000000000f2e376 rsp = 0x00007ffff7549d80
Found by: stack scanning
30 phantomjs!JSC::call [CallData.cpp : 38 + 0x21]
rbx = 0x00007ffff7549fb0 r12 = 0x00007f8ef1f0c5b0
r13 = 0x00007f8f3b5c45d0 r14 = 0x00007f8f3efa8150
r15 = 0x00007f8f3c6168f0 rip = 0x0000000000f7291d
rsp = 0x00007ffff7549e50 rbp = 0x00007f8ef1bcd790
Found by: call frame info
31 phantomjs!JSC::Bindings::QtConnectionObject::execute [qt_runtime.cpp : 1863 + 0x17]
rbx = 0x00007ffff7549fb0 r12 = 0x00007f8ef1f0c5b0
r13 = 0x00007f8f3b5c45d0 r14 = 0x00007f8f3efa8150
r15 = 0x00007f8f3c6168f0 rip = 0x0000000000542ece
rsp = 0x00007ffff7549e60 rbp = 0x00007f8ef1bcd790
Found by: call frame info
32 phantomjs!JSC::Bindings::QtConnectionObject::qt_metacall [qt_runtime.cpp : 1807 + 0xa]
rbx = 0x0000000000000000 r12 = 0x00007ffff754a230
r13 = 0x000000000405f440 r14 = 0x0000000000000004
r15 = 0x0000000004079b10 rip = 0x000000000054390c
rsp = 0x00007ffff754a0e0 rbp = 0x0000000000000000
Found by: call frame info
33 phantomjs!QMetaObject::activate [qobject.cpp : 3555 + 0x1f]
rbx = 0x0000000000000004 r12 = 0x000000000401ec01
r13 = 0x0000000000000000 r14 = 0x0000000000000004
r15 = 0x0000000004079b10 rip = 0x0000000001cf7160
rsp = 0x00007ffff754a110 rbp = 0x000000000405f440
Found by: call frame info
34 phantomjs!WebPage::resourceError [moc_webpage.cpp : 470 + 0x4]
rbx = 0x0000000000000004 r12 = 0x000000000401ec00
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x000000000046a033
rsp = 0x00007ffff754a230 rbp = 0x00007ffff754a4a0
Found by: call frame info
35 phantomjs!WebPage::qt_static_metacall [moc_webpage.cpp : 215 + 0x8]
rbx = 0x0000000000000004 r12 = 0x000000000401ec00
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x000000000046aa69
rsp = 0x00007ffff754a250 rbp = 0x00007ffff754a4a0
Found by: call frame info
36 phantomjs!QMetaObject::activate [qobject.cpp : 3539 + 0x1e]
rbx = 0x0000000000000004 r12 = 0x00007ffff754a501
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x0000000001cf6ef8
rsp = 0x00007ffff754a380 rbp = 0x000000000401ec00
Found by: call frame info
37 phantomjs!NetworkAccessManager::resourceError [moc_networkaccessmanager.cpp : 278 + 0x4]
rbx = 0x00007ffff754a570 r12 = 0x00007ffff754a5b0
r13 = 0x00007ffff754a5a0 r14 = 0x00007ffff754a590
r15 = 0x00007ffff754a5c0 rip = 0x000000000046c483
rsp = 0x00007ffff754a4a0 rbp = 0x00007ffff754a560
Found by: call frame info
38 phantomjs!NetworkAccessManager::handleNetworkError [networkaccessmanager.cpp : 397 + 0x1f]
rbx = 0x00007ffff754a570 r12 = 0x00007ffff754a5b0
r13 = 0x00007ffff754a5a0 r14 = 0x00007ffff754a590
r15 = 0x00007ffff754a5c0 rip = 0x000000000042b903
rsp = 0x00007ffff754a4c0 rbp = 0x00007ffff754a560
Found by: call frame info
39 phantomjs!QMetaObject::activate [qobject.cpp : 3539 + 0x1e]
rbx = 0x000000000000000e r12 = 0x00007f8ef2504b01
r13 = 0x000000000046c540 r14 = 0x0000000000000008
r15 = 0x00000000043b12f0 rip = 0x0000000001cf6ef8
rsp = 0x00007ffff754a780 rbp = 0x00000000040211d0
Found by: call frame info
40 phantomjs!QNetworkReply::error [moc_qnetworkreply.cpp : 173 + 0x4]
rbx = 0x00000000043afb10 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000001ba74c1
rsp = 0x00007ffff754a8a0 rbp = 0x00000000043afd50
Found by: call frame info
41 phantomjs!QNetworkReplyImpl::abort [qnetworkreplyimpl.cpp : 882 + 0xf]
rbx = 0x00000000043afb10 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000001b32888
rsp = 0x00007ffff754a8d0 rbp = 0x00000000043afd50
Found by: call frame info
42 phantomjs!WebCore::QNetworkReplyHandler::abort [QNetworkReplyHandler.cpp : 416 + 0xb]
rbx = 0x00000000043afb10 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000ada4f2
rsp = 0x00007ffff754a900 rbp = 0x00000000043add60
Found by: call frame info
43 phantomjs!WebCore::ResourceHandle::cancel [ResourceHandleQt.cpp : 107 + 0x4]
rbx = 0x00007f8ef251a600 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000ad7f79
rsp = 0x00007ffff754a920 rbp = 0x00007ffff754a970
Found by: call frame info
44 phantomjs!WebCore::ResourceLoader::didCancel [ResourceLoader.cpp : 360 + 0x5]
rbx = 0x00007f8ef2538000 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x00000000008856c5
rsp = 0x00007ffff754a930 rbp = 0x00007ffff754a970
Found by: call frame info
45 phantomjs!WebCore::MainResourceLoader::didCancel [MainResourceLoader.cpp : 114 + 0xa]
rbx = 0x00007f8ef2538000 r12 = 0x00007f8ef2504b78
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000872e44
rsp = 0x00007ffff754a950 rbp = 0x00007ffff754a970
Found by: call frame info
46 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 381 + 0x7]
rbx = 0x0000000000872e00 r12 = 0x00007f8ef2504b00
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000885527
rsp = 0x00007ffff754a970 rbp = 0x00007f8ef2538000
Found by: call frame info
47 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 371 + 0x1]
rbx = 0x00007f8ef253d000 r12 = 0x00007f8ef2504b00
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000884e69
rsp = 0x00007ffff754a9c0 rbp = 0x00007f8ef2538000
Found by: call frame info
48 phantomjs!WebCore::DocumentLoader::stopLoading [DocumentLoader.cpp : 252 + 0x7]
rbx = 0x00007f8ef253d000 r12 = 0x00007f8ef2504b00
r13 = 0x00007f8ef253d008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000843932
rsp = 0x00007ffff754aa10 rbp = 0x00007f8ef2538000
Found by: call frame info
49 phantomjs!WebCore::NavigationScheduler::schedule [NavigationScheduler.cpp : 405 + 0x4]
rbx = 0x00007f8ef2504e98 r12 = 0x0000000000000001
r13 = 0x00007f8ef1ddaea8 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000882e86
rsp = 0x00007ffff754aac0 rbp = 0x00007ffff754aba0
Found by: call frame info
50 phantomjs!WebCore::NavigationScheduler::scheduleLocationChange [NavigationScheduler.cpp : 333 + 0x4]
rbx = 0x00007f8ef2504e98 r12 = 0x0000000000000001
r13 = 0x00007f8ef1ddaea8 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000883fd9
rsp = 0x00007ffff754aae0 rbp = 0x00007ffff754aca0
Found by: call frame info
51 phantomjs!WebCore::SubframeLoader::loadOrRedirectSubframe [SubframeLoader.cpp : 238 + 0x19]
rbx = 0x0000000000000000 r12 = 0x00007ffff754ac10
r13 = 0x00000000043a8370 r14 = 0x0000000000000000
r15 = 0x00007f8ef2504b00 rip = 0x000000000088e41f
rsp = 0x00007ffff754ac00 rbp = 0x00007f8f3aed9d10
Found by: call frame info
52 phantomjs!WebCore::SubframeLoader::requestFrame [SubframeLoader.cpp : 83 + 0x1d]
rbx = 0x0000000000000000 r12 = 0x0000000000000001
r13 = 0x00007f8f3aed9d10 r14 = 0x0000000000000000
r15 = 0x0000000000000000 rip = 0x000000000088efa8
rsp = 0x00007ffff754ac80 rbp = 0x00007f8ef22e57e0
Found by: call frame info
53 phantomjs!WebCore::HTMLFrameElementBase::openURL [HTMLFrameElementBase.cpp : 108 + 0x23]
rbx = 0x0000000000000001 r12 = 0x00000000043a82e0
r13 = 0x00000000043a82e0 r14 = 0x00000000043a8368
r15 = 0x0000000000000000 rip = 0x000000000077b238
rsp = 0x00007ffff754ade0 rbp = 0x00007f8ef33ab240
Found by: call frame info
54 phantomjs!WebCore::HTMLFrameElementBase::parseMappedAttribute [HTMLFrameElementBase.cpp : 116 + 0x16]
rbx = 0x0000000000000001 r12 = 0x00000000043a82e0
r13 = 0x00000000043a82e0 r14 = 0x0000000000000000
r15 = 0x00000000043a82e0 rip = 0x000000000077b5c7
rsp = 0x00007ffff754ae40 rbp = 0x00007f8ef33ab240
Found by: call frame info
55 phantomjs!WebCore::HTMLIFrameElement::parseMappedAttribute [HTMLIFrameElement.cpp : 131 + 0x4]
rbx = 0x0000000000000001 r12 = 0x00007f8ef33ab240
r13 = 0x00000000043a82e0 r14 = 0x0000000000000000
r15 = 0x00000000043a82e0 rip = 0x00000000010e170f
rsp = 0x00007ffff754aed0 rbp = 0x00007f8ef33ab240
Found by: call frame info
56 phantomjs!WebCore::StyledElement::attributeChanged [StyledElement.cpp : 188 + 0xf]
rbx = 0x0000000000000001 r12 = 0x00007f8ef33ab240
r13 = 0x00000000043a82e0 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x00000000010734a4
rsp = 0x00007ffff754af30 rbp = 0x0000000000000001
Found by: call frame info
57 phantomjs!WebCore::Element::setAttribute [Element.cpp : 692 + 0x15]
rbx = 0x00007f8ef33ab240 r12 = 0x0000000000000008
r13 = 0x00007f8f38325868 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x0000000000634762
rsp = 0x00007ffff754afd0 rbp = 0x0000000000000000
Found by: call frame info
58 phantomjs!WebCore::jsElementPrototypeFunctionSetAttribute [JSElement.cpp : 1614 + 0x1a]
rbx = 0x000000000000000a r12 = 0x00007f8f3b7e11b8
r13 = 0x00007f8ef1de0000 r14 = 0x00007f8f3b9f53b0
r15 = 0x0000000000000001 rip = 0x0000000000c2def4
rsp = 0x00007ffff754b070 rbp = 0x00007f8f3b9f53c8
Found by: call frame info
59 0x7f8ef40001e7
rbx = 0x00007f8f3b606e60 r12 = 0x0000000000283c53
r13 = 0x00007f8f3b9f5338 r14 = 0xffff000000000000
r15 = 0xffff000000000002 rip = 0x00007f8ef40001e8
rsp = 0x00007ffff754b140 rbp = 0x00007ffff754b1f0
Found by: call frame info
60 phantomjs!JSC::Interpreter::execute [JITCode.h : 77 + 0x7]
rip = 0x0000000000f2c1eb rsp = 0x00007ffff754b200
Found by: stack scanning
61 phantomjs!JSC::evaluate [Completion.cpp : 64 + 0x1b]
rbx = 0x00007f8f3c62b868 r12 = 0x00007f8ef1ee2250
r13 = 0x00007ffff754b4c8 r14 = 0x00007f8f3efa8940
r15 = 0x00007f8f3b606e60 rip = 0x00000000004f9538
rsp = 0x00007ffff754b280 rbp = 0x00007f8f3efa8958
Found by: call frame info
62 phantomjs!WebCore::ScriptController::evaluateInWorld [JSMainThreadExecState.h : 54 + 0x4]
rbx = 0x0000000002a9ed10 r12 = 0x00007f8f3efa8958
r13 = 0x00007f8f3aed9f18 r14 = 0x00007f8f3aed9b00
r15 = 0x00007f8f3b5e25d0 rip = 0x00000000005291da
rsp = 0x00007ffff754b330 rbp = 0x00007f8f3efa8958
Found by: call frame info
63 phantomjs!WebCore::ScriptController::evaluate [ScriptController.cpp : 167 + 0x15]
rbx = 0x00007ffff754b4c0 r12 = 0x00007f8f3aed9f18
r13 = 0x00007ffff754b440 r14 = 0x0000000000000001
r15 = 0x0000000000000001 rip = 0x00000000005294d2
rsp = 0x00007ffff754b420 rbp = 0x00007ffff754b520
Found by: call frame info
64 phantomjs!WebCore::ScriptController::executeScript [ScriptControllerBase.cpp : 64 + 0xd]
rbx = 0x00007f8f3aed9f18 r12 = 0x00007ffff754b4c0
r13 = 0x00007f8f3aed9b00 r14 = 0x0000000000000001
r15 = 0x0000000000000001 rip = 0x000000000051a475
rsp = 0x00007ffff754b440 rbp = 0x00007ffff754b520
Found by: call frame info
65 phantomjs!QWebFrame::evaluateJavaScript [qwebframe.cpp : 1556 + 0x23]
rbx = 0x00007ffff754b4c0 r12 = 0x0000000000000001
r13 = 0x00007f8ef1f61e50 r14 = 0x00007ffff754b5c0
r15 = 0x00007f8f3aed9f18 rip = 0x000000000047fb3c
rsp = 0x00007ffff754b4a0 rbp = 0x0000000000000001
Found by: call frame info
66 phantomjs!WebPage::evaluateJavaScript [webpage.cpp : 683 + 0x13]
rbx = 0x00007f8eec1b4c80 r12 = 0x00007f8eec1b4c90
r13 = 0x00007ffff754b720 r14 = 0x00007ffff754b5d0
r15 = 0x000000000401ec00 rip = 0x000000000041f1b9
rsp = 0x00007ffff754b5c0 rbp = 0x00007ffff754b650
Found by: call frame info
67 phantomjs!WebPage::qt_static_metacall [moc_webpage.cpp : 224 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aa42
rsp = 0x00007ffff754b720 rbp = 0x00007ffff754c0e0
Found by: call frame info
68 phantomjs!WebPage::qt_metacall [moc_webpage.cpp : 338 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aebf
rsp = 0x00007ffff754b850 rbp = 0x00007ffff754c0e0
Found by: call frame info
69 phantomjs!JSC::Bindings::QtRuntimeMetaMethod::call [qt_runtime.cpp : 1455 + 0x17]
rbx = 0x00007ffff754beb0 r12 = 0x00007ffff754c0e0
r13 = 0x00007f8ef1eebe10 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x0000000000552f07
rsp = 0x00007ffff754b9e0 rbp = 0x0000000000000002
Found by: call frame info
70 phantomjs!JSC::cti_op_call_NotJSFunction [JITStubs.cpp : 2191 + 0x3]
rbx = 0x0000000000000000 r12 = 0x00007f8f3b9f4f60
r13 = 0x00007f8ef276d260 r14 = 0x00007ffff754c7e8
r15 = 0x00007f8f3b9f4ee0 rip = 0x0000000000f5d075
rsp = 0x00007ffff754c790 rbp = 0x00007ffff754c7f0
Found by: call frame info
71 0x7f8ef473b440
rbx = 0x00007f8ef473b37b r12 = 0x0000000000000178
r13 = 0x00007f8f3b9f4ee0 r14 = 0xffff000000000000
r15 = 0xffff000000000002 rip = 0x00007f8ef473b441
rsp = 0x00007ffff754c7f0 rbp = 0x00007ffff754c890
Found by: call frame info
72 libc-2.11.3.so + 0x76c6f
rip = 0x00007f8f3d9bbc70 rsp = 0x00007ffff754c830
rbp = 0x00007ffff754c890
Found by: stack scanning
73 phantomjs!JSC::Interpreter::executeCall [JITCode.h : 77 + 0x16]
rip = 0x0000000000f2e376 rsp = 0x00007ffff754c8a0
Found by: stack scanning
74 phantomjs!JSC::call [CallData.cpp : 38 + 0x21]
rbx = 0x00007ffff754cad0 r12 = 0x00007f8f3b6061e8
r13 = 0x00007f8f3b5c45d0 r14 = 0x00007f8f3efa8150
r15 = 0x00007f8f3c6168f0 rip = 0x0000000000f7291d
rsp = 0x00007ffff754c970 rbp = 0x00007f8ef1f04790
Found by: call frame info
75 phantomjs!JSC::Bindings::QtConnectionObject::execute [qt_runtime.cpp : 1863 + 0x17]
rbx = 0x00007ffff754cad0 r12 = 0x00007f8f3b6061e8
r13 = 0x00007f8f3b5c45d0 r14 = 0x00007f8f3efa8150
r15 = 0x00007f8f3c6168f0 rip = 0x0000000000542ece
rsp = 0x00007ffff754c980 rbp = 0x00007f8ef1f04790
Found by: call frame info
76 phantomjs!JSC::Bindings::QtConnectionObject::qt_metacall [qt_runtime.cpp : 1807 + 0xa]
rbx = 0x0000000000000000 r12 = 0x00007ffff754cd50
r13 = 0x000000000405f440 r14 = 0x0000000000000004
r15 = 0x0000000004079b10 rip = 0x000000000054390c
rsp = 0x00007ffff754cc00 rbp = 0x0000000000000000
Found by: call frame info
77 phantomjs!QMetaObject::activate [qobject.cpp : 3555 + 0x1f]
rbx = 0x0000000000000004 r12 = 0x000000000401ec01
r13 = 0x0000000000000000 r14 = 0x0000000000000004
r15 = 0x0000000004079b10 rip = 0x0000000001cf7160
rsp = 0x00007ffff754cc30 rbp = 0x000000000405f440
Found by: call frame info
78 phantomjs!WebPage::resourceError [moc_webpage.cpp : 470 + 0x4]
rbx = 0x0000000000000004 r12 = 0x000000000401ec00
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x000000000046a033
rsp = 0x00007ffff754cd50 rbp = 0x00007ffff754cfc0
Found by: call frame info
79 phantomjs!WebPage::qt_static_metacall [moc_webpage.cpp : 215 + 0x8]
rbx = 0x0000000000000004 r12 = 0x000000000401ec00
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x000000000046aa69
rsp = 0x00007ffff754cd70 rbp = 0x00007ffff754cfc0
Found by: call frame info
80 phantomjs!QMetaObject::activate [qobject.cpp : 3539 + 0x1e]
rbx = 0x0000000000000004 r12 = 0x00007ffff754d001
r13 = 0x000000000046a210 r14 = 0x0000000000000008
r15 = 0x0000000004057dc0 rip = 0x0000000001cf6ef8
rsp = 0x00007ffff754cea0 rbp = 0x000000000401ec00
Found by: call frame info
81 phantomjs!NetworkAccessManager::resourceError [moc_networkaccessmanager.cpp : 278 + 0x4]
rbx = 0x00007ffff754d090 r12 = 0x00007ffff754d0d0
r13 = 0x00007ffff754d0c0 r14 = 0x00007ffff754d0b0
r15 = 0x00007ffff754d0e0 rip = 0x000000000046c483
rsp = 0x00007ffff754cfc0 rbp = 0x00007ffff754d080
Found by: call frame info
82 phantomjs!NetworkAccessManager::handleNetworkError [networkaccessmanager.cpp : 397 + 0x1f]
rbx = 0x00007ffff754d090 r12 = 0x00007ffff754d0d0
r13 = 0x00007ffff754d0c0 r14 = 0x00007ffff754d0b0
r15 = 0x00007ffff754d0e0 rip = 0x000000000042b903
rsp = 0x00007ffff754cfe0 rbp = 0x00007ffff754d080
Found by: call frame info
83 phantomjs!QMetaObject::activate [qobject.cpp : 3539 + 0x1e]
rbx = 0x000000000000000e r12 = 0x00007f8ef2d94001
r13 = 0x000000000046c540 r14 = 0x0000000000000008
r15 = 0x00007f8eec04df70 rip = 0x0000000001cf6ef8
rsp = 0x00007ffff754d2a0 rbp = 0x00000000040211d0
Found by: call frame info
84 phantomjs!QNetworkReply::error [moc_qnetworkreply.cpp : 173 + 0x4]
rbx = 0x00007f8eec04f7b0 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000001ba74c1
rsp = 0x00007ffff754d3c0 rbp = 0x00007f8eec0393d0
Found by: call frame info
85 phantomjs!QNetworkReplyImpl::abort [qnetworkreplyimpl.cpp : 882 + 0xf]
rbx = 0x00007f8eec04f7b0 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000001b32888
rsp = 0x00007ffff754d3f0 rbp = 0x00007f8eec0393d0
Found by: call frame info
86 phantomjs!WebCore::QNetworkReplyHandler::abort [QNetworkReplyHandler.cpp : 416 + 0xb]
rbx = 0x00007f8eec04f7b0 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000ada4f2
rsp = 0x00007ffff754d420 rbp = 0x00007f8eec07e230
Found by: call frame info
87 phantomjs!WebCore::ResourceHandle::cancel [ResourceHandleQt.cpp : 107 + 0x4]
rbx = 0x00007f8ef251a918 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000ad7f79
rsp = 0x00007ffff754d440 rbp = 0x00007ffff754d490
Found by: call frame info
88 phantomjs!WebCore::ResourceLoader::didCancel [ResourceLoader.cpp : 360 + 0x5]
rbx = 0x00007f8ef239b900 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x00000000008856c5
rsp = 0x00007ffff754d450 rbp = 0x00007ffff754d490
Found by: call frame info
89 phantomjs!WebCore::MainResourceLoader::didCancel [MainResourceLoader.cpp : 114 + 0xa]
rbx = 0x00007f8ef239b900 r12 = 0x00007f8ef2d94078
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000872e44
rsp = 0x00007ffff754d470 rbp = 0x00007ffff754d490
Found by: call frame info
90 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 381 + 0x7]
rbx = 0x0000000000872e00 r12 = 0x00007f8ef2d94000
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000885527
rsp = 0x00007ffff754d490 rbp = 0x00007f8ef239b900
Found by: call frame info
91 phantomjs!WebCore::ResourceLoader::cancel [ResourceLoader.cpp : 371 + 0x1]
rbx = 0x00007f8ef2399000 r12 = 0x00007f8ef2d94000
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000884e69
rsp = 0x00007ffff754d4e0 rbp = 0x00007f8ef239b900
Found by: call frame info
92 phantomjs!WebCore::DocumentLoader::stopLoading [DocumentLoader.cpp : 252 + 0x7]
rbx = 0x00007f8ef2399000 r12 = 0x00007f8ef2d94000
r13 = 0x00007f8ef2399008 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000843932
rsp = 0x00007ffff754d530 rbp = 0x00007f8ef239b900
Found by: call frame info
93 phantomjs!WebCore::NavigationScheduler::schedule [NavigationScheduler.cpp : 405 + 0x4]
rbx = 0x00007f8ef2d94398 r12 = 0x0000000000000001
r13 = 0x00007f8ef227e1c0 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000882e86
rsp = 0x00007ffff754d5e0 rbp = 0x00007ffff754d6c0
Found by: call frame info
94 phantomjs!WebCore::NavigationScheduler::scheduleLocationChange [NavigationScheduler.cpp : 333 + 0x4]
rbx = 0x00007f8ef2d94398 r12 = 0x0000000000000001
r13 = 0x00007f8ef227e1c0 r14 = 0x0000000000000001
r15 = 0x00007f8f3c6ab3f0 rip = 0x0000000000883fd9
rsp = 0x00007ffff754d600 rbp = 0x00007ffff754d7c0
Found by: call frame info
95 phantomjs!WebCore::SubframeLoader::loadOrRedirectSubframe [SubframeLoader.cpp : 238 + 0x19]
rbx = 0x0000000000000000 r12 = 0x00007ffff754d730
r13 = 0x0000000004280f20 r14 = 0x0000000000000000
r15 = 0x00007f8ef2d94000 rip = 0x000000000088e41f
rsp = 0x00007ffff754d720 rbp = 0x00007f8f3aed9d10
Found by: call frame info
96 phantomjs!WebCore::SubframeLoader::requestFrame [SubframeLoader.cpp : 83 + 0x1d]
rbx = 0x0000000000000000 r12 = 0x0000000000000001
r13 = 0x00007f8f3aed9d10 r14 = 0x0000000000000000
r15 = 0x0000000000000000 rip = 0x000000000088efa8
rsp = 0x00007ffff754d7a0 rbp = 0x00007f8ef22e57e0
Found by: call frame info
97 phantomjs!WebCore::HTMLFrameElementBase::openURL [HTMLFrameElementBase.cpp : 108 + 0x23]
rbx = 0x0000000000000001 r12 = 0x0000000004280e90
r13 = 0x0000000004280e90 r14 = 0x0000000004280f18
r15 = 0x0000000000000000 rip = 0x000000000077b238
rsp = 0x00007ffff754d900 rbp = 0x00007f8ef2ea3700
Found by: call frame info
98 phantomjs!WebCore::HTMLFrameElementBase::parseMappedAttribute [HTMLFrameElementBase.cpp : 116 + 0x16]
rbx = 0x0000000000000001 r12 = 0x0000000004280e90
r13 = 0x0000000004280e90 r14 = 0x0000000000000000
r15 = 0x0000000004280e90 rip = 0x000000000077b5c7
rsp = 0x00007ffff754d960 rbp = 0x00007f8ef2ea3700
Found by: call frame info
99 phantomjs!WebCore::HTMLIFrameElement::parseMappedAttribute [HTMLIFrameElement.cpp : 131 + 0x4]
rbx = 0x0000000000000001 r12 = 0x00007f8ef2ea3700
r13 = 0x0000000004280e90 r14 = 0x0000000000000000
r15 = 0x0000000004280e90 rip = 0x00000000010e170f
rsp = 0x00007ffff754d9f0 rbp = 0x00007f8ef2ea3700
Found by: call frame info
100 phantomjs!WebCore::StyledElement::attributeChanged [StyledElement.cpp : 188 + 0xf]
rbx = 0x0000000000000001 r12 = 0x00007f8ef2ea3700
r13 = 0x0000000004280e90 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x00000000010734a4
rsp = 0x00007ffff754da50 rbp = 0x0000000000000001
Found by: call frame info
101 phantomjs!WebCore::Element::setAttribute [Element.cpp : 692 + 0x15]
rbx = 0x00007f8ef2ea3700 r12 = 0x0000000000000005
r13 = 0x00007f8f3b42aad0 r14 = 0x0000000000000000
r15 = 0x0000000002a98e38 rip = 0x0000000000634762
rsp = 0x00007ffff754daf0 rbp = 0x0000000000000002
Found by: call frame info
102 phantomjs!WebCore::jsElementPrototypeFunctionSetAttribute [JSElement.cpp : 1614 + 0x1a]
rbx = 0x000000000000000a r12 = 0x00007f8f3b7e11b8
r13 = 0x00007f8ef22e57e0 r14 = 0x00007f8f3b9f4790
r15 = 0x0000000000000001 rip = 0x0000000000c2def4
rsp = 0x00007ffff754db90 rbp = 0x00007f8f3b9f47a8
Found by: call frame info
103 0x7f8ef40001e7
rbx = 0x00007f8ef29feed0 r12 = 0x0000000000019d40
r13 = 0x00007f8f3b9f4718 r14 = 0xffff000000000000
r15 = 0xffff000000000002 rip = 0x00007f8ef40001e8
rsp = 0x00007ffff754dc60 rbp = 0x00007ffff754dd10
Found by: call frame info
104 libstdc++.so.6.0.13 + 0xcb6bc
rip = 0x00007f8f3e20a6bd rsp = 0x00007ffff754dca0
rbp = 0x00007ffff754dd10
Found by: stack scanning
105 phantomjs!JSC::Interpreter::execute [JITCode.h : 77 + 0x7]
rip = 0x0000000000f2c1eb rsp = 0x00007ffff754dd20
Found by: stack scanning
106 phantomjs!JSC::evaluate [Completion.cpp : 64 + 0x1b]
rbx = 0x00007f8f3c62b868 r12 = 0x00007f8ef2a52650
r13 = 0x00007ffff754dfe8 r14 = 0x00007f8f3efa8940
r15 = 0x00007f8f3b606e60 rip = 0x00000000004f9538
rsp = 0x00007ffff754dda0 rbp = 0x00007f8f3efa8958
Found by: call frame info
107 phantomjs!WebCore::ScriptController::evaluateInWorld [JSMainThreadExecState.h : 54 + 0x4]
rbx = 0x0000000002a9ed10 r12 = 0x0000000000000000
r13 = 0x00007f8f3aed9f18 r14 = 0x00007f8f3aed9b00
r15 = 0x00007f8f3b5e25d0 rip = 0x00000000005291da
rsp = 0x00007ffff754de50 rbp = 0x00007f8f3efa8958
Found by: call frame info
108 phantomjs!WebCore::ScriptController::evaluate [ScriptController.cpp : 167 + 0x15]
rbx = 0x00007ffff754dfe0 r12 = 0x00007f8f3aed9f18
r13 = 0x00007ffff754df60 r14 = 0x0000000000000000
r15 = 0x0000000000000001 rip = 0x00000000005294d2
rsp = 0x00007ffff754df40 rbp = 0x00007ffff754e040
Found by: call frame info
109 phantomjs!WebCore::ScriptController::executeScript [ScriptControllerBase.cpp : 64 + 0xd]
rbx = 0x00007f8f3aed9f18 r12 = 0x00007ffff754dfe0
r13 = 0x00007f8f3aed9b00 r14 = 0x0000000000000000
r15 = 0x0000000000000001 rip = 0x000000000051a475
rsp = 0x00007ffff754df60 rbp = 0x00007ffff754e040
Found by: call frame info
110 phantomjs!QWebFrame::evaluateJavaScript [qwebframe.cpp : 1556 + 0x23]
rbx = 0x00007ffff754dfe0 r12 = 0x0000000000000001
r13 = 0x00007f8ef28348d0 r14 = 0x00007ffff754e0e0
r15 = 0x00007f8f3aed9f18 rip = 0x000000000047fb3c
rsp = 0x00007ffff754dfc0 rbp = 0x0000000000000001
Found by: call frame info
111 phantomjs!WebPage::evaluateJavaScript [webpage.cpp : 683 + 0x13]
rbx = 0x00007f8eec08f8c0 r12 = 0x00007f8eec08f8d0
r13 = 0x00007ffff754e240 r14 = 0x00007ffff754e0f0
r15 = 0x000000000401ec00 rip = 0x000000000041f1b9
rsp = 0x00007ffff754e0e0 rbp = 0x00007ffff754e170
Found by: call frame info
112 phantomjs!WebPage::qt_static_metacall [moc_webpage.cpp : 224 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aa42
rsp = 0x00007ffff754e240 rbp = 0x00007ffff754ec00
Found by: call frame info
113 phantomjs!WebPage::qt_metacall [moc_webpage.cpp : 338 + 0xe]
rbx = 0x0000000000000000 r12 = 0x000000000401ec00
r13 = 0x0000000000000011 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x000000000046aebf
rsp = 0x00007ffff754e370 rbp = 0x00007ffff754ec00
Found by: call frame info
114 phantomjs!JSC::Bindings::QtRuntimeMetaMethod::call [qt_runtime.cpp : 1455 + 0x17]
rbx = 0x00007ffff754e9d0 r12 = 0x00007ffff754ec00
r13 = 0x00007f8ef29b1390 r14 = 0x0000000000000001
r15 = 0x0000000001d5fd40 rip = 0x0000000000552f07
rsp = 0x00007ffff754e500 rbp = 0x0000000000000002
Found by: call frame info
115 phantomjs!WTF::fastMalloc [FastMalloc.cpp : 3063 + 0xa]
rip = 0x00000000004e0a40 rsp = 0x00007ffff754e5f0
Found by: stack scanning
Thread 1
0 libc-2.11.3.so + 0xa0c5d
rbx = 0x00007f8f3cf0cd80 r12 = 0x00000000ffffffff
r13 = 0x00007f8f3b5ec200 r14 = 0x0000000002a7e1a0
r15 = 0x0000000002a7e170 rip = 0x00007f8f3d9e5c5d
rsp = 0x00007f8f3cf0cc50 rbp = 0x00007f8f3cf0cd00
Found by: given as instruction pointer in context
1 libc-2.11.3.so + 0xa0acf
rip = 0x00007f8f3d9e5ad0 rsp = 0x00007f8f3cf0cc60
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
2 phantomjs + 0x18b5c6f
rip = 0x0000000001cb5c70 rsp = 0x00007f8f3cf0cc68
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
3 ld-2.11.3.so + 0x9009
rip = 0x00007f8f3ef4100a rsp = 0x00007f8f3cf0cca0
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
4 ld-2.11.3.so + 0x982b
rip = 0x00007f8f3ef4182c rsp = 0x00007f8f3cf0cce0
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
5 libpthread-2.11.3.so + 0xefef
rip = 0x00007f8f3e461ff0 rsp = 0x00007f8f3cf0ccf8
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
6 ld-2.11.3.so + 0x9af1
rip = 0x00007f8f3ef41af2 rsp = 0x00007f8f3cf0cd00
rbp = 0x00007f8f3cf0cd00
Found by: stack scanning
7 libc-2.11.3.so + 0xe217
rip = 0x00007f8f3d953218 rsp = 0x00007f8f3cf0cd48
Found by: stack scanning
8 phantomjs!WebCore::SerializedScriptValue::create [WTFThreadData.h : 153 + 0x6]
rip = 0x0000000000536632 rsp = 0x00007f8f3cf0ce10
Found by: stack scanning
9 libpthread-2.11.3.so + 0x10fbf
rbx = 0x00007f8f3e463fc0 r12 = 0x00000000000002d9
r13 = 0x00007f8f3cf0d700 r14 = 0x0000000000000000
r15 = 0x0000000000000000 rip = 0x00007f8f3e463fc0
rsp = 0x00007f8f3cf0cea0 rbp = 0x0000000000000000
Found by: call frame info
10 libpthread-2.11.3.so + 0x68c9
rip = 0x00007f8f3e4598ca rsp = 0x00007f8f3cf0ced0
Found by: stack scanning
11 libpthread-2.11.3.so + 0x10fbf
rip = 0x00007f8f3e463fc0 rsp = 0x00007f8f3cf0cf78
Found by: stack scanning
12 libc-2.11.3.so + 0xcfb6c
rip = 0x00007f8f3da14b6d rsp = 0x00007f8f3cf0cfe0
Found by: stack scanning
Thread 2
0 libc-2.11.3.so + 0xc93e3
rbx = 0x0000000000000000 r12 = 0x00007f8f3c5f3da0
r13 = 0x0000000000000008 r14 = 0x0000000003e886c8
r15 = 0x0000000003e88960 rip = 0x00007f8f3da0e3e3
rsp = 0x00007f8f3c5f3b70 rbp = 0x0000000003e88510
Found by: given as instruction pointer in context
1 phantomjs!qt_safe_select [qcore_unix.cpp : 83 + 0x15]
rip = 0x0000000001d06327 rsp = 0x00007f8f3c5f3b80
Found by: stack scanning
2 phantomjs!QEventDispatcherUNIXPrivate::doSelect [qeventdispatcher_unix.cpp : 223 + 0xd]
rbx = 0x0000000003e8e3f0 r12 = 0x00007f8f3c5f3da0
r13 = 0x0000000003e88510 r14 = 0x0000000000000000
r15 = 0x0000000003e88960 rip = 0x0000000001d075e7
rsp = 0x00007f8f3c5f3c10 rbp = 0x0000000003e88510
Found by: call frame info
3 phantomjs!QEventDispatcherUNIX::processEvents [qeventdispatcher_unix.cpp : 937 + 0x4]
rbx = 0x0000000003e8e3f0 r12 = 0x00007f8f3c5f3da0
r13 = 0x0000000003eafc90 r14 = 0x0000000000000001
r15 = 0x0000000004009178 rip = 0x0000000001d08092
rsp = 0x00007f8f3c5f3d50 rbp = 0x0000000003e88510
Found by: call frame info
4 phantomjs!QEventLoop::processEvents [qeventloop.cpp : 149 + 0x1]
rbx = 0x0000000003e8e3f0 r12 = 0x00007f8f3c5f3e30
r13 = 0x00007f8f3c5f3e40 r14 = 0x0000000003e47230
r15 = 0x0000000004009178 rip = 0x0000000001cdd4b5
rsp = 0x00007f8f3c5f3da0 rbp = 0x00007f8f3c5f3de0
Found by: call frame info
5 phantomjs!QEventLoop::exec [qeventloop.cpp : 200 + 0x4]
rbx = 0x0000000003e8e3f0 r12 = 0x00007f8f3c5f3e30
r13 = 0x00007f8f3c5f3e40 r14 = 0x0000000003e47230
r15 = 0x0000000004009178 rip = 0x0000000001cdd918
rsp = 0x00007f8f3c5f3dd0 rbp = 0x00007ffff7550390
Found by: call frame info
6 phantomjs!QThread::exec [qthread.cpp : 542 + 0xc]
rbx = 0x00000000040090f0 r12 = 0x0000000004009178
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001bdfdf6
rsp = 0x00007f8f3c5f3e30 rbp = 0x0000000004009178
Found by: call frame info
7 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x0000000004009178 r12 = 0x0000000003e8d190
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8f3c5f3e80 rbp = 0x0000000003ff0220
Found by: call frame info
8 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007ffff754f870
r13 = 0x00007f8f3c5f4700 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8f3c5f3ed0 rbp = 0x0000000000000000
Found by: call frame info
Thread 3
0 libc-2.11.3.so + 0xc93e3
rbx = 0x00007f8f3ad90d50 r12 = 0x0000000001b389ff
r13 = 0x0000000000000038 r14 = 0x00007f8f3400c6a8
r15 = 0x00007f8f3400c940 rip = 0x00007f8f3da0e3e3
rsp = 0x00007f8f3ad90b70 rbp = 0x0000000000013d14
Found by: given as instruction pointer in context
1 phantomjs!qt_safe_select [qcore_unix.cpp : 93 + 0x16]
rip = 0x0000000001d06228 rsp = 0x00007f8f3ad90b80
Found by: stack scanning
2 phantomjs!QEventDispatcherUNIXPrivate::doSelect [qeventdispatcher_unix.cpp : 223 + 0xd]
rbx = 0x00007f8f3400cec0 r12 = 0x00007f8f3ad90da0
r13 = 0x00007f8f3400c4f0 r14 = 0x00007f8f3ad90d50
r15 = 0x00007f8f3400c940 rip = 0x0000000001d075e7
rsp = 0x00007f8f3ad90c10 rbp = 0x00007f8f3400c4f0
Found by: call frame info
3 phantomjs!QEventDispatcherUNIX::processEvents [qeventdispatcher_unix.cpp : 937 + 0x4]
rbx = 0x00007f8f3400cec0 r12 = 0x00007f8f3ad90da0
r13 = 0x00007f8f3400c4d0 r14 = 0x0000000000000001
r15 = 0x0000000004062c28 rip = 0x0000000001d08092
rsp = 0x00007f8f3ad90d50 rbp = 0x00007f8f3400c4f0
Found by: call frame info
4 phantomjs!QEventLoop::processEvents [qeventloop.cpp : 149 + 0x1]
rbx = 0x00007f8f3400cec0 r12 = 0x00007f8f3ad90e30
r13 = 0x00007f8f3ad90e40 r14 = 0x0000000003e47230
r15 = 0x0000000004062c28 rip = 0x0000000001cdd4b5
rsp = 0x00007f8f3ad90da0 rbp = 0x00007f8f3ad90de0
Found by: call frame info
5 phantomjs!QEventLoop::exec [qeventloop.cpp : 200 + 0x4]
rbx = 0x00007f8f3400cec0 r12 = 0x00007f8f3ad90e30
r13 = 0x00007f8f3ad90e40 r14 = 0x0000000003e47230
r15 = 0x0000000004062c28 rip = 0x0000000001cdd918
rsp = 0x00007f8f3ad90dd0 rbp = 0x00007ffff7550390
Found by: call frame info
6 phantomjs!QThread::exec [qthread.cpp : 542 + 0xc]
rbx = 0x0000000004062ba0 r12 = 0x0000000004062c28
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001bdfdf6
rsp = 0x00007f8f3ad90e30 rbp = 0x0000000004062c28
Found by: call frame info
7 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x0000000004062c28 r12 = 0x0000000004062ce0
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8f3ad90e80 rbp = 0x0000000004062b80
Found by: call frame info
8 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007ffff754c020
r13 = 0x00007f8f3ad91700 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8f3ad90ed0 rbp = 0x0000000000000000
Found by: call frame info
Thread 4
0 libpthread-2.11.3.so + 0xb569
rbx = 0x00007f8f34013640 r12 = 0x0000000000000031
r13 = 0x00007f8f3a58fde0 r14 = 0x00007f8f34013668
r15 = 0x0000000000000000 rip = 0x00007f8f3e45e569
rsp = 0x00007f8f3a58fd68 rbp = 0x00007f8f34013970
Found by: given as instruction pointer in context
1 linux-gate.so + 0x7ac
rip = 0x00007ffff75ff7ad rsp = 0x00007f8f3a58fd90
Found by: stack scanning
2 phantomjs!QWaitCondition::wait [qwaitcondition_unix.cpp : 84 + 0x7]
rip = 0x0000000001be39d7 rsp = 0x00007f8f3a58fdc0
Found by: stack scanning
3 phantomjs!QThreadPoolThread::run [qthreadpool.cpp : 141 + 0x1d]
rbx = 0x00007f8f34013560 r12 = 0x00007f8f340135e9
r13 = 0x0000000000000002 r14 = 0x00007f8f34041000
r15 = 0x0000000000000003 rip = 0x0000000001bd783d
rsp = 0x00007f8f3a58fe40 rbp = 0x00007f8f34013970
Found by: call frame info
4 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x0000000004014c40 r12 = 0x00007f8f34013b10
r13 = 0x0000000000000001 r14 = 0x00007f8f34013cf0
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8f3a58fe80 rbp = 0x00007f8f34013970
Found by: call frame info
5 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007f8f3ad8f630
r13 = 0x00007f8f3a590700 r14 = 0x00007f8f34013cf0
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8f3a58fed0 rbp = 0x0000000000000000
Found by: call frame info
Thread 5
0 libpthread-2.11.3.so + 0xb569
rbx = 0x00007f8f34013640 r12 = 0x0000000000000033
r13 = 0x00007f8f39760de0 r14 = 0x00007f8f34013668
r15 = 0x0000000000000000 rip = 0x00007f8f3e45e569
rsp = 0x00007f8f39760d68 rbp = 0x00007f8eec010e20
Found by: given as instruction pointer in context
1 linux-gate.so + 0x7ac
rip = 0x00007ffff75ff7ad rsp = 0x00007f8f39760d90
Found by: stack scanning
2 phantomjs!QWaitCondition::wait [qwaitcondition_unix.cpp : 84 + 0x7]
rip = 0x0000000001be39d7 rsp = 0x00007f8f39760dc0
Found by: stack scanning
3 phantomjs!QThreadPoolThread::run [qthreadpool.cpp : 141 + 0x1d]
rbx = 0x00007f8f34013560 r12 = 0x00007f8f340135e9
r13 = 0x0000000000000002 r14 = 0x00007f8f34072f00
r15 = 0x0000000000000003 rip = 0x0000000001bd783d
rsp = 0x00007f8f39760e40 rbp = 0x00007f8eec010e20
Found by: call frame info
4 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x00000000040c48f0 r12 = 0x00007f8eec011120
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8f39760e80 rbp = 0x00007f8eec010e20
Found by: call frame info
5 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007f8f3ad8f5b0
r13 = 0x00007f8f39761700 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8f39760ed0 rbp = 0x0000000000000000
Found by: call frame info
Thread 6
0 libpthread-2.11.3.so + 0xb569
rbx = 0x00007f8f34013640 r12 = 0x0000000000000030
r13 = 0x00007f8f38f58de0 r14 = 0x00007f8f34013668
r15 = 0x0000000000000000 rip = 0x00007f8f3e45e569
rsp = 0x00007f8f38f58d68 rbp = 0x00007f8eec015f10
Found by: given as instruction pointer in context
1 linux-gate.so + 0x7ac
rip = 0x00007ffff75ff7ad rsp = 0x00007f8f38f58d90
Found by: stack scanning
2 phantomjs!QWaitCondition::wait [qwaitcondition_unix.cpp : 84 + 0x7]
rip = 0x0000000001be39d7 rsp = 0x00007f8f38f58dc0
Found by: stack scanning
3 phantomjs!QThreadPoolThread::run [qthreadpool.cpp : 141 + 0x1d]
rbx = 0x00007f8f34013560 r12 = 0x00007f8f340135e9
r13 = 0x0000000000000002 r14 = 0x00007f8f3405ea80
r15 = 0x0000000000000003 rip = 0x0000000001bd783d
rsp = 0x00007f8f38f58e40 rbp = 0x00007f8eec015f10
Found by: call frame info
4 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x00000000040c79f0 r12 = 0x00007f8eec016300
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8f38f58e80 rbp = 0x00007f8eec015f10
Found by: call frame info
5 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007f8f3ad8f5b0
r13 = 0x00007f8f38f59700 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8f38f58ed0 rbp = 0x0000000000000000
Found by: call frame info
Thread 7
0 libpthread-2.11.3.so + 0xb1fc
rbx = 0x000000000411a350 r12 = 0x004189374bc6a7ef
r13 = 0xffffffffffffffff r14 = 0x000000000411a378
r15 = 0x000000000411a190 rip = 0x00007f8f3e45e1fc
rsp = 0x00007f8ef3ffed38 rbp = 0x00007f8f3b211190
Found by: given as instruction pointer in context
1 phantomjs!QWaitCondition::wait [qwaitcondition_unix.cpp : 86 + 0xa]
rip = 0x0000000001be39f5 rsp = 0x00007f8ef3ffed60
rbp = 0x00007f8f3b211190
Found by: stack scanning
2 phantomjs!WTF::ThreadCondition::timedWait [ThreadingQt.cpp : 253 + 0x12]
rbx = 0x00007f8f3b211188 r12 = 0x0000000000000003
r13 = 0x00007f8f3b211188 r14 = 0x00007f8f3b211190
r15 = 0x00007f8ef3ffee40 rip = 0x00000000004e70ea
rsp = 0x00007f8ef3ffede0 rbp = 0x00007f8f3b211190
Found by: call frame info
3 phantomjs!WTF::MessageQueue<WebCore::LocalStorageTask>::waitForMessage [MessageQueue.h : 137 + 0x12]
rbx = 0x0000000000000003 r12 = 0x0000000000000003
r13 = 0x00007f8f3b211188 r14 = 0x00007f8f3b211190
r15 = 0x00007f8ef3ffee40 rip = 0x0000000001185269
rsp = 0x00007f8ef3ffee00 rbp = 0x00007f8f3b211198
Found by: call frame info
4 phantomjs!WebCore::LocalStorageThread::threadEntryPoint [LocalStorageThread.cpp : 68 + 0xa]
rbx = 0x00007f8ef2e28c80 r12 = 0x000000000411aeb0
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001185493
rsp = 0x00007f8ef3ffee40 rbp = 0x00007f8f3b211188
Found by: call frame info
5 phantomjs!WTF::ThreadPrivate::run [ThreadingQt.cpp : 66 + 0x6]
rbx = 0x000000000411a630 r12 = 0x000000000411aeb0
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00000000004e68fb
rsp = 0x00007f8ef3ffee70 rbp = 0x000000000411a630
Found by: call frame info
6 phantomjs!QThreadPrivate::start [qthread_unix.cpp : 338 + 0x9]
rbx = 0x000000000411adf8 r12 = 0x000000000411aeb0
r13 = 0x0000000000000001 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x0000000001be2cc2
rsp = 0x00007f8ef3ffee80 rbp = 0x000000000411a630
Found by: call frame info
7 libpthread-2.11.3.so + 0x68c9
rbx = 0x0000000000000000 r12 = 0x00007ffff754eb50
r13 = 0x00007f8ef3fff700 r14 = 0x00007f8f3f156040
r15 = 0x0000000000000003 rip = 0x00007f8f3e4598ca
rsp = 0x00007f8ef3ffeed0 rbp = 0x0000000000000000
Found by: call frame info
Loaded modules:
0x00400000 - 0x0268dfff phantomjs ??? (main)
0x7f8f380bf000 - 0x7f8f380cdfff VeraSe.ttf ???
0x7f8f381ea000 - 0x7f8f381f5fff VeraMoBd.ttf ???
0x7f8f3868f000 - 0x7f8f3869ffff Vera.ttf ???
0x7f8f386a0000 - 0x7f8f386aefff VeraBd.ttf ???
0x7f8f39762000 - 0x7f8f39967fff libnss_dns-2.11.3.so ???
0x7f8f39968000 - 0x7f8f39b74fff libnss_files-2.11.3.so ???
0x7f8f39b75000 - 0x7f8f39d88fff libresolv-2.11.3.so ???
0x7f8f3b5cf000 - 0x7f8f3b5dbfff VeraMono.ttf ???
0x7f8f3cf0e000 - 0x7f8f3d164fff libssl.so.0.9.8 ???
0x7f8f3d165000 - 0x7f8f3d501fff libcrypto.so.0.9.8 ???
0x7f8f3d506000 - 0x7f8f3d72dfff libexpat.so.1.5.2 ???
0x7f8f3d72e000 - 0x7f8f3d944fff libz.so.<IP_ADDRESS> ???
0x7f8f3d945000 - 0x7f8f3dca1fff libc-2.11.3.so ???
0x7f8f3dca7000 - 0x7f8f3debcfff libgcc_s.so.1 ???
0x7f8f3debd000 - 0x7f8f3e13efff libm-2.11.3.so ???
0x7f8f3e13f000 - 0x7f8f3e43dfff libstdc++.so.6.0.13 ???
0x7f8f3e453000 - 0x7f8f3e66afff libpthread-2.11.3.so ???
0x7f8f3e66f000 - 0x7f8f3e876fff librt-2.11.3.so ???
0x7f8f3e877000 - 0x7f8f3ea7afff libdl-2.11.3.so ???
0x7f8f3ea7b000 - 0x7f8f3ecaffff libfontconfig.so.1.4.4 ???
0x7f8f3ecb0000 - 0x7f8f3ef37fff libfreetype.so.6.6.0 ???
0x7f8f3ef38000 - 0x7f8f3ef55fff ld-2.11.3.so ???
0x7f8f3efd2000 - 0x7f8f3f146fff locale-archive ???
0x7f8f3f150000 - 0x7f8f3f152fff de156ccd2eddbdc19d37a45b8b2aac9c-le64.cache-3 ???
0x7ffff75ff000 - 0x7ffff75fffff linux-gate.so ???
I'm building from source and will see if the patch applies and resolves the situation.
I expect that the patch did not resolve the situation. Am I right?
Fixed in 2.1.1
|
2025-04-01T06:37:57.300833
| 2016-06-26T01:23:33
|
162307551
|
{
"authors": [
"3zzy",
"Dingo64",
"Vitallium",
"ariya",
"arrmo",
"elhamdaoui",
"ip1981",
"istinspring",
"jglogan",
"jorgesinval",
"michaelkl",
"ricohumme",
"seachanged",
"wojciechmorawski",
"zackw"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3842",
"repo": "ariya/phantomjs",
"url": "https://github.com/ariya/phantomjs/issues/14376"
}
|
gharchive/issue
|
QXcbConnection: Could not connect to display
Which version of PhantomJS are you using? Tip: run phantomjs --version.
2.1.1
What steps will reproduce the problem?
Take a working javascript file (e.g. rasterize.js)
Call the file from a script file (e.g. executable file, #!/bin/sh at the start)
Now execute the script file as a cron job
Which operating system are you using?
Ubuntu 16.04 LTS
Did you use binary PhantomJS or did you compile it from source?
Binary - installed it as part of Ubuntu (using apt).
Please provide any additional information below.
This is very similar to https://github.com/ariya/phantomjs/issues/14240, but I am using the stock version of PhantomJS (with Ubuntu). Below is the output from running the cron job (which calls the script file, to generate a pdf file once a day) ...
QXcbConnection: Could not connect to display
PhantomJS has crashed. Please read the bug reporting guide at
http://phantomjs.org/bug-reporting.html and file a bug report.
Aborted (core dumped)
Thanks!
Please check your environment variable QT_QPA_PLATFORM. It must be empty.
Yep, it is - in both cases. Does this need to be manually forced for some reason?
Thanks!
I would send a patch to reset all the troublesome environment variables before initializing Qt ... if I could find a list of them anywhere.
But it sounds like this isn't actually the problem?
It doesn't sound like it, but I could definitely be wrong.
Trying different things, no luck. Is there a setting of QT_QPA_PLATFORM that you believe may help?
Thanks!
Hi,
Having same issue here.
Working with 1.9 on my dev machine, works fine.
But on the live server I can't even download 1.9.x so I'm stuck now.
Where can I find this settings of QT? How to get there. So I can check that on my side.
Thanks!
The offscreen platform is a usable workaround for me with the stock Ubuntu package. This works for me:
QT_QPA_PLATFORM=offscreen phantomjs rasterize.js 2i8zaNg04d9B41Zir2kT3J/output.html 2i8zaNg04d9B41Zir2kT3J.png
can confirm export QT_QPA_PLATFORM=offscreen before selenium scripts work's well.
We finally got things working by downloading the binary at https://bitbucket.org/ariya/phantomjs/downloads/phantomjs-2.1.1-linux-x86_64.tar.bz2 and extracting the phantomjs to /usr/bin
I had hard times with ubuntu 16.04 they basically distribute half finished system with lots of broken packages.
I'm running on 16.04 too, but the problem with the repo's is that they deliver with x11, which fails. So with the binary I don't have that issue and it works perfectly now.
Works here now also - thanks!!!
Debian/Ubuntu has a modified version of PhantomJS that can work headlessly, hence the problem with QXcb. See https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=817277 for details.
Unfortunately, this is not something that we (=PhantomJS team) can fix.
@ricohumme Can you provide the steps (commands) to install it? I tried:
export PHANTOM_JS_VERSION=2.1.1
wget "https://bitbucket.org/ariya/phantomjs/downloads/phantomjs-$PHANTOM_JS_VERSION-linux-x86_64.tar.bz2"
tar xvfj "./phantomjs-$PHANTOM_JS_VERSION-linux-x86_64.tar.bz2"
ln -sf "./phantomjs-$PHANTOM_JS_VERSION-linux-x86_64/bin/phantomjs" "/usr/bin"
but still couldn't get it working
@3zzy the steps I used are as follows:
wget https://bitbucket.org/ariya/phantomjs/downloads/phantomjs-2.1.1-linux-x86_64.tar.bz2
bzip2 -d phantomjs-2.1.1-linux-x86_64.tar.bz2
tar -xvf phantomjs-2.1.1-linux-x86_64.tar
cp phantomjs-2.1.1-linux-x86_64/bin/phantomjs /usr/bin/phantomjs
Does this help your case?
Thanks for help, solution above works well :)
I'm the same error but in Odoo
xvfb-run worked best for me, from Ariya's link posted 8/2016.
xvfb is "providing an unobtrusive way to run applications that don’t really need an X server but insist on having one anyway."
@jglogan Thank you for solution with QT_QPA_PLATFORM=offscreen. It let me run phantomjs itself, but it was failing for me with EOFError: end of file reached message on certain pages.
After some experiments with different platforms I found out that it works like a charm with QT_QPA_PLATFORM=minimal.
Hope this will help someone.
Thanks a lot, guys! QT_QPA_PLATFORM=offscreen works great on Debian 8 jessie amd64 with PhantomJS 2.1.1
# QT_QPA_PLATFORM=offscreen phantomjs --version
2.1.1
Debian/Ubuntu has a modified version of PhantomJS that can work headlessly, hence the problem with QXcb. See https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=817277 for details.
This is quite the opposite:
Unfortunately it can not be fixed in Debian. To achieve headless-ness
upstream statically link with customised QT + Webkit. We don't want to ship
forks of those projects. It would be great to eventually convince upstream to
use standard libraries.
There is also a suggestion to use xvfb-run, which is right.
Every time I try to plot anything with RStudio Cloud using the webshot package I obtain an error. An example:
#install webshot
library(webshot)
webshot::install_phantomjs()
# Make the graph
my_graph=wordcloud2(demoFreq, size=1.5)
# save it in html
library("htmlwidgets")
saveWidget(my_graph,"tmp.html",selfcontained = F)
# and in png
webshot("tmp.html","fig_1.png", delay =5, vwidth = 480, vheight=480)
The error says:
QXcbConnection: Could not connect to display :0
PhantomJS has crashed. Please read the bug reporting guide at
<http://phantomjs.org/bug-reporting.html> and file a bug report.
Error in webshot("tmp.html", "fig_1.png", delay = 5, vwidth = 480, vheight = 480) :
webshot.js returned failure value: -6
Any thoughts?
|
2025-04-01T06:37:57.349280
| 2019-11-19T23:51:28
|
525335449
|
{
"authors": [
"Prodian0013",
"awick"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3843",
"repo": "arkime/arkime",
"url": "https://github.com/arkime/arkime/issues/1316"
}
|
gharchive/issue
|
Feature Request: Add option for writing to index alias
Feature: Provide an option to allow sessions2 index to write to an index alias.
Use Case: This will allow users to manage the sessions2 index through ilm, curator or other means in elastic rather than being forced to use the built in moloch time based rotation.
Arkime already supports ilm, just not auto roll over. With performance testing we've found that auto rollover increases search time.
|
2025-04-01T06:37:57.357999
| 2024-01-18T23:43:45
|
2089288865
|
{
"authors": [
"ssalbdivad",
"yankeeinlondon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3844",
"repo": "arktypeio/arktype",
"url": "https://github.com/arktypeio/arktype/pull/905"
}
|
gharchive/pull-request
|
Fix symbolic links in Beta branch
fixed symbolics links with absolute path which had blocked portability
fixed all linting warnings (these were all pre-existing warnings)
fixed all formatting issues with prettier
The test suite still fails in a number of places but I'm fairly certain that this PR doesn't introduce any problems that weren't already there.
Hi thanks so much for this PR! There are some helpful changes in here for sure and I'm not sure which build script I have is screwing with those symlinks- I had thought I fixed it.
That said, there's a lot here and most of it can be attributed to beta being a development branch. E.g. in the cases of those unused variables, most of them are unused because of something I still need to do before I merge the branch, before which time those lint errors would have to be addressed.
If you want to create a PR that's not a straightforward bug fix (especially on a dev branch), I'd recommend running it by me first. Again, I really do appreciate the initiative to help out!
Sure no problem; i needed to do most of this to get it to build locally and figured it might be useful as a PR.
I will make a commit now fixing the symlink issue. If you run into anything else that blocks the build let me know!
|
2025-04-01T06:37:57.381933
| 2019-09-26T10:28:14
|
498799441
|
{
"authors": [
"arlac77"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3845",
"repo": "arlac77/npm-template-sync",
"url": "https://github.com/arlac77/npm-template-sync/pull/784"
}
|
gharchive/pull-request
|
mkpr
Applied mkpr on package.json
sed s/"commander": "^.*"/"commander": "^3.0.2"/
:tada: This PR is included in version 8.5.3 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:37:57.434085
| 2018-05-01T07:19:39
|
319134934
|
{
"authors": [
"arlac77",
"coveralls"
],
"license": "0BSD",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3846",
"repo": "arlac77/timeseries-sqlite2leveldb",
"url": "https://github.com/arlac77/timeseries-sqlite2leveldb/pull/98"
}
|
gharchive/pull-request
|
merge package from arlac77/npm-package-template
README.md
docs(README): update from template
Coverage remained the same at 97.143% when pulling 3f5e8b79e5609de01534b9d2ac83d3c036ece2d2 on template-sync-1 into<PHONE_NUMBER>5545a50622bc46bf6964de8f02b886 on master.
:tada: This PR is included in version 1.0.6 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:37:57.489132
| 2018-12-02T06:32:26
|
386532191
|
{
"authors": [
"Ishan1742",
"arnav-t",
"mathrulestheworld"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3848",
"repo": "arnav-t/Shooting-Game",
"url": "https://github.com/arnav-t/Shooting-Game/issues/22"
}
|
gharchive/issue
|
Prevent stopping at walls
Can make the player move in the general direction of walls, instead of stopping. When a player hits a wall.
yes the only stop happens when he hits perpendicular to the wall.and we have to deal with edges also.
sent a PR for this one
Fixed in #73
|
2025-04-01T06:37:57.516276
| 2015-05-23T22:18:05
|
79969195
|
{
"authors": [
"aron-bordin"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3851",
"repo": "aron-bordin/kivy-designer",
"url": "https://github.com/aron-bordin/kivy-designer/issues/78"
}
|
gharchive/issue
|
MainApp().run() import module
To load the python code, the app MUST be using this format:
Works:
MainApp().run()
This example doesn't work:
app = MainApp()
app.run()
https://github.com/kivy/kivy-designer/blob/master/designer/project_loader.py#L994
|
2025-04-01T06:37:57.522964
| 2021-07-21T04:51:54
|
949321533
|
{
"authors": [
"arpit456jain",
"jigar-sable"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3852",
"repo": "arpit456jain/Amazing-Js-Projects",
"url": "https://github.com/arpit456jain/Amazing-Js-Projects/issues/45"
}
|
gharchive/issue
|
Password Generator
Issue Type:
Feature Request
Describe the feature
I would like to add Password Generator project.
Go ahead @jigar-sable
|
2025-04-01T06:37:57.566277
| 2023-11-06T11:05:21
|
1978866591
|
{
"authors": [
"labmonkey",
"tlbueno"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3853",
"repo": "artemiscloud/activemq-artemis-broker-kubernetes-image",
"url": "https://github.com/artemiscloud/activemq-artemis-broker-kubernetes-image/issues/83"
}
|
gharchive/issue
|
Any plans to create a Metrics Plugin?
Hello. I was browsing this project and since the repositories are based on the equivalents from RedHat AMQ Broker I was wondering if there is any reason why the Metrics plugin is still using the one published by RedHat instead of having its own "mirrored" one in artemiscloud just like other repositories? Generally I am asking if there are plans to create this plugin repository here.
https://github.com/artemiscloud/activemq-artemis-broker-kubernetes-image/blob/89a5d7600961569c4e3eef0541d977e57d6d5777/modules/activemq-artemis-launch/added/launch.sh#L299
Hi @labmonkey
Currently it's the opposite. RedHat AMQ Broker Operator is based on ArtemisCloud project and RedHat AMQ Broker is based on Apache ActiveMQ Artemis.
Currently there is no plan to have the ArtemisPrometheusMetricsPlugin under ArtemisCloud project since the project from RedHat is open source, has Apache license and its artifacts are published in RedHat public maven repository.
|
2025-04-01T06:37:57.602279
| 2019-08-05T08:33:26
|
476726318
|
{
"authors": [
"lopopolo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3855",
"repo": "artichoke/artichoke",
"url": "https://github.com/artichoke/artichoke/issues/128"
}
|
gharchive/issue
|
Rename artichoke-backend to artichoke-backend-mruby
We are getting close to the point where we can start experimenting with additional interpreter implementations.
Rename artichoke-backend to artichoke-backend-mruby to reflect this new multi-VM implementation state.
"we are getting close" 😅
|
2025-04-01T06:37:57.603265
| 2021-05-25T06:33:36
|
900339592
|
{
"authors": [
"shayb-datumate"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3856",
"repo": "articodeltd/angular-cesium",
"url": "https://github.com/articodeltd/angular-cesium/pull/385"
}
|
gharchive/pull-request
|
Fix update moving point on editors that don't work properly in tablets
Update editors "moving point" when adding a new point. This is done because if the client doesn't have a mouse, the moving point doesn't update, which can cause multiple issues in creating map entities.
Thank you very much 🙏🏻
|
2025-04-01T06:37:57.619517
| 2015-08-01T13:45:53
|
98538771
|
{
"authors": [
"alenofx",
"artman"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3857",
"repo": "artman/Signals",
"url": "https://github.com/artman/Signals/issues/6"
}
|
gharchive/issue
|
Missing Podspec in repo
Please add the Podspec to the repo (useful to fetch a different branch directly from the Podfile without using a local repo, e.g. the "swift-2" branch).
Thank you.
Thanks for noticing. Fixed.
|
2025-04-01T06:37:57.623589
| 2021-07-12T17:35:00
|
942288291
|
{
"authors": [
"aarongustafson",
"artstorm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3858",
"repo": "artstorm/eleventy-plugin-seo",
"url": "https://github.com/artstorm/eleventy-plugin-seo/pull/32"
}
|
gharchive/pull-request
|
Config option to enable page numbers to be excluded from the title
Added tests and documentation.
All tests (render + Liquid + Nunjucks) are passing and I have tested the Nunjucks implementation on my own site as well using the global config and the front matter override.
@aarongustafson I've reviewed your PR. Excellent work! 👍🏻 It's a really convenient and nice addition to the plugin. I'll have it merged and publish a new release of the plugin.
|
2025-04-01T06:37:57.626595
| 2015-01-12T16:51:39
|
54077228
|
{
"authors": [
"broskoski",
"craigspaeth",
"dzucconi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3859",
"repo": "artsy/2014.artsy.net",
"url": "https://github.com/artsy/2014.artsy.net/pull/5"
}
|
gharchive/pull-request
|
@broskoski: Simplest possible swiping carousel
There's still some issues with this I'm working out but it works alright. Would gladly swap this out with a library if I could find one that is halfway decent (suggestions?).
:+1: looking good, can I use it in microgravity?
This is interesting: http://dev.w3.org/csswg/css-snappoints/
There were some component-js swipe components that looked pretty good http://component.github.io/?q=swipe. Although I can't personally vouch for any of them. Maybe this is an opportunity to release our own open source swipe component :)
|
2025-04-01T06:37:57.629254
| 2020-10-01T16:25:56
|
712996383
|
{
"authors": [
"abhitip",
"artsyit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3860",
"repo": "artsy/cohesion",
"url": "https://github.com/artsy/cohesion/pull/108"
}
|
gharchive/pull-request
|
Adds article screen owner type
This adds article and articles as a screen owner type.
:rocket: PR was released in v1.40.0 :rocket:
|
2025-04-01T06:37:57.631261
| 2024-07-23T22:28:31
|
2426225556
|
{
"authors": [
"ArtsyOpenSource",
"artsyit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3861",
"repo": "artsy/eigen",
"url": "https://github.com/artsy/eigen/pull/10511"
}
|
gharchive/pull-request
|
chore: update metaphysics graphql schema
Greetings human :robot: this PR was automatically created as part of metaphysics' deploy process. #nochangelog
Warnings
:warning:
✅ No changelog changes
Generated by :no_entry_sign: dangerJS against 89316bd2f23f778ebe74d35c04421c396a8f3573
|
2025-04-01T06:37:57.636523
| 2017-02-13T17:47:11
|
207290331
|
{
"authors": [
"ArtsyOpenSource",
"alloy",
"sarahscott"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3862",
"repo": "artsy/eigen",
"url": "https://github.com/artsy/eigen/pull/2169"
}
|
gharchive/pull-request
|
[WIP] WorksForYou QA Items
Closes https://github.com/artsy/eigen/issues/1572
[x] Use sale_message from gravity instead of native logic
[ ] Address margins/layout issues
[ ] Fix tests
Regarding the sale message: institutional works don't have them in gravity, so the current pr work looks like this (note the Not For Sale work on the right and the price-label-less work on the left):
This is what eigen currently shows (no label at all for sold or institutional works):
I will add a 'Not For Sale' message to institutional works so they aren't left looking like something is missing from them for now, but perhaps it's something that would make sense in gravity (i.e. not ever leaving the sale_message empty and returning 'Not For Sale' if no info is provided)
Fixes #1572
1 Warning
:warning:
PR is classed as Work in Progress
Generated by :no_entry_sign: danger
I can definitely address the single-artwork notification QA now, but for the masonry layout in general, I need to do a bigger refactor than I thought.
@alloy looks good on-device 👍
Superkalifragilistic 👌
|
2025-04-01T06:37:57.643989
| 2022-02-03T13:55:29
|
1123115772
|
{
"authors": [
"ArtsyOpenSource",
"brainbicycle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3863",
"repo": "artsy/eigen",
"url": "https://github.com/artsy/eigen/pull/6155"
}
|
gharchive/pull-request
|
fix: match iOS sentry release name in fastfile #trivial
The type of this PR is: fix
Description
I changed a '+' to a '-'.
I'll fix this manually for the outgoing builds 🙄
PR Checklist (tick all before merging)
[x] I have included screenshots or videos to illustrate my changes, or I have not changed anything that impacts the UI.
[x] I have tested my changes on iOS and Android.
[x] I have added tests/stories for my changes, or my changes don't require testing/stories, or I have included a link to a separate Jira ticket covering the tests.
[x] I have added a feature flag, or my changes don't require a feature flag. (How do I add one?)
[x] I have documented any follow-up work that this PR will require, or it does not require any.
[x] I have added an app state migration, or my changes do not require one. (What are migrations?)
[x] I have added a changelog entry below or my changes do not require one.
To the reviewers 👀
[ ] I would like at least one of the reviewers to run this PR on the simulator or device.
Changelog updates
Changelog updates
Cross-platform user-facing changes
iOS user-facing changes
Android user-facing changes
Dev changes
Fix sentry release name on iOS - Brian
This PR contains the following changes:
Dev changes (Fix sentry release name on iOS - Brian)
Generated by :no_entry_sign: dangerJS against dec6f1478be64dc2fed650b5746d318cdb1c7a38
|
2025-04-01T06:37:57.654843
| 2023-04-12T15:54:54
|
1664829550
|
{
"authors": [
"MounirDhahri",
"dariakoko",
"olerichter00"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3864",
"repo": "artsy/eigen",
"url": "https://github.com/artsy/eigen/pull/8498"
}
|
gharchive/pull-request
|
feat(CX-3598): add new onboarding cards to home screen
This PR resolves CX-3598
Description
https://user-images.githubusercontent.com/36167539/231513604-9d69b90b-cf7e-4707-8009-69b24ba3f216.mp4
PR Checklist
[x] I have tested my changes on iOS and Android.
[x] I hid my changes behind a feature flag, or they don't need one.
[x] I have included screenshots or videos, or I have not changed the UI.
[ ] I have added tests, or my changes don't require any.
[ ] I added an app state migration, or my changes do not require one.
[ ] I have documented any follow-up work that this PR will require, or it does not require any.
[ ] I have added a changelog entry below, or my changes do not require one.
To the reviewers 👀
[ ] I would like at least one of the reviewers to run this PR on the simulator or device.
Changelog updates
Changelog updates
Cross-platform user-facing changes
added new Artsy onboarding rail to the home screen -daria
iOS user-facing changes
Android user-facing changes
Dev changes
Need help with something? Have a look at our docs, or get in touch with us.
@olerichter00 You mean the images themselves, not the way we display them in the component? Not sure, I downloaded them from figma and they have the same size as the ones on the old cards. You would suggest reducing the size?
@olerichter00 You mean the images themselves, not the way we display them in the component? Not sure, I downloaded them from figma and they have the same size as the ones on the old cards. You would suggest reducing the size?
Ok, let's add them as they are. I remember @MounirDhahri was working on the image sizes in Eigen a while ago. @MounirDhahri, do you think it is possible and needed to reduce the size of the images?
@MounirDhahri @olerichter00 maybe we could document somewhere the optimal image sizes? Do we have a readme for similar purposes?
I don't think it's possible to document that in numbers but it's probably a good idea to write this somewhere in our docs. The rationale behind it comes from some studies that emerged a few years ago about how app size matters. Overall, the smaller the app size, the more people convert into installing the app. In our case, since we have a fair amount of images in the app, our bundle size has been increasing quickly and it's important for us to try to cut the image size to avoid affecting our conversion
|
2025-04-01T06:37:57.663246
| 2023-07-25T09:49:51
|
1819948048
|
{
"authors": [
"ArtsyOpenSource",
"gkartalis"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3865",
"repo": "artsy/eigen",
"url": "https://github.com/artsy/eigen/pull/9048"
}
|
gharchive/pull-request
|
build(deps): update flipper and dev deps
This PR resolves []
Description
Updates flipper dep and some other deps 🐙
PR Checklist
[x] I have tested my changes on iOS and Android.
[x] I hid my changes behind a feature flag, or they don't need one.
[x] I have included screenshots or videos, or I have not changed the UI.
[x] I have added tests, or my changes don't require any.
[x] I added an app state migration, or my changes do not require one.
[x] I have documented any follow-up work that this PR will require, or it does not require any.
[x] I have added a changelog entry below, or my changes do not require one.
To the reviewers 👀
[ ] I would like at least one of the reviewers to run this PR on the simulator or device.
Changelog updates
Changelog updates
Cross-platform user-facing changes
iOS user-facing changes
Android user-facing changes
Dev changes
update flipper and dev deps - gkartalis
Need help with something? Have a look at our docs, or get in touch with us.
This PR contains the following changes:
Dev changes (update flipper and dev deps - gkartalis)
Generated by :no_entry_sign: dangerJS against 9a64039a79e01056bee93adab9616000291b91d8
|
2025-04-01T06:37:57.667380
| 2022-12-21T21:34:26
|
1506939654
|
{
"authors": [
"damassi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3866",
"repo": "artsy/energy",
"url": "https://github.com/artsy/energy/pull/395"
}
|
gharchive/pull-request
|
fix(offline): Add missing show queries to syncManager
Description
This adds a few missing queries to the syncManager:
ShowArtworks, showArtworksQuery ($slug: String!, $imageSize: Int!)
ShowInstalls, showInstallsQuery ($slug: String!, $imageSize: Int!)
ShowDocuments, showDocumentsQuery ($slug: String!, $partnerID: String!)
cc @artsy/mobile-platform
PR Checklist
[ ] I tested my changes on iOS and Android.
[ ] I added screenshots or videos to illustrate my changes.
[ ] I added Tests and Stories for my changes.
To the reviewers 👀
[ ] I would like at least one of the reviewers to run this PR on the simulator or device.
Need help with something? Have a look at our docs, or get in touch with us.
This is pretty trivial so going to merge and if anyone has any feedback I can loop back tomorrow and address
|
2025-04-01T06:37:57.668673
| 2019-05-21T16:45:07
|
446721057
|
{
"authors": [
"eessex",
"sweir27"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3867",
"repo": "artsy/force",
"url": "https://github.com/artsy/force/pull/4074"
}
|
gharchive/pull-request
|
[WIP] Use /auth2 for Consign signup/login flow
Relates to work to add reCAPTCHA to all auth forms, replaces the one-off signup/login implementation for /consign with the new auth components.
WIP while I confirm that analytics is working as expected.
😍 This looks so great!!! So happy to see all that deleted code.
|
2025-04-01T06:37:57.669718
| 2018-03-13T18:18:10
|
304880514
|
{
"authors": [
"anandaroop",
"ashfurrow"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3868",
"repo": "artsy/meta",
"url": "https://github.com/artsy/meta/pull/20"
}
|
gharchive/pull-request
|
Adds March 13 notes, makes Show&Tell optional
This adds the three things we saw today, as well as marks the meeting as optional by our engineers.
👍 thanks!
|
2025-04-01T06:37:57.686748
| 2022-07-29T05:00:25
|
1321762308
|
{
"authors": [
"codecov-commenter",
"noahziheng"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3869",
"repo": "artusjs/core",
"url": "https://github.com/artusjs/core/pull/155"
}
|
gharchive/pull-request
|
feat(scanner): filter config with env
resolve #154
临时方案,暂时解决 BUG,后续需对 Scanner 进行重构,理清递归过程
Codecov Report
Merging #155 (1addba3) into master (d9d5ee2) will increase coverage by 0.11%.
The diff coverage is 100.00%.
@@ Coverage Diff @@
## master #155 +/- ##
==========================================
+ Coverage 88.82% 88.93% +0.11%
==========================================
Files 50 51 +1
Lines 1083 1094 +11
Branches 174 177 +3
==========================================
+ Hits 962 973 +11
Misses 121 121
Impacted Files
Coverage Δ
src/loader/impl/config.ts
88.88% <100.00%> (-1.81%)
:arrow_down:
src/loader/impl/framework_config.ts
100.00% <100.00%> (ø)
src/loader/impl/plugin_config.ts
92.30% <100.00%> (+0.30%)
:arrow_up:
src/loader/utils/config_file_meta.ts
100.00% <100.00%> (ø)
src/scanner/scan.ts
94.85% <100.00%> (+0.27%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update d9d5ee2...1addba3. Read the comment docs.
|
2025-04-01T06:37:57.709942
| 2020-04-02T18:03:36
|
592825489
|
{
"authors": [
"arufian",
"jbleyleSF"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3870",
"repo": "arufian/LWC-Component-COVID19",
"url": "https://github.com/arufian/LWC-Component-COVID19/issues/2"
}
|
gharchive/issue
|
USA only accepts "US"
None of these values work:
USA
United States
United States of America
Hi @jbleyleSF
I will add these as synonyms for "U.S" in the API
I have add synonyms system to the API.
Please look at this gif.
I will close this and #4 issue.
|
2025-04-01T06:37:57.714007
| 2015-03-21T10:04:11
|
63384690
|
{
"authors": [
"marcdibold",
"thinklinux",
"xaiki"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3871",
"repo": "arunoda/meteor-up",
"url": "https://github.com/arunoda/meteor-up/pull/328"
}
|
gharchive/pull-request
|
Add suport for Debian 'the Debian way'
This enables deployments using as much debian-vodo as possible and compatible with systemd
+1
As a side note - as of 15.04 Ubuntu comes with systemd too.
+1
|
2025-04-01T06:37:57.716884
| 2016-03-01T20:56:39
|
137684514
|
{
"authors": [
"arve0"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3872",
"repo": "arve0/codeclub_lesson_builder",
"url": "https://github.com/arve0/codeclub_lesson_builder/issues/201"
}
|
gharchive/issue
|
complete translation of UI
The UI should handle:
[ ] index.md in different languages
[ ] show only courses / lessons for selected language
[ ] save / restore state with localStorage
[ ] propose lesson in different language if languages is set in YAML
---
title: Straffespark
languages:
nn-NO: straffespark.nn.md
en-GB: penalty.md
---
Ref: https://github.com/kodeklubben/oppgaver/issues/170
Will be fixed in codeclub-viewer
|
2025-04-01T06:37:57.725690
| 2019-09-26T12:10:42
|
498844239
|
{
"authors": [
"aryzhov",
"pspatil16"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3873",
"repo": "aryzhov/flutter-expandable",
"url": "https://github.com/aryzhov/flutter-expandable/issues/25"
}
|
gharchive/issue
|
Expand the card by default
As the name suggest, how would you keep a card expanded by default ?
You can wrap your ExpandablePanel with an ExpandableNotifier and set initialExpanded to true:
ExpandableNotifier(
initialExpanded: true,
child: ExpandablePanel(
...
)
)
|
2025-04-01T06:37:57.734218
| 2021-02-15T12:57:39
|
808517618
|
{
"authors": [
"sanemogi"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3874",
"repo": "asand3r/zbx-hpmsa",
"url": "https://github.com/asand3r/zbx-hpmsa/issues/38"
}
|
gharchive/issue
|
Getting : ERROR: 2 : Invalid sessionkey when running ./zbx-hpmsa.py
Hello,
When trying to test the connection to MSA using the ./zbx-hpmsa.py lld <IP_ADDRESS> disks
or ./zbx-hpmsa.py full <IP_ADDRESS> volumes command I get the following error
ERROR: 2 : Invalid sessionkey
Could you please help to resolve this ?
thank you in advance,
Spyros
Issue resolved by using custom user/password
|
2025-04-01T06:37:57.736880
| 2015-11-19T15:05:06
|
117832534
|
{
"authors": [
"DanielJMaher",
"FBRTMaka"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3875",
"repo": "asascience-open/ooi-ui",
"url": "https://github.com/asascience-open/ooi-ui/pull/597"
}
|
gharchive/pull-request
|
Mc cleanup
@DanielJMaher @birdage , minor fixes:
Cleaned up the plot control buttons (got messed up a little on some css changes)
Fixed the event class column on the plotting page (now shows up again)
Added an info button next to the subscription button (per Eoin)
Disabled the selection of different download options for now (per Eoin)
4 might cause us some issues. There are requirements for Json and CSV. But since they are going through ERDDAP at some point, we would be fine.
I'm a little worried about doing this, but ithink it might be fine.
@DanielJMaher great find that this works now for the download! Just re-enabled it!
reviewed with Dan
|
2025-04-01T06:37:57.740289
| 2020-10-04T21:18:08
|
714404775
|
{
"authors": [
"Niwla23",
"asayah-san"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3876",
"repo": "asayah-san/fokus-android",
"url": "https://github.com/asayah-san/fokus-android/issues/5"
}
|
gharchive/issue
|
Version 2 unusable: App crashes when pressing +
Since version 2.0.0 the app crashes immidietaly when you click on + at tasks. I tried to wipe data, didnt help.
Android Version: 9.0
Model: Nokia 3 TA-1032
GMS: manually removed
I am terribly sorry for your experience. However, I cannot reproduce the issue you are having. Can you update to the 2.0.1 release?
2.0.1 has the same problem for me. I will try Debugging tomorrow maybe.
Btw, maybe you could add an error Catcher showing the exception in an extra Activity
Update: I have reproduced the issue, it seems that this is only present in Android 9.0 Pie. The TwoLineRadioButton component is throwing a NullPointerException.
Okay, is this fixable?
Yes, working on a fix right now. :)
Update: will release the update with the hotfix later
Hotfix is now live in the Releases section :)
|
2025-04-01T06:37:57.743386
| 2017-05-24T07:19:27
|
230943088
|
{
"authors": [
"Baam25",
"aschuch",
"erickva"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3877",
"repo": "aschuch/StatefulViewController",
"url": "https://github.com/aschuch/StatefulViewController/issues/57"
}
|
gharchive/issue
|
UIActivityIndicatorView Always on top left
UIActivityIndicatorView as loadingView Always on top left even when it's set with a larger frame, it is always small in the corner.
Is there a work around this? As the insets method applies for all the state views.
Are you using the latest version from master?
Can you post some example code?
3.0, sure, in my test I am simply giving it a frame and assigning it to the loadingView:
let loading = UIActivityIndicatorView(frame: CGRect(x: 100, y: 100, width: 400, height: 400))
loading.startAnimating()
loadingView = loading
But regardless of the size of the View it is always in the top left corner:
Hm, please try with the latest master branch, there are some changes in there that did not make it into an official release yet. Are you using a vanilla UIViewController or UITableViewController/UICollectionViewController?
It works perfectly with the master branch! Thank you very much aschuch!
@aschuch Can you release a new version from the master branch which will include these changes?
|
2025-04-01T06:37:57.768534
| 2021-02-07T02:15:37
|
802836459
|
{
"authors": [
"drlippman",
"kozross"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3878",
"repo": "asciimath/asciimathml",
"url": "https://github.com/asciimath/asciimathml/issues/124"
}
|
gharchive/issue
|
Inconsistent whitespace sensitivity with numbers
If you type the equivalent of "1.2 00" into the test environment, it renders as '1.200' (i.e. one number). This is consistent with doing "12 00", which renders as '1200'. However, if you try "1. 200", you instead get '1. 200' (i.e two 'numbers').
This raises multiple questions regarding parsing of whitespace and numbers:
Is whitespace significant for numbers or not? The current situation appears to be 'depends', which is unhelpful.
Is it valid to have something like "12." parse as a number? I would argue that this doesn't really make sense.
Furthermore, even though some specification of the grammar exists, how numbers are parsed is not specified: I had to discover myself that, for example, insignificant leading or trailing zeroes remain. This should likely get fixed too, and it relates to this issue, which is why I mention it here.
The short answer is that whitespace breaks tokens. So "1200" becomes <mn>1200<mn>, while "12 00" becomes <mn>12</mn><mn>00</mn>. So in the MathML it is two numbers, but it's up the renderer (FireFox or MathJax if you're using that) how to display that, and it appears they display them without a space. That's not something AsciiMathML controls, so not anything to "fix" here.
Now, interestingly, it does appear "1.2" becomes <mn>1.2</mn> while "12." becomes <mn>12</mn><mo>.</mo>. So currently "12." does not get parsed as a number, but as a number and an operator, and the rendering engine seems to add a space after operators, which is why "1. 200" is appearing to you as two numbers. I would argue that parsing is actually wrong, and "12." should get parsed as a number, since an ending decimal point can be used to indicate trailing zeros are significant digits in science fields. So if anything needs fixing, it'd be getting "12." to get parsed as a single number.
@drlippman Thank you for that - it's very useful information. I feel it would help a lot if all this got added to the description of the grammar on the main AsciiMath site.
Feel free to make a PR to the website repo with suggested changes.
|
2025-04-01T06:37:57.797592
| 2022-10-18T08:10:18
|
1412766503
|
{
"authors": [
"ascopes"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3879",
"repo": "ascopes/java-compiler-testing",
"url": "https://github.com/ascopes/java-compiler-testing/issues/100"
}
|
gharchive/issue
|
Rename path classes
Current path classes (NioPath, RamPath, SubPath, etc) are currently a bit ambiguous as they may be implied to be types of java.nio paths. In reality, they are a wrapper around them instead.
Files currently exist at https://github.com/ascopes/java-compiler-testing/tree/bd99311b0de4508537b11708d5dc5ec6c6c12092/java-compiler-testing/src/main/java/io/github/ascopes/jct/paths
Instead, I'd like to rename them to something more meaningful. Haven't got a final set of names in mind just yet but something like this might make sense:
PathLike -> PathHolder
NioPath -> PathPathHolder
SubPath -> NestedPathHolder
RamPath -> RamPathHolder
The package should also be renamed accordingly.
Examples in the README will need updating as well when this is done.
Closed by f1437b9ff8775212f8af75861754663775b300c9
|
2025-04-01T06:37:57.813590
| 2017-04-27T08:32:29
|
224710477
|
{
"authors": [
"bergel",
"danilofes"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3880",
"repo": "aserg-ufmg/RefDiff",
"url": "https://github.com/aserg-ufmg/RefDiff/issues/1"
}
|
gharchive/issue
|
Tutorial?
It would be fantastic if a kind of tutorial could be provided in the README file. Currently, it is hard to grasp the expressiveness of RefDiff just by reading the README file
Dear Bergel,
I have added more info in the README, along with a first version of a tutorial:
https://github.com/aserg-ufmg/RefDiff/blob/master/doc/Tutorial1.md
We intend to keep improving the docs. Let me know if you have other suggestions.
Thanks
|
2025-04-01T06:37:57.826464
| 2016-09-29T17:05:44
|
180107449
|
{
"authors": [
"ashish-chopra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3881",
"repo": "ashish-chopra/angular-gauge",
"url": "https://github.com/ashish-chopra/angular-gauge/issues/9"
}
|
gharchive/issue
|
label text animation with animating gauge is not supported
When the gauge starts animating the text inside the label shows the hard value, instead of showing the value at the current frame of animation.
I am leaving this requirement altogether. I don't think providing an extra label anywhere in the gauge is useful. The users might put gauge inside a div where they have the liberty to put label anywhere. If i add this feature, i can only support two places: above or below only.
|
2025-04-01T06:37:57.827475
| 2018-04-16T16:36:47
|
314732063
|
{
"authors": [
"alichtman",
"ashishb"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3882",
"repo": "ashishb/osx-and-ios-security-awesome",
"url": "https://github.com/ashishb/osx-and-ios-security-awesome/issues/10"
}
|
gharchive/issue
|
Add FSMonitor - macOS app that monitors all changes in the file system
http://fsmonitor.com/
Thanks for the suggestion, why don't send a pull request?
|
2025-04-01T06:37:57.869358
| 2017-11-28T20:54:22
|
277534559
|
{
"authors": [
"OterLabb",
"askmike",
"cmroche",
"patjk",
"welly59"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3883",
"repo": "askmike/gekko",
"url": "https://github.com/askmike/gekko/issues/1375"
}
|
gharchive/issue
|
Issue when starting Gekko - Windows 10
Installed Gekko using this guide:
https://gekko.wizb.it/docs/installation/installing_gekko_on_windows_with_bash_on_windows_10.html
Install seems to work ok except the handle.js file is different. Line 53 is different so i modified WAL to DEL any (note the error is the same regardless of whether i modify handle.js or not)
When i start using 👍
n**ode gekko --ui**
I get the output below:
**TAlib is enabled
TULIP indicators is enabled
Serving Gekko UI on http://localhost:3000/
(node:808) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 1): Error: Exited with code 3**
When i then navigate with my browser it seems ot be working except as soon as i start a new live Gekk and then press start i get this:
**_events.js:160
throw er; // Unhandled 'error' event
^
Error: SQLITE_IOERR: disk I/O error
at Error (native)
RECEIVED ERROR IN GEKKO<PHONE_NUMBER>398145
Child process has died._**
and a popup in my browser saying:
GEKKO ERROR: Child process has died.
Hi, the problem here is that your SQLITE3 plugin is using WAL journaling mode, which doesn't work well on windows 10 and creates this problem when you try to open the history database for reading.
The develop branch has a fix to auto-select the DELETE journaling mode when windows is detected, alternatively change the 'WAL' entry in sqlite.journalMode in your config file to 'DELETE', you will need to delete and reimport your history.
However there are other problems you will experience on Windows 10, even with Delete mode the write to the databse will be very slow and data may become corrupt during an import (data will be missing from the end of the import our the journal is corrupted which can use your import to be broken into multiple date ranges). There is a pull request oustanding to address this, #1369
@askmike I think the PR is going to be important for Windows users on BASH. Any chance of prioritizing a review?
I did modify WAL to DEL on line 53 of handle.js. Should I have made that DELETE instead?
I amended line 53 to read:
db.run('PRAGMA journal_mode = ' + config.sqlite.journalMode||'DELETE');
but it is still getting errors.
On Wed, Nov 29, 2017 at 1:01 PM, welly59<EMAIL_ADDRESS>wrote:
I amended line 53 to read:
db.run('PRAGMA journal_mode = ' + config.sqlite.journalMode||'DELETE');
but it is still getting errors.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/askmike/gekko/issues/1375#issuecomment-347761666, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AA7MD77M442fCeVe1Y3dAMUu7iwrCyVgks5s7PM0gaJpZM4Qt5Y5
.
--
PGP key at keybase.io/mikevanrossum
https://keybase.io/mikevanrossum/key.asc
@askmike This is one of the problems I received when trying to sort out corruption problems of my SQLITE3 setup on Windows on BASH. It may not be the disk in this case.
@welly59 Delete your history files, then merge the PR #1369 and let me know if that addresses your issue.
cmroche, would you mind providing instructions on how to merge the PR?
@welly59 If you are using GitKraken I think you can right click on the PR (in the PR view) and it will give you an option to merge it to your current branch. If you sync to the develop branch you should be able to do this without any conflicts.
I will give that a try. I installed using git via bash so a cmd line method would be ideal
ok i think i have done correctly. I am not getting disk i/o error anymore and have set up 2 paper traders. I'll report back once they have run for a bit
How long does a typical import take? I've tried various imports for just the last day or two, and none successfully go through. What's the best way to troubleshoot?
ok i/o error solved but now failing to load getCandles. It keeps spawning node processes until my machine dies.
Ive installed ubuntu in a VM instead which is working fine
Got this same error running gekko on bash on windows, on two different computers. Switching to the develop branch seems to work for now.
|
2025-04-01T06:37:57.881031
| 2018-11-07T18:14:39
|
378408761
|
{
"authors": [
"pitbullgti",
"sahni619",
"thecccut"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3884",
"repo": "askmike/gekko",
"url": "https://github.com/askmike/gekko/issues/2646"
}
|
gharchive/issue
|
Unable to Backtest on Ubuntu
Note: this is the technical bug tracker, please use other platforms for getting support and starting a (non technical) discussion. See the getting help page for details.
I'm submitting a ...
[ X ] bug report
[ ] question about the decisions made in the repository
Action taken (Tried Backtesting a Dataset)
Expected result (Would have shown the backtest results)
Actual result (TypeError: Cannot read property 'to' of undefined)
Other information (e.g. detailed explanation, stacktraces, related issues, suggestions how to fix, links for us to have context, eg. stackoverflow, etc)
Error :
` from: r.to,
^
TypeError: Cannot read property 'to' of undefined
at _.map.r (/root/gekko/core/tools/dateRangeScanner.js:117:25)
at Function.map (/root/gekko/node_modules/lodash/dist/lodash.js:3509:27)
at async.whilst (/root/gekko/core/tools/dateRangeScanner.js:115:15)
at /root/gekko/node_modules/async/dist/async.js:988:16
at Object.whilst (/root/gekko/node_modules/async/dist/async.js:5092:25)
at async.parallel (/root/gekko/core/tools/dateRangeScanner.js:68:13)
at /root/gekko/node_modules/async/dist/async.js:3853:9
at /root/gekko/node_modules/async/dist/async.js:484:16
at iterateeCallback (/root/gekko/node_modules/async/dist/async.js:1013:24)
at /root/gekko/node_modules/async/dist/async.js:988:16
at /root/gekko/node_modules/async/dist/async.js:3850:13
at apply (/root/gekko/node_modules/async/dist/async.js:41:25)
at /root/gekko/node_modules/async/dist/async.js:76:12
at Statement. (/root/gekko/plugins/sqlite/reader.js:135:5) `
Tried on Gekko UI. Haven't tried command-line.
More Findings I retried backtesting and it worked once and then again when I tried it gave out below error
Error: non-error thrown: Child process has died. at Object.onerror (/root/gekko/node_modules/koa/lib/context.js:105:40) at <anonymous> at process._tickCallback (internal/process/next_tick.js:189:7)
Plz note am using default RSI strategy to backtest with only modifications in the TOML file.
I restarted my bot and after 2 days it crashed again with the same error. It happens all the time.
I'm having the same issue running on a Raspberry Pi 3 npm v6.4.1 and nodejs 8.12.0
Heres the error:
`<-- POST /api/scansets
/home/qwe/gekko/core/tools/dateRangeScanner.js:117
from: r.to,
^
TypeError: Cannot read property 'to' of undefined
at _.map.r (/home/qwe/gekko/core/tools/dateRangeScanner.js:117:25)
at Function.map (/home/qwe/gekko/node_modules/lodash/dist/lodash.js:3509:27)
at async.whilst (/home/qwe/gekko/core/tools/dateRangeScanner.js:115:15)
at /home/qwe/gekko/node_modules/async/dist/async.js:988:16
at Object.whilst (/home/qwe/gekko/node_modules/async/dist/async.js:5092:25)
at async.parallel (/home/qwe/gekko/core/tools/dateRangeScanner.js:68:13)
at /home/qwe/gekko/node_modules/async/dist/async.js:3853:9
at /home/qwe/gekko/node_modules/async/dist/async.js:484:16
at iterateeCallback (/home/qwe/gekko/node_modules/async/dist/async.js:1013:24)
at /home/qwe/gekko/node_modules/async/dist/async.js:988:16
at /home/qwe/gekko/node_modules/async/dist/async.js:3850:13
at apply (/home/qwe/gekko/node_modules/async/dist/async.js:41:25)
at /home/qwe/gekko/node_modules/async/dist/async.js:76:12
at Statement. (/home/qwe/gekko/plugins/sqlite/reader.js:135:5)
--> POST /api/scansets 200 6,668ms 254b
`
This happens as soon as I try to scan datasets.
Me to :(
|
2025-04-01T06:37:57.884886
| 2018-08-19T12:38:46
|
351899830
|
{
"authors": [
"askmike",
"mariodantas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3885",
"repo": "askmike/gekko",
"url": "https://github.com/askmike/gekko/pull/2460"
}
|
gharchive/pull-request
|
Stable
What kind of change does this PR introduce? (Bug fix, feature, docs update, ...)
feature
What is the current behavior? (You can also link to an open issue here)
When gekko service ir restarted (manually) we lost live gekkos (are empty)
What is the new behavior (if this is a feature change)?
When service is stopped (i.e. for a git pull) and then restarted, gekko recovers all running gekkos and stir running them
Other information:
Hey, it seems something went wrong with this PR. Just to be sure: a PR is a pull request which has some new code that you'd like to add to the project.
When service is stopped (i.e. for a git pull) and then restarted, gekko recovers all running gekkos and stir running them
I can't find code for this in the changes. But I do find a lot of other stuff. Could you clarify what you were trying to do?
Closing this, feel free to comment and answer the questions above if you want to reopen.
|
2025-04-01T06:37:57.887379
| 2023-02-22T10:02:51
|
1594809031
|
{
"authors": [
"bharath-bhushan-glean",
"sriram-vudayagiri-glean"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3886",
"repo": "askscio/atlassian-plugins",
"url": "https://github.com/askscio/atlassian-plugins/pull/19"
}
|
gharchive/pull-request
|
[Confluence Plugin] Add info endpoint in plugin
This change adds a new /info endpoint that can be called by any logged in user.
By default it returns the following fields:
User Info: Returns information regarding the logged in user such as name, key, email, isAdmin, etc.
Instance info: Returns information regarding the Confluence Server instance such as version, baseUrl, etc.
Plugin info: Returns information regarding the Scio plugin such as target, version, lastResponseCode, lastSuccessTime, lastFailureTime, etc.
[x] Test locally
[x] Test on actual instance
Latest changes LGTM.
|
2025-04-01T06:37:57.890372
| 2015-03-22T16:57:16
|
63552877
|
{
"authors": [
"ahmadsalim",
"mrkkrp"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3887",
"repo": "aslatter/parsec",
"url": "https://github.com/aslatter/parsec/issues/32"
}
|
gharchive/issue
|
Allow more expressive functions to be given as input in operator tables for buildExpressionParser
Currently, the Operator constructors used for building expression parsers are required to be pure functions wrapped in the ParsecT monad, i.e.:
Infix (ParsecT s u m (a -> a -> a)) Assoc
Prefix (ParsecT s u m (a -> a))
Postfix (ParsecT s u m (a -> a))
This however limits what can be done with the operators, for example it is not possible to perform additional checks of the parsed argument under the ParsecT monad and then fail using fail or unexpected, or build up new information in the complete term using the parsed sub-terms.
I would therefore like to ask for a feature request that allows taking more expressive functions where only the resulting term needs to be in the ParsecT monad. In order to not break backwards compatibility, one could imagine having an M postfix to each new constructor, so the following constructors are added:
InfixM (a -> a -> ParsecT s u m a) Assoc
PrefixM (a -> ParsecT s u m a)
PostfixM (a -> ParsecT s u m a)
I would happily submit a PR if the feature request is considered.
@ahmadsalim, this is interesting. We're considering replacing of old constructors with something like this in Megaparsec. You can discuss it here: mrkkrp/megaparsec#22.
|
2025-04-01T06:37:57.900848
| 2022-09-12T10:22:26
|
1369637455
|
{
"authors": [
"paullewis"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3888",
"repo": "aspect-build/rules_js",
"url": "https://github.com/aspect-build/rules_js/issues/446"
}
|
gharchive/issue
|
js_binary + ESM + hermeticity
Assuming something like this in a BUILD file:
load("@aspect_rules_ts//ts:defs.bzl", "ts_project")
load("@aspect_rules_js//js:defs.bzl", "js_binary")
load("@aspect_rules_swc//swc:defs.bzl", "swc_transpiler")
load("@bazel_skylib//lib:partial.bzl", "partial")
ts_project(
name = "src",
srcs = [
"src.ts",
],
declaration = True,
transpiler = partial.make(
swc_transpiler,
source_maps = "true",
swcrc = "//:swcrc_esm",
),
tsconfig = "//:tsconfig",
deps = [
"//:node_modules/@types/node",
],
)
js_binary(
name = "src_bin",
entry_point = ":src.js",
)
And assuming that the transpilation of src.ts makes an ESM bundle in bazel-dist, running the js_binary fails because the generated src.js file is neither src.mjs, nor does Node know to run as "type": "module".
In other words, bazel run //src:src_bin will fail:
(node:92736) Warning: To load an ES module, set "type": "module" in the package.json or use the .mjs extension.
However, if in the source tree I update the package.json file to have "type": "module", the js_binary will now run.
I'm not sure if this is a bug or a question, but I'm surprised that a change to the package.json file in the source tree affects the js_binary. I assumed (perhaps incorrectly) that there would be a sandbox for the node process underpinning js_binary, and that it wouldn't have access to a package.json in the source tree.
Am I missing something?
Ah, I see. Happy to close this if it's a dupe... wdyt?
|
2025-04-01T06:37:57.911049
| 2017-09-30T22:51:59
|
261877945
|
{
"authors": [
"dnfclas",
"guardrex"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3889",
"repo": "aspnet/ApiDocs",
"url": "https://github.com/aspnet/ApiDocs/pull/29"
}
|
gharchive/pull-request
|
Update descriptions for API docs
Fixes #2197
This has the smell :nose: of an auto-generated document. If so, plz close. If not, this fills in the MIA descriptions.
@guardrex,
Thanks for having already signed the Contribution License Agreement. Your agreement was validated by .NET Foundation. We will now review your pull request.
Thanks,
.NET Foundation Pull Request Bot
Looks like it's covered by https://github.com/aspnet/ApiDocs/pull/23.
|
2025-04-01T06:37:57.912214
| 2019-11-14T17:50:57
|
523012456
|
{
"authors": [
"guardrex"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3890",
"repo": "aspnet/AspNetCore.Docs",
"url": "https://github.com/aspnet/AspNetCore.Docs/pull/15692"
}
|
gharchive/pull-request
|
Hosted services package reference
Fixes #15649
I think we can merely call out the package is added/present for the Worker SDK and implicit for the Web SDK.
Strip out the bits on a Web Host and Generic Host pair of samples. Those went away recently (3.0 updates era), so those bits no longer apply.
How about that :point_up"?
|
2025-04-01T06:37:57.914990
| 2019-07-06T13:34:50
|
464854984
|
{
"authors": [
"endeffects",
"mkArtakMSFT"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3891",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/11937"
}
|
gharchive/issue
|
Blazor Server-Side: Api-Controllers of an other assembly got not mapped
I'm running a blazor server side application on the latest .net core preview6.
When i'm moving the api controllers to an other referenced assembly
the controller got not mapped to any route at runtime. It works when i move them
back to the host or when i remove all the blazor server hosting stuff.
Thank you for filing this issue. In order for us to investigate this issue, please provide a minimalistic repro project that illustrates the problem.
I have provided a sample here: https://1drv.ms/u/s!AvfU-VGBgXuLg-o4LRpQynSJvInKcA?e=Pb00ls
There is an API Project with the moved Controller. By loading the FetchData Page you'll see that the controller was not found. By removing the external controler and activating the original one on the server project you'll see that the FetchData Page is working.
We're closing this as this seems to be a dupe of https://github.com/aspnet/AspNetCore/issues/11921
|
2025-04-01T06:37:57.917029
| 2019-07-12T21:36:49
|
467609947
|
{
"authors": [
"BrennanConroy",
"anurse",
"mikaelm12"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3892",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/12135"
}
|
gharchive/issue
|
StreamId Map should be cleared when connection is stopped
Currently we don't clear the internal dictionary for mapping stream ids to active streams when the client disconnects. It's a silent issue because we will override any stale values with new streams, but regardless, we should be clearing out the map on connection close.
And removing on stream closes?
We could. We'd have to have another map though correlating invocation ids to their stream ids. So we'd have to push the invocation id into the checkUploadStreamlogic where we allocate the stream Ids so we could group them together. I don't feel to strongly about this tbh.
Sure, but it's technically a memory leak on the client.
*sighs in agreement
Consider for preview 8, relatively low priority.
|
2025-04-01T06:37:57.926123
| 2019-08-08T12:01:53
|
478427422
|
{
"authors": [
"Bambofy",
"SteveSandersonMS",
"mkArtakMSFT",
"rynowak"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3893",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/12969"
}
|
gharchive/issue
|
Blazor - Javascript interop in thread from Invoke() breaks.
Hi,
If you try to change the pages html from any function/thread starting from Invoke() it breaks the page with no errors.
e.g.
public void ButtonClicked()
{
// do stuff...
Invoke(() => {
this.StateHasChanged();
UpdateHTMLDOM(); // <---------- this breaks the entire page.
});
}
private async void UpdateHTMLDOM()
{
await JSRuntime.invokemethod<>("CreateImage", "myImageString.png");
}
Can you provide a minimal example that reproduces the problem? Given the default project template, what is the minimal set of steps needed to update it to show the problem you're describing?
Using the default project template "Server-side razor" given in the visual studio 2019 preview version. Using netcore sdk version "3.0.0-preview6-27804-01".
Replace the counter.razor code with the following:
@page "/counter"
@inject IJSRuntime JSRuntime
<h1>Counter</h1>
<p>Current count: @currentCount</p>
<button @onclick="@ButtonClicked">Increment Counter.</button>
@code {
private int currentCount = 0;
private string Title = "";
bool PreventRenderLoop = false;
protected override void OnAfterRender()
{
if (PreventRenderLoop) return; // stop the page from entering a render loop.
Task getDataTask = Task.Run(GetData);
getDataTask.Wait();
}
public async void GetData()
{
// demo just load some variables.
Title = "Hello World";
// e.g. data = MyService.LoadData();
PreventRenderLoop = true;
// refresh the page after loading new values
Invoke(() =>
{
this.StateHasChanged();
// now the page has the new values loaded from the service
// we want to display the DOM which is relavent
Task.Run(DisplayDOM).Wait();
});
}
public async Task DisplayDOM()
{
currentCount = -1;
await JSRuntime.InvokeAsync<string>("Notify", "Successfully displayed dom!", "success");
}
public void ButtonClicked()
{
currentCount++;
}
}
And the code for the javascript notification (which goes in wwwroot/js and you reference in _Host.cshtml) is:
function Notify(pStrMessage, pStrMode)
{
alert(pStrMessage + " _ " + pStrMode);
}
If you run the above code, when you click the button to increment the counter, nothing happens. I resolved this by changing it to this:
@page "/counter"
@inject IJSRuntime JSRuntime
<h1>Counter</h1>
<p>Current count: @currentCount</p>
<button @onclick="@ButtonClicked">Increment Counter.</button>
@code {
private int currentCount = 0;
private string Title = "";
bool PreventRenderLoop = false;
protected override void OnAfterRender()
{
if (PreventRenderLoop) return; // stop the page from entering a render loop.
Task getDataTask = Task.Run(GetData);
getDataTask.Wait();
}
public async void GetData()
{
// demo just load some variables.
Title = "Hello World";
// e.g. data = MyService.LoadData();
PreventRenderLoop = true;
// refresh the page after loading new values
await Invoke(() =>
{
this.StateHasChanged();
});
// now the page has the new values loaded from the service
// we want to display the DOM which is relavent
Task domTask = Task.Run(DisplayDOM);
await domTask;
}
public async Task DisplayDOM()
{
currentCount = -1;
await JSRuntime.InvokeAsync<string>("Notify", "Successfully displayed dom!", "success");
}
public void ButtonClicked()
{
currentCount++;
}
}
I notice now that i set the PreventRender boolean and never unset it, therefore the OnAfterRender() function is constantly blocked. Could that cause the rest of the program to stop working? if so why does the above code allow the counter to work?
Cheers! Richard
Thanks for contacting us, @Bambofy.
It's not clear what you're trying to do as it most probably be simpler than this. Give us some more context, so that we are able to guide you better.
i'm trying to load data from a service in the OnAfterRender function and then refreshing the page to update the values.
Why use OnAfterRender? Typically the pattern for loading data asynchronously in Blazor looks like this:
@if (myData == null)
{
<div>something to show while data is loading</div>
}
else
{
<div>show your data here</div>
}
@code {
private MyData myData;
protected override async Task OnInitialized()
{
myData = await LoadMyData();
}
}
This pattern will render twice, once to show a "loading" UI and then again after the LoadMyData() method completes.
Do you have something special that you're trying to do here?
Why use OnAfterRender? Typically the pattern for loading data asynchronously in Blazor looks like this:
@if (myData == null)
{
<div>something to show while data is loading</div>
}
else
{
<div>show your data here</div>
}
@code {
private MyData myData;
protected override async Task OnInitialized()
{
myData = await LoadMyData();
}
}
This pattern will render twice, once to show a "loading" UI and then again after the LoadMyData() method completes.
Do you have something special that you're trying to do here?
Hey i did not find the root cause of the original problem to be honest!
The way i'm making it work at the moment is like this:
private static Type AccountType = null;
private static string EmailHash = "";
private static string PasswordHash = "";
protected override async Task OnAfterRenderAsync()
{
// Ensure user is authenticated.
if (AccountType == null)
{
await Authenticate();
// Show new data.
await Invoke(() =>
{
this.StateHasChanged();
});
}
await JSRuntime.InvokeAsync<string>("WebpageReady");
}
protected async Task Authenticate()
{
/*
* This authorization check can potentially be overriden since this code is clientside
* Therefore we only use this clientside authorization check to render the page's DOM.
*/
EmailHash = await JSRuntime.InvokeAsync<string>("GetSessionStorage", "EMAIL");
PasswordHash = await JSRuntime.InvokeAsync<string>("GetSessionStorage", "PASSWORD");
// Get the account type.
AccountType = AccountService.GetAccountType(EmailHash, PasswordHash);
}
|
2025-04-01T06:37:57.929122
| 2019-09-27T06:21:39
|
499257727
|
{
"authors": [
"hannespreishuber",
"mkArtakMSFT"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3894",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/14498"
}
|
gharchive/issue
|
Visual Studio 2019 Preview miss Blazor Template
Clean new INstall of Visual Studio 2019 Preivew- with all updates
Create Blazor Server App
.Razor files are uncolored in edtor
Visual Studio installer dialog comes up- is already running-
closing dialog- see nothing happen
Starting Project "Missing asp.net core-modul for hosting … iis express"
2nd machine everything works fine
Thanks for contacting us, @hannespreishuber.
This is not something runtime related. Please use VS Feedback to report this issue so the appropriate team can investigate this further.
|
2025-04-01T06:37:57.933306
| 2017-07-31T08:09:51
|
392384106
|
{
"authors": [
"AsValeO",
"cwe1ss",
"davidfowl",
"srollinet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3895",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/5900"
}
|
gharchive/issue
|
Don't swallow IHostedService.StartAsync exceptions
As discussed briefly in https://github.com/aspnet/Hosting/issues/1085#issuecomment-305717425 I propose that the HostedServiceExecutor should not catch exceptions in StartAsync.
By swallowing exceptions, the hosting-layer forces all services to be optional which is undesired IMO. By not swallowing exceptions we give services the option to decide this on its own. If a service is truly optional, it can swallow exceptions in its own StartAsync code.
I also think that StartAsync-exceptions should not be combined into an AggregateException. Instead it should just stop on the first exception.
One thing to consider is that right now this code is executed AFTER the server has been started. Not sure what effects this has when exceptions are no longer caught?!
I still hold the opinion that hosted services should be started before the server is started (a service could always decide to kick off another task in its StartAsync if it wants to be delayed) but this has also been discussed briefly in #1085 and declined.
Workaround for the 2.x version
public static class ServiceCollectionUtils
{
public static void AddCriticalHostedService<TService, TImplementation>(this IServiceCollection services)
where TService : class
where TImplementation : class, IHostedService, TService
{
services.AddSingleton<TService, TImplementation>();
services.AddHostedService<CriticalHostedServiceWrapper<TService>>();
}
private class CriticalHostedServiceWrapper<TService> : IHostedService
{
private readonly IHostedService _hostedService;
private readonly ILogger _logger;
public CriticalHostedServiceWrapper(TService hostedService, ILoggerFactory loggerFactory)
{
_hostedService = (IHostedService)hostedService;
_logger = loggerFactory.CreateLogger(_hostedService.GetType());
}
public async Task StartAsync(CancellationToken cancellationToken)
{
try
{
await _hostedService.StartAsync(cancellationToken);
}
catch (Exception e)
{
_logger.LogCritical(e, "Cannot start critical service. The application will exit.");
Environment.Exit(1);
}
}
public Task StopAsync(CancellationToken cancellationToken)
{
return _hostedService.StopAsync(cancellationToken);
}
}
}
Looks like behavior still the same in 3.0.
Update to use the generic host?
It works with generic host. Damn, I'm so far behind..
|
2025-04-01T06:37:57.937688
| 2019-02-12T06:01:27
|
409134326
|
{
"authors": [
"dougbu",
"glennc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3896",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/7491"
}
|
gharchive/issue
|
Remove CodeGenerator metadata in default service reference experience
Aim here is to remove the current requirement to specify the CodeGenerator metadata on <Service...Reference> (soon <OpenApiReference> and <ProtobufReference>) items in a project file.
Some metadata must remain to say "give me TypeScript" (instead of project's default language) but this should be the exception and not the rule. Fortunately, the type of document is going to be explicit (again <OpenApiReference> and <ProtobufReference>) and we won't have to choose among multiple tools. Probably best to make TargetLanguage the main override we provide in this area. (@glennc agreed?)
This change will require our Open API code generation partners (NSwag, AutoRest, ...) either all implement the same target name or include a well-known property specifying their target's prefix. Either way, the last file that declares a target or updates a property wins. So, unless our partners object, let's avoid the indirection and go with a single Open API target name. (gRPC doesn't have multiple partners for the foreseeable future and should be fine with a single target for that scenario.)
@glennc I put this in Preview4. That make sense to you?
@mkArtakMSFT the "Small" cost here doesn't mean calendar time due to the need for NSwag changes.
From discussion today. Generator tag isn't required by default. There will be a language property a customer can add with the values "Default", and "TypeScript" where default means whatever the extensions/language the project is for and TypeScript is TypeScript.
Partially addressed in ce8f053af7e9. Will add language metadata soon.
After offline discussion with @glennc, we've decided to close this issue and await custom input on the feature's overall usability
|
2025-04-01T06:37:57.949558
| 2019-02-25T15:41:01
|
414167433
|
{
"authors": [
"SteveSandersonMS",
"danroth27",
"fred-perkins",
"tylerhartwig"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3897",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/issues/7917"
}
|
gharchive/issue
|
Assembly Load Context throws errors in Blazor App 0.8
Describe the bug
When trying to load an assembly using the AssemblyLoadContext.Default a NotImplementedException is thrown. The Static property Default and it's other static methods don't seem to have any implementations.
To Reproduce
Steps to reproduce the behavior:
Download Sample code from: fred-perkins/BlazorLoading
Run application
Select Fetch Assembly in the left panel
Click "Fetch Assembly using AssemblyLoadContext" Button
Observe error in console: WASM: [System.NotImplementedException] The method or operation is not implemented.
Expected behavior
AssemblyLoadContext should attempt to load the DLL in the default context.
Additional context
We need to be able to dynamically load libraries one at a time within Blazor. Using Assembly.Load to read from bytes presents a problem where assemblies can be loaded several times, which causes a InvalidCastException to be thrown when resolving a type from one of those libraries.
Using AssemblyLoadContext it is possible to avoid this, however it seems unsupported when running in Webassembly on Blazor - But functions within a console application.
This is likely a mono issue:
Looking at mono it seems there are two implementations of the class: AssemblyLoadContext Facade
& AssemblyLoadContext.
@danroth27 - We spoke about this a week ago (Part of a derivco call). This is blocking us from migrating our framework to Blazor for our POC, as we need to be able to dynamically load libraries in the background. Is there a known alternative we could use other than Assembly.Load - Which doesn't solve our problem
@fred-perkins Thanks for sharing this feedback with us!
@SteveSandersonMS @lewing Thoughts on this one? Do we have any support currently for dynamically loading .NET assemblies in a Blazor app? Is this really all part of supporting lazy loading of application areas?
Could you clarify what the problem with Assembly.Load is? If your code caches the results of loading the assembly, is that not sufficient to ensure you only load each one once?
@SteveSandersonMS / @danroth27 So after a bit of testing I got Assembly.Load(byte[]) to work for me as i wanted. I think there was a misunderstanding about how Assembly.Load varies compared to the API on DotNet core - Where a byte array loaded assembly is anonymous and causes issues where assemblies would be loaded twice. This occurred when one was manually loaded, and the second where an assembly used a type for a dependent assembly - Which results in a InvalidCastException with duplicate types.
This means that for me to work around the issue I've had to use AssemblyLoadContext in .Net Core & Assembly.Load for the web assembly portion.
Ideally i'd prefer to use the AssemblyLoadContext Api across the board so i don't have to do fallback code in my .NetStandard libraries like the following:
public Assembly LoadFromStream(byte[] assemblySource)
{
if (AssemblyLoadContext.Default != null)
{
return AssemblyLoadContext.Default.LoadFromStream(new MemoryStream(assemblySource))
}
return Assembly.Load(assemblySource);
}
I did have a dig through the mono source code in the end. It looks like they don't fully support the AssemblyLoadContext Api, and where they have partial support it only works for when you specify the assemblyName or filePath - The LoadFromStream throws a not supported exception as mentioned above.
Thanks for the update, @fred-perkins! I'm glad you have a workable solution for now.
Longer term, any inconsistencies between .NET Core and Mono are definitely worth reporting. In this specific case it sounds like Mono ideally would support more of AssemblyLoadContext, and might even want tweaks to ensure that Assembly.Load behaves the same as it does on .NET Core (even if that behavior isn't what you wanted in your scenario, it should still be consistent). If you were able to report any such API gaps or inconsistencies to http://github.com/mono/mono that would be really helpful.
I'll close this as external since it's really about the underlying runtime, not about Blazor.
@SteveSandersonMS Awesome, thanks for the reply. I'll get a ticket raised with the Mono guys :)
@fred-perkins
Can you link to the related mono issue?
I'm struggling to properly resolve a type existing in Assembly A, that depends on Assembly B.
Both are loaded into the default context already, but I'm looking to ensure A uses the type from B, rather than what's already loaded.
|
2025-04-01T06:37:57.951457
| 2019-08-28T21:13:11
|
486612056
|
{
"authors": [
"BrennanConroy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3898",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/pull/13524"
}
|
gharchive/pull-request
|
Add a FetchHttpClient
Fixes https://github.com/aspnet/AspNetCore/issues/9444
Taking over PR #12599, I'll try to keep the two original commits when merging.
Bueller
|
2025-04-01T06:37:57.953028
| 2019-09-16T11:00:41
|
493980448
|
{
"authors": [
"adrianwright109",
"pranavkm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3899",
"repo": "aspnet/AspNetCore",
"url": "https://github.com/aspnet/AspNetCore/pull/14028"
}
|
gharchive/pull-request
|
Remove Newtonsoft JSON.NET from Blazor WASM template
Addresses #13422
@adrianwright109 could you cherry-pick your commit on the release/3.1 branch?
Closing this PR in favour of #14061 that correctly targets release/3.1 branch.
|
2025-04-01T06:37:57.957907
| 2014-09-16T23:18:14
|
42945280
|
{
"authors": [
"BrennanConroy",
"Tratcher"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3900",
"repo": "aspnet/Caching",
"url": "https://github.com/aspnet/Caching/issues/17"
}
|
gharchive/issue
|
Harden trigger and callback registrations in CacheEntry
There are possible race conditions in CacheEntry between AttachTriggers and DetatchTriggers. Similar for the callback APIs. These need to be hardened to prevent potential stress/race crashes.
Every call to InvokeEvictionCallbacks is wrapped in the ReaderWriterLock in MemoryCache so it looks impossible to invoke more than once
#126 and #132
|
2025-04-01T06:37:57.960384
| 2016-06-08T18:55:59
|
159244617
|
{
"authors": [
"GuardRex",
"Rick-Anderson",
"Tratcher"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3901",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/issues/1358"
}
|
gharchive/issue
|
new topic RUNNING ASP.NET CORE APPLICATIONS WITH IIS AND ANTARES (AZURE WEBSITES)
[ ] New article focused on Azure websites. Get info from RUNNING ASP.NET CORE APPLICATIONS WITH IIS AND ANTARES (AZURE WEBSITES)
https://docs.asp.net/en/latest/publishing/iis.html could also include some of these details.
@Tratcher There are a few nice troubleshooting adds for sure. I'll open an issue and make a list, including talking about the publish-iis tooling, which still isn't in that doc. Given that people assume it actually publishes apps to IIS, yeah, it probably should get a few words.
|
2025-04-01T06:37:57.961771
| 2017-02-09T03:24:41
|
206393452
|
{
"authors": [
"Rick-Anderson"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3902",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/issues/2709"
}
|
gharchive/issue
|
aspnetcore/hosting/directory-structure.md
https://docs.microsoft.com/en-us/aspnet/core/hosting/directory-structure
fixed by https://github.com/aspnet/Docs/commit/b6ff637c34bca1eba67c5a1cb7a4892ca201edc9
|
2025-04-01T06:37:57.967918
| 2018-06-19T07:49:11
|
333556073
|
{
"authors": [
"Rick-Anderson",
"m-henderson",
"yamazaroon"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3903",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/issues/7132"
}
|
gharchive/issue
|
Missing "Select Department" line of code, or, code description is inaccurate.
There appears to be one line of code missing in the following section:
https://docs.microsoft.com/en-us/aspnet/core/data/ef-rp/update-related-data?view=aspnetcore-2.1#update-the-courses-edit-page
Under the section that instructs us to "Update Pages/Courses/Edit.cshtml" the instructions state that the "preceding markup" has added a "Select Department" statement to the pull-down list. However, it appears that the line of code to add the statement is left out in the example.
On the other hand, it may have been intentionally left out for editing section as a Department would have already been selected when the Course was created. If this is the case, then the statement describing what the preceding code has added should not include the statement:
"Adds the "Select Department" option. This change renders "Select Department" rather than the first department.
Removing the "Select Department" option would also seem to render the validation statement unnecessary, as there would be no way to unselect a department before saving changes.
Once again, thank you for the great tutorials.
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 140ce637-4d58-995f-c69e-b1de51f1dbb5
Version Independent ID: 9bc46ffa-60db-bdee-0984-fab183c8f503
Content: Razor Pages with EF Core in ASP.NET Core - Update Related Data - 7 of 8
Content Source: aspnetcore/data/ef-rp/update-related-data.md
Product: asp.net-core
GitHub Login: @Rick-Anderson
Microsoft Alias: riande
@Rick-Anderson can you assign this issue to me so I can fix it up and submit a PR?
@m-henderson sorry missed this. I got it fixed. Thanks for the offer.
|
2025-04-01T06:37:57.973670
| 2018-12-05T17:28:55
|
387865396
|
{
"authors": [
"Grauenwolf",
"guardrex"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3904",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/issues/9844"
}
|
gharchive/issue
|
"Add providers" example doesn't match current ASP.NET Core Template
None of the configuration examples matches the code found in the current ASP.NET Core template used by Visual Studio.
public static class Program
{
/// <summary>Creates the web host builder.</summary>
/// <param name="args">The arguments.</param>
/// <returns></returns>
public static IWebHostBuilder CreateWebHostBuilder(string[] args) =>
WebHost.CreateDefaultBuilder(args)
.UseStartup<Startup>();
/// <summary>Defines the entry point of the application.</summary>
/// <param name="args">The arguments.</param>
public static void Main(string[] args)
{
CreateWebHostBuilder(args)
.Build().Run();
}
}
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 7184a308-a254-9350-a5cd-5bced1f369ae
Version Independent ID: 726e3bf1-f367-d733-8933-bccc04da0e16
Content: Logging in ASP.NET Core
Content Source: aspnetcore/fundamentals/logging/index.md
Product: aspnet-core
GitHub Login: @tdykstra
Microsoft Alias: tdykstra
@Grauenwolf ... It's somewhat common across the repo for a few reasons. For one thing, the sample apps trail the framework releases. Another is that samples concentrate on the concepts that the topics cover, which often don't strictly require the app to follow the template conventions. In this case (logging), the common thread is just hanging the ConfigureLogging method calls off of the host builder ... that's the same pattern regardless of the 2.x flavor.
I'll mark this issue on the sample update tracking issue. The doc author who takes on the 2.2 updates for this sample will see this at that time. Thanks for commenting on the topic.
|
2025-04-01T06:37:57.975150
| 2016-03-06T09:15:12
|
138772615
|
{
"authors": [
"Rick-Anderson",
"SychevIgor"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3905",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/pull/1066"
}
|
gharchive/pull-request
|
Update dependency-injection.rst
ASP.NET 5 to ASP.NET Core 1
Thanks for your interest. Closing as this will be updated in the next sweep.
Show all checks
|
2025-04-01T06:37:57.978521
| 2015-11-13T21:52:24
|
116860777
|
{
"authors": [
"Rick-Anderson",
"bitcrazed",
"danroth27",
"dnfclas"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3906",
"repo": "aspnet/Docs",
"url": "https://github.com/aspnet/Docs/pull/649"
}
|
gharchive/pull-request
|
Fixed instructions for configuring social logins
Fixes #648
Fixes instructions on how to dnvm use 1.0.0-beta 8
Fixes instructions on how to install user-secrets
Will have to update these instructions for RC too, but can deal with that after it's released.
Hi @bitcrazed, I'm your friendly neighborhood .NET Foundation Pull Request Bot (You can call me DNFBOT). Thanks for your contribution!
This seems like a small (but important) contribution, so no Contribution License Agreement is required at this point. Real humans will now evaluate your PR.
TTYL, DNFBOT;
@rustd @Rick-Anderson Please review
Looks good :shipit:
Thanks! 05fc5fd2b5d92dc48f89cb696e1adedc522dad76
|
2025-04-01T06:37:57.981018
| 2016-10-28T21:38:05
|
186028179
|
{
"authors": [
"dnfclas",
"v-anpasi"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3907",
"repo": "aspnet/EntityFramework.Docs",
"url": "https://github.com/aspnet/EntityFramework.Docs/pull/286"
}
|
gharchive/pull-request
|
Migration qa - index/toc fixes
I also removed the _Files too.
Hi @v-anpasi, I'm your friendly neighborhood .NET Foundation Pull Request Bot (You can call me DNFBOT). Thanks for your contribution!
In order for us to evaluate and accept your PR, we ask that you sign a contribution license agreement. It's all electronic and will take just minutes. I promise there's no faxing. https://cla2.dotnetfoundation.org.
TTYL, DNFBOT;
|
2025-04-01T06:37:57.998706
| 2016-04-11T16:32:55
|
147483762
|
{
"authors": [
"kall2sollies",
"maumar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3908",
"repo": "aspnet/EntityFramework",
"url": "https://github.com/aspnet/EntityFramework/issues/5033"
}
|
gharchive/issue
|
Include() ThenInclude() throws “Sequence contains more than one matching element” exception
I have a model which involves parent-child relations on 3 levels :
Corporations have companies
Companies belong to a corporation and have factories
Factories belong to a company
Since these 3 entities share a lot in common, they all inherit from an abstract BaseOrganization entity.
When I try to list all the factories, including their mother companies, and then including their mother corporations, I have these two different scenarios :
Without including BaseOrganization into the context, code-first creates three tables (TPC). Include() and ThenInclude() work fine, and I can list factories and traverse relations as expected.
Including BaseOrganization into the context, code-first creates one table (TPH) with a discriminator field. Include() and ThenInclude() throw a Sequence contains more than one matching element exception.
This issue (without the inheritance pattern) was already adressed here ( https://github.com/aspnet/EntityFramework/issues/1460).
So I think there is an issue with multi-level relations Include when these relations involve same base types AND when the tables are generated along the Table Per Hierarchy strategy.
Below is the full reproduction code :
using System;
using System.Collections.Generic;
using System.Linq;
using Microsoft.Data.Entity;
namespace MultiLevelTest
{
// All places share name and Id
public abstract class BaseOrganization
{
public int Id { get; set; }
public string Name { get; set; }
}
// a corporation (eg : Airbus Group)
public class Corporation : BaseOrganization
{
public virtual ICollection<Company> Companies { get; set; } = new List<Company>();
}
// a company (eg : Airbus, Airbus Helicopters, Arianespace)
public class Company : BaseOrganization
{
public virtual Corporation Corporation { get; set; }
public virtual ICollection<Factory> Factories { get; set; } = new List<Factory>();
}
// a factory of a company (Airbus Toulouse, Airbus US...)
public class Factory : BaseOrganization
{
public virtual Company Company { get; set; }
}
// setup DbContext
public class MyContext : DbContext
{
// if this line is commented, then code first creates 3 tables instead of one, and everything works fine.
public DbSet<BaseOrganization> BaseOrganizationCollection { get; set; }
public DbSet<Corporation> Corporations { get; set; }
public DbSet<Company> Companies { get; set; }
public DbSet<Factory> Factories { get; set; }
protected override void OnConfiguring(DbContextOptionsBuilder optionsBuilder)
{
optionsBuilder.UseSqlServer(
@"Server=(localdb)\mssqllocaldb;Database=MultiLevelTest;Trusted_Connection=True;MultipleActiveResultSets=true");
}
protected override void OnModelCreating(ModelBuilder modelBuilder)
{
base.OnModelCreating(modelBuilder);
modelBuilder.Entity<Corporation>().HasMany(c => c.Companies).WithOne(c => c.Corporation);
modelBuilder.Entity<Company>().HasMany(c => c.Factories).WithOne(c => c.Company);
modelBuilder.Entity<Factory>().HasOne(f => f.Company);
}
}
public class Program
{
public static void Main(string[] args)
{
using (var ctx = new MyContext())
{
ctx.Database.EnsureDeleted();
ctx.Database.EnsureCreated();
// Add a corporation with companies then factories (this works fine)
if (!ctx.Corporations.Any()) CreateOrganizationGraph(ctx);
// Get all the factories without including anything (this is still working fine)
var simpleFactories = ctx.Factories.ToList();
foreach(var f in simpleFactories) Console.WriteLine(f.Name);
// Get all the factories including their mother company, then their mother corporation
var fullFactories = ctx.Factories
.Include(f => f.Company)
.ThenInclude(c => c.Corporation)
.ToList();
foreach (var f in fullFactories) Console.WriteLine($"{f.Company.Corporation.Name} > {f.Company.Name} > {f.Name}");
}
}
public static void CreateOrganizationGraph(MyContext ctx)
{
var airbusCorp = new Corporation()
{
Name = "Airbus Group",
Companies = new List<Company>()
{
new Company
{
Name = "Airbus",
Factories = new List<Factory>()
{
new Factory {Name = "Airbus Toulouse (FR)"},
new Factory {Name = "Airbus Hambourg (DE)"}
}
},
new Company
{
Name = "Airbus Helicopters",
Factories = new List<Factory>()
{
new Factory {Name = "Eurocopter Marignane (FR)"},
new Factory {Name = "Eurocopter Deutschland (DE)"}
}
}
}
};
ctx.Corporations.Add(airbusCorp);
ctx.SaveChanges();
}
}
}
Has anyone had the opportunity to reproduce, and state if that is an actual bug, or an improper approach on my own ?
@kall2sollies I ran the scenario on our current bits and it seems to be working. I guess it must have been addressed in one of the Include improvements that we did after RC1 was released. Repro produces the following output:
Airbus Toulouse (FR)
Airbus Hambourg (DE)
Eurocopter Marignane (FR)
Eurocopter Deutschland (DE)
Airbus Group > Airbus > Airbus Toulouse (FR)
Airbus Group > Airbus > Airbus Hambourg (DE)
Airbus Group > Airbus Helicopters > Eurocopter Marignane (FR)
Airbus Group > Airbus Helicopters > Eurocopter Deutschland (DE)
Hello,
It must be as you say, since my code was run against rc1, and also ef7rc1.
So the next release will probably solve that.
Thanks for the reply.
Calendau GUQUET
Via mon iPhone
Le 20 avr. 2016 à 23:57, Maurycy Markowski<EMAIL_ADDRESS>a écrit :
@kall2sollies I ran the scenario on our current bits and it seems to be working. I guess it must have been addressed in one of the Include improvements that we did after RC1 was released. Repro produces the following output:
Airbus Toulouse (FR)
Airbus Hambourg (DE)
Eurocopter Marignane (FR)
Eurocopter Deutschland (DE)
Airbus Group > Airbus > Airbus Toulouse (FR)
Airbus Group > Airbus > Airbus Hambourg (DE)
Airbus Group > Airbus Helicopters > Eurocopter Marignane (FR)
Airbus Group > Airbus Helicopters > Eurocopter Deutschland (DE)
—
You are receiving this because you were mentioned.
Reply to this email directly or view it on GitHub
I am unable to migrate my test project to RC2 (ou RC3) builds of EF.
Here's what I did:
Added rererence to the beta nuget channel (https://www.myget.org/F/aspnetvnext/api/v3/index.json)
Changed package names to reflect the new naming and version reset
Here's my project.json file:
{
"version": "1.0.0-*",
"description": "LinqKitIssue Console Application",
"authors": [ "cguquet" ],
"tags": [ "" ],
"projectUrl": "",
"licenseUrl": "",
"compilationOptions": {
"emitEntryPoint": true
},
"dependencies": {
"LinqKit": "<IP_ADDRESS>",
"Microsoft.EntityFrameworkCore": "1.0.0-rc2-*",
"Microsoft.EntityFrameworkCore.Commands": "1.0.0-rc2-*",
"Microsoft.EntityFrameworkCore.SqlServer": "1.0.0-rc2-*",
"Microsoft.Extensions.Caching.Abstractions": "1.0.0-rc2-*"
},
"commands": {
"LinqKitIssue": "LinqKitIssue"
},
"frameworks": {
"dnx451": { }
}
}
It results in having unresolved dependencies, because they are not compatible with DNX version 4.5.1.
I have tried to change the framework to everything possible or impossible (net451, dnx46, net46, dnx462, net462) but the result would always be the same. Here's the output of dnu list:
Microsoft .NET Development Utility Clr-x86-1.0.0-rc1-16609
Listing dependencies for LinqKitIssue (D:\Documents\Visual Studio 2015\Projects\LinqKitIssue\src\LinqKitIssue\project.json)
[Target framework DNX,Version=v4.6.2 (dnx462)]
Framework references:
fx/Microsoft.CSharp
fx/mscorlib
fx/System
fx/System.ComponentModel.DataAnnotations
fx/System.Core
fx/System.Data
fx/System.Transactions
Package references:
EntityFramework 6.1.3
Ix-Async 1.2.5
* LinqKit <IP_ADDRESS>
Microsoft.AspNetCore.Hosting.Abstractions 1.0.0-rc2-20466
Microsoft.AspNetCore.Hosting.Server.Abstractions 1.0.0-rc2-20466
Microsoft.AspNetCore.Http.Abstractions 1.0.0-rc2-20466
Microsoft.AspNetCore.Http.Features 1.0.0-rc2-20466
* Microsoft.EntityFrameworkCore 1.0.0-rc2-20466
* Microsoft.EntityFrameworkCore.Commands 1.0.0-rc2-20466
Microsoft.EntityFrameworkCore.Relational 1.0.0-rc2-20466
Microsoft.EntityFrameworkCore.Relational.Design 1.0.0-rc2-20466
* Microsoft.EntityFrameworkCore.SqlServer 1.0.0-rc2-20466
* Microsoft.Extensions.Caching.Abstractions 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.Caching.Memory 1.0.0-rc2-20466
Microsoft.Extensions.Configuration.Abstractions 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.DependencyInjection 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.DependencyInjection.Abstractions 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.FileProviders.Abstractions 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.Logging 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.Logging.Abstractions 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.Options 1.0.0-rc2-20466 - Unresolved
Microsoft.Extensions.PlatformAbstractions 1.0.0-rc2-20466
Microsoft.Extensions.Primitives 1.0.0-rc2-20466 - Unresolved
Remotion.Linq 2.0.2
System.Diagnostics.DiagnosticSource 4.0.0-rc2-23931
System.Text.Encodings.Web 4.0.0-rc2-23931 - Unresolved
@kall2sollies Visual Studio is missing the tooling support for RC2 stuff, you can try using .NET cli which is the replacement for dnx, and a command line. Here is the link to the project, which contains some getting started instructions:
http://dotnet.github.io/getting-started/
Among all the things I thought I could try, it has been the only one I didn't, because I thought it was a lightweight version of the runtime, and not the future runtime itself, so ok tomorrow I'll go for it, thanks!
Ok I have installed the redist installer and can now run the .Net CLI with dotnet commands (.NET Command Line Tools (1.0.0-beta-001598).
No problem scaffolding a new projet with the dotnet new command, but I am still unable to build or run this simple project mentioned above.
Below is my current project.json file. Current framework moniker is set to net46, but I also tried net461, netcore50 (as in default CLI hello world project, but it still fails because it either cannot resolve dependencies (when targetting net46), or tells dependencies are not compatible with netcore50 (especially LinqKit, which brings EF6 with).
{
"version": "1.0.0-*",
"description": "LinqKitIssue Console Application",
"authors": [ "cguquet" ],
"tags": [ "" ],
"projectUrl": "",
"licenseUrl": "",
"compilationOptions": {
"emitEntryPoint": true
},
"dependencies": {
"NETStandard.Library": "1.0.0-rc2-23811",
"LinqKit": "<IP_ADDRESS>",
"Microsoft.EntityFrameworkCore": "1.0.0-rc3-20637",
"Microsoft.EntityFrameworkCore.Commands": "1.0.0-rc3-20637",
"Microsoft.EntityFrameworkCore.SqlServer": "1.0.0-rc3-20637",
"Microsoft.Extensions.Caching.Abstractions": "1.0.0-rc3-20637"
},
"commands": {
"LinqKitIssue": "LinqKitIssue"
},
"frameworks": {
"net46": { }
}
}
|
2025-04-01T06:37:58.009444
| 2017-01-04T14:52:32
|
198725330
|
{
"authors": [
"KieranDevlinSycous",
"Remleo",
"adduss",
"ajcvickers",
"divega",
"romfir",
"rowanmiller",
"smitpatel",
"starychfojtu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3909",
"repo": "aspnet/EntityFramework",
"url": "https://github.com/aspnet/EntityFramework/issues/7350"
}
|
gharchive/issue
|
Explicit loading for list of entries: context.Entries(...)...Load()
For explicit loading EF Core offers (from docs):
var blog = context.Blogs
.Single(b => b.BlogId == 1);
context.Entry(blog)
.Reference(b => b.Owner)
.Load();
If I have, for expample, list of 10 entries, I need to 10 times call Reference(...).Load() in foreach, that generate 10 SQL queries to DB.
How about optimized method Entries():
var blogs = context.Blogs
.Where(...)
.ToList();
context.Entries(blogs)
.Reference(b => b.Owner)
.Load();
which make a single SQL query like: select .... where [BlogOwner].[BlogId] in (?, ?, ?, ?, ?, ?)
Sorry, but I have not found similar functionality. Thanks
To things you can try:
Use eager loading, i.e. Include(b => b.Owner)
Write a query like this (I haven't actually tested it, but it should work):
var ids = context.ChangeTracker
.Entries<Blog>()
.Select(e => e.Property(b => b.BlogId))
.ToList();
context.Owner.Where(o => ids.Contains(o.BlogId)).Load();
@Remleo BTW, we would be interested in understanding your scenario better, e.g. why didn't you use eager loading in this case.
Sometimes may need to eager/explicit load a relationship after the parent model has already been retrieved. For example, this may be useful if I need to dynamically decide whether to load related models:
// This code is responsible for retrieve specific blog entities
// But it have no idea about logic in BlackBox.SomeMethod()
// That is why it dont load Navigations
var neededBlogs = context.Blogs.Where(....).ToList();
....
class BlackBox {
public void SomeMethod(IEnumerable<Blog> blogs)
{
// There might be code that ensure Owner loading,
// because this method have no idea about is `blogs` was load `Owners` or not.
// Also `Loader` should be intelligent enough for load only empty Navigations
// so calling this method multiple times is safe
context.Entries(blogs).Reference(b => b.Owner).Load();
foreach (var blog in blogs)
{
if (someDynamicCondition && blog.Owner != null)
{
SendNotificationToOwner(blog.Owner.email, "Alert!");
}
}
}
}
This code works:
var ids = context.ChangeTracker
.Entries<Blog>()
.Select(e => e.Property(b => b.BlogId))
.ToList();
context.Owner.Where(o => ids.Contains(o.BlogId)).Load();
but code is not "dry". where statement need to be hardcoded and match FK for Navigation-property. Context knows about FK and it is his "job" to load Navigation-properties.
Sorry for my english (
Reopening so we can visit this in triage.
Closing but will reconsider if we see more requests. We would consider a PR with the feature. You could also look at implementing it as an extension method.
+1
This is totally valid example for any TPT inheritance. Base model does not contains navigation properties and if we woluld like to display list with all inherited types with the common property fe. "Name", but loaded from different related entities of derivered models. We can't use include because there is no navigation property to do so, and if we load directly EF is not clever enough just to get related entites IDs by FK, but it will get whole tables, and performance will be not acceptable. In that case we have to change whole structure to TPH or create view, map it to new model and at the end we will end up with two different models to describe exactly the same entity :/
Visualisation ;):
A (entity with all 3 digit numbers - like dictionary of all of them)
B (base model to keep numbers assigments - abstract)
--> C (derived) --> User (additional relation to entity with Name)
--> D (derived) --> Company (additional relation to entity with Name)
--> etc.
I would like to see this as there are many query methods that use the same query but don't use the same included references therefore if you have one method to return the query and add all references that are used in all instances then you end up with a method that works but is really inefficient. For example:
public IQueryable<Account> GetAccountsCreatedOnDate(DateTime time) {
context.Account.Where(x => x.CreatedDateTime == time);
}
public IEnumerable<Post> GetPostsByUsersCreatedOnDate(DateTime time) {
var users = GetAccountsCreatedOnDate(time);
context.Entries(users).Reference(b => b.Posts).Load();
return users.SelectMany(x => x.Posts);
}
@KieranDevlinSycous - Just try this.
public IEnumerable<Post> GetPostsByUsersCreatedOnDate(DateTime time) {
return GetAccountsCreatedOnDate(time).SelectMany(x => x.Posts).AsEnumerable();
}
I find this feature very useful. For example if you don't want to use eager loading too much, because it generates JOIN queries and if some other query already fetched the data, you are still stuck with lot of joins instead of simple selects. For example you have entity A with property C and entity B with property C and you eager load A with C and then you still have too eager load C with B instead of just selecting B. So you either have costly queries or you have many many methods on your repositories. So it would be much simpler to just load the necessary references in place. I would love this feature to be implemented. Now I use this instead
public Query<TEntity> LoadBy<TForeignEntity>(
IEnumerable<TForeignEntity> foreignEntities,
Func<TForeignEntity, TEntity> entitySelector,
Func<TForeignEntity, Guid?> entityIdSelector,
bool unrestricted = false)
{
var ids = foreignEntities.Where(e => entitySelector(e) == null).Select(e => entityIdSelector(e).ToOption());
return Select(unrestricted).Where(e => e.Id, ids);
}
Our team would really need this feature, instead of it we must manually download entries by using id of a main entry and set its collection state to be loaded (Context.Entry(obj).Collection(c => c.Collections).IsLoaded = true)
|
2025-04-01T06:37:58.011945
| 2016-09-28T23:16:59
|
179915199
|
{
"authors": [
"ajcvickers"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3910",
"repo": "aspnet/EntityFramework",
"url": "https://github.com/aspnet/EntityFramework/pull/6632"
}
|
gharchive/pull-request
|
Implement complex type usages and properties
A ComplexTypeUsage represents the reference to a complex type on an entity type or on another complex type. It contains the ComplexProperty instances that represent the properties of the complex type and their facets in this usage.
/cc @AndriySvyryd (But anyone can review...)
Agreed. We probably should review it with others--I'll make a note to talk about it when we discuss in the design meeting.
|
2025-04-01T06:37:58.021249
| 2018-09-08T18:33:24
|
358318171
|
{
"authors": [
"ajcvickers",
"crhairr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3911",
"repo": "aspnet/EntityFrameworkCore",
"url": "https://github.com/aspnet/EntityFrameworkCore/issues/13265"
}
|
gharchive/issue
|
EntityQueryModelVisitor does not properly use CreateTransparentIdentifierType
As a 3rd-party EF Core provider maintainer, I need to be able to control the transparent identifier type used by EntityQueryModelVisitor. I can override CreatetTransparentIdentifierType, but this has no effect on VisitAdditionalFromClause because it does not properly call that method:
https://github.com/aspnet/EntityFrameworkCore/blob/146863a3bdc25c1ea1cb224ba0611594398c8459/src/EFCore/Query/EntityQueryModelVisitor.cs#L983
Instead, it directly references the internal TransparentIdentifier<,> type.
This causes type mismatches when overriding CreateTransparentIdentifierType and attempting to use that type in other visitor methods. The net effect is having to override more methods that strictly necessary just to ensure that the correct type is used.
Further technical details
EF Core version: ALL historical versions
Database Provider: Non-specific (problem is in EFCore)
Tracking this for 3.0 as part of the the larger query refactoring that is tracked by #12795
|
2025-04-01T06:37:58.023878
| 2019-08-08T09:38:59
|
478358886
|
{
"authors": [
"TehWardy",
"ajcvickers"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3912",
"repo": "aspnet/EntityFrameworkCore",
"url": "https://github.com/aspnet/EntityFrameworkCore/issues/17025"
}
|
gharchive/issue
|
Getting a string[] from a UDF
I can't seem to find anything about table valued functions in EF core is this even possible.
Here's a sample of what I would like to achieve ...
I have a role class that looks like this ...
[Table("Roles", Schema = "Security")]
public class Role
{
[Key]
public Guid Id { get; set; }
// this is a csv list of "privilege keys" that the role grants
public string Privs { get; set; }
...
}
Add this to my DbContext ...
[DbFunction("[DMS].[GetFolderPrivList]")]
public static string[] GetFolderPrivList(string userId, Guid folderId)
{
throw new Exception();
}
Then I have a UDF like this to compute a list of privs for a folder "path" (which recursively crawls the tree for inherited permissions) ...
|
2025-04-01T06:37:58.030581
| 2019-07-30T15:38:15
|
474681325
|
{
"authors": [
"DenisPimenov",
"pakrym",
"rynowak"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3913",
"repo": "aspnet/Extensions",
"url": "https://github.com/aspnet/Extensions/issues/2113"
}
|
gharchive/issue
|
Strange DI behaviour with ASPNETCORE_ENVIRONMENT
Describe the bug
I have aspnetcore application, where singleton type has a dependency on scoped type. When a variable is defined ASPNETCORE_ENVIRONMENT=Development, I get an InvalidOperationException: Cannot consume scoped service 'B' from singleton 'A', but when the ASPNETCORE_ENVIRONMENT='Anything', an exception is not raised and application started.
To Reproduce
Steps to reproduce the behavior:
Using this version of ASP.NET Core '2.2.6'
Run this code https://gist.github.com/DenisPimenov/fbde9d1f91170d2423cee1089a29028c
With these arguments ASPNETCORE_ENVIRONMENT="Production"
Application started without exceptions
Expected behavior
Application should be terminated with InvalidOperationException
Additional context
dotnet --info
.NET Core SDK (reflecting any global.json):
Version: 2.2.301
Commit: 70d6be0814
Runtime Environment:
OS Name: Mac OS X
OS Version: 10.14
OS Platform: Darwin
RID: osx.10.14-x64
Base Path: /usr/local/share/dotnet/sdk/2.2.301/
In development mode we run some extra validations as the container is built. The fact that you're seeing an error means that something is wrong.
In this case you're using a scoped service from a singleton - meaning that you're creating a long-lived instance of something that you expect to be scoped.
We don't run this validation at startup time in production scenarios because it's slow.
This validations has public api?
@pakrym
You can use UseDefaultServiceProvider method to configure the scope validation to always be on:
public static IWebHostBuilder CreateHostBuilder(string[] args) =>
WebHost.CreateDefaultBuilder(args)
.UseDefaultServiceProvider(options => options.ValidateScopes = true)
.UseStartup<Startup>();
Thank you for your comments.
|
2025-04-01T06:37:58.035327
| 2018-03-22T13:38:49
|
307646814
|
{
"authors": [
"TonyValenti",
"Tratcher"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3914",
"repo": "aspnet/HttpSysServer",
"url": "https://github.com/aspnet/HttpSysServer/issues/436"
}
|
gharchive/issue
|
Unlisted Dependency
Hi ASPNet Team,
I am trying to use the HttpSysServer in a project and I'm getting the error message below.
Itseems as though my project is missing a reference, but I'm noticing that System.Threading.Overlapped is not listed as a package dependency for this project. Are there any other unlisted dependencies I should add?
System.IO.FileNotFoundException: Could not load file or assembly 'System.Threading.Overlapped, Version=<IP_ADDRESS>, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The system cannot find the file specified.
File name: 'System.Threading.Overlapped, Version=<IP_ADDRESS>, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a'
at Microsoft.AspNetCore.Server.HttpSys.RequestQueue..ctor(UrlGroup urlGroup, ILogger logger)
at Microsoft.AspNetCore.Server.HttpSys.HttpSysListener..ctor(HttpSysOptions options, ILoggerFactory loggerFactory)
at Microsoft.AspNetCore.Server.HttpSys.MessagePump..ctor(IOptions`1 options, ILoggerFactory loggerFactory, IAuthenticationSchemeProvider authentication)
--- End of stack trace from previous location where exception was thrown ---
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, ServiceProvider provider)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(IServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitScoped(ScopedCallSite scopedCallSite, ServiceProvider provider)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitSingleton(SingletonCallSite singletonCallSite, ServiceProvider provider)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(IServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceProvider.<>c__DisplayClass22_0.<RealizeService>b__0(ServiceProvider provider)
at Microsoft.Extensions.DependencyInjection.ServiceProvider.GetService(Type serviceType)
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType)
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider)
at Microsoft.AspNetCore.Hosting.Internal.WebHost.EnsureServer()
at Microsoft.AspNetCore.Hosting.Internal.WebHost.BuildApplication()
at Microsoft.AspNetCore.Hosting.WebHostBuilder.Build()
at AlphaDrive.Services.AspNetCore.AspNetCoreExtension.<Start>d__14.MoveNext() in C:\Users\Tony Valenti\source\repos\AlphaDrive\AlphaDrive.Services.AspNetCore\AspNetCoreExtension.cs:line 60
What package version is this?
I am using this DLL:
C:\Users\Tony Valenti\source\repos\AlphaDrive\packages\Microsoft.AspNetCore.Server.HttpSys.2.0.3\lib\netstandard2.0\Microsoft.AspNetCore.Server.HttpSys.dll
and it has this version:
<IP_ADDRESS>
And you're running on .NET Core 2 or .NET 4.x?
I'm not sure how to answer that question. The library is a cross platform library that I'm compiling for .NET 4.6.1.
Oddly, my app works fine on Windows 10, but on Windows 7 I get the error mentioned.
Actually, I guess I see the error happening on windows 10 too. Apparently it is just not happening on my dev box.
This seems like an issue specific to your project dependency setup. Can you share a dummy project on github that reproduces it?
|
2025-04-01T06:37:58.040754
| 2017-01-18T15:31:56
|
201606423
|
{
"authors": [
"aL3891",
"moozzyk",
"vijayrkn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3915",
"repo": "aspnet/IISIntegration",
"url": "https://github.com/aspnet/IISIntegration/issues/317"
}
|
gharchive/issue
|
Webdeploy parameter file support/tool
Hello,
I've been working on deploying a asp.net core app to azure via webdeploy (in vsts) recently and I think the tooling could be improved here, in particular with respects to webdeploy and its parameter file syntax.
The current docs recommends simply zipping up the publish contents and this works fine if you don't want to use the parameters file support provided by web deploy and the vsts website tasks. however a very nice feature webdeloy is to be able to specify a Setparameters.xml file at release time for different environments. When building with msbuild this file is generated by the publish target, but to my knowledge this is not supported with the dotnet cli.
Proposal: Add a new tool to Microsoft.AspNetCore.Server.IISIntegration.Tools that generates a SetParameter.xml file
In addition to doing this I think it would be nice to have tools for actually managing web deploy parameters, but that is kind of an aside.
I went looking for the code for publish-iis and I did found it in the master branch in this repo, but not in the dev branch, has it moved somewhere else? (also, do you think this is a worth while idea at all?)
I know webdeploy isn't very cutting edge but it is a very useful tool especially combined with the vsts tooling.
/cc @vijayrkn
I think a better repo to report this is https://github.com/aspnet/websdk. This is also a new home for the publish-iis tool which was converted to an MSBuild task (https://github.com/aspnet/websdk/tree/dev/src/Publish/Microsoft.NETCore.Sdk.Publish.Tasks)
We have added support for adding parameters file for web deploy publish. https://github.com/aspnet/websdk/blob/dev/src/Publish/Microsoft.NET.Sdk.Publish.Targets/netstandard1.0/PublishTargets/Microsoft.NET.Sdk.Publish.MSDeploy.targets#L237
The next VS 2017 release should have this.
aha nice to see it coming! also I didn't know they opened up the websdk repo, that's great!
|
2025-04-01T06:37:58.045918
| 2018-11-16T22:55:49
|
381789469
|
{
"authors": [
"HaoK",
"natemcmaster",
"pranavkm",
"ryanbrandenburg"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3916",
"repo": "aspnet/Identity",
"url": "https://github.com/aspnet/Identity/pull/2072"
}
|
gharchive/pull-request
|
Upgrade projects to netcoreapp3.0
Changes:
Upgrade dependencies
Change TFM on Microsoft.AspNetCore.Identity to netcoreapp3.0
Remove .NET Framework tests
Part of https://github.com/aspnet/AspNetCore/issues/3754
Whoa, lots of test failures. Are these already known issues @HaoK?
That looks to be errors from the razor UI stuff, maybe the upgrade to 3.0 broke something there? @javiercn @pranavkm any ideas?
Can't find'C:\projects\identity\test\Identity.FunctionalTests\bin\Release\netcoreapp3.0\Identity.DefaultUI.WebSite.deps.json'
Does the 3.0 Sdk no longer generate a deps file?
Does the 3.0 Sdk no longer generate a deps file?
This sounds like something @ryanbrandenburg was investigating. Is this the issue you were seeing, Ryan?
I don't recall having missing deps.json files.
This might have been it: https://github.com/aspnet/Identity/pull/2072/commits/43bae5e91f15bffa63eadef179d7292ff7384a64 Tests pass locally now.
|
2025-04-01T06:37:58.050301
| 2016-09-27T17:29:24
|
179559820
|
{
"authors": [
"Eilon",
"kichalla"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3917",
"repo": "aspnet/JsonPatch",
"url": "https://github.com/aspnet/JsonPatch/issues/37"
}
|
gharchive/issue
|
Patching of dictionaries should be case sensitive on keys
The current code handles all (dynamic(example: ExpandoObject implements IDictionary<string,object>) or not) dictionaries which implement IDictionary<string, object> in a case Insensitive way. This is an incorrect behavior.
For example:
When Json.net deserializes data into a regular IDictionary<,> it considers the keys case-sensitive.
When Json.net deserializes data into a regular Poco type, it considers the property names case-Insensitive.
Considering the above behavior, we should do the following in json patch:
For dynamic types (like ExpandoObject, DynamicObject etc.), the matching of keys should be case-Insensitive. This is to keep the behavior consistent with a Poco type where the property names are considered case-Insensitive.
For dictionaries, keep the default json.net behavior, i.e. keep them case sensitive.
@dougbu @rynowak
@kichalla can you link to the PR/commit where you previously fixed this and then close this? Thanks!
81931e75d48370cf9163254523b1f6c4bcc92acb
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.