id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1222106589 | GLEmulation.ANGLE_GLES20 not so much friendly with Ubuntu 22.04 and Nvidia drivers
Issue details
With new Ubuntu update Wayland has become the default display server and probably some issues with Nvidia proprietary drivers are now rising. This is what I've noticed using LWJGL3 backend and ANGLE extension:
GLEmulation.ANGLE_GLES20 with Wayland and trying to active the Nvidia card: simply no effect. The default Intel IGPU is used (bad performances)
GLEmulation.ANGLE_GLES20 with X11 and trying to active the Nvidia card: Crash happens related to LWJGL
[LWJGL] GLFW_PLATFORM_ERROR error
Description : EGL: Failed to create window surface: Arguments are inconsistent
Stacktrace :
org.lwjgl.glfw.GLFW.nglfwCreateWindow(GLFW.java:2024)
org.lwjgl.glfw.GLFW.glfwCreateWindow(GLFW.java:2197)
com.badlogic.gdx.backends.lwjgl3.Lwjgl3Application.createGlfwWindow(Lwjgl3Application.java:524)
com.badlogic.gdx.backends.lwjgl3.Lwjgl3Application.createWindow(Lwjgl3Application.java:462)
com.badlogic.gdx.backends.lwjgl3.Lwjgl3Application.createWindow(Lwjgl3Application.java:448)
com.badlogic.gdx.backends.lwjgl3.Lwjgl3Application.<init>(Lwjgl3Application.java:162)
games.rednblack.editor.Main.main(Main.java:50)
Everything works fine with any other configuration (GL20 both Nvidia and Intel, GL30 both Nvidia and Intel, ANGLE_GLES20 with Intel IGPU).
This issue looks like something related more to drivers (https://www.phoronix.com/scan.php?page=news_item&px=Ubuntu-22.04-NVIDIA-XOrg-Back) or LWJGL itself and so hard to fix on libGDX side, but I think it's good to point out in any case.
Version of libGDX and/or relevant dependencies
libGDX: 1.10.1-SNAPSHOT
GPU: nvidia-driver-510
Please select the affected platforms
[ ] Android
[ ] iOS
[ ] HTML/GWT
[ ] Windows
[X] Linux
[ ] macOS
That line appears to be the culprit
https://github.com/libgdx/libgdx/blob/aca13b245c2832da425a5c484be427bac3ad19df/backends/gdx-backend-lwjgl3/src/com/badlogic/gdx/backends/lwjgl3/Lwjgl3Application.java#L501
If commented it doesn't crash anymore and GLES is correctly used by libGDX.. No idea why it doesn't work, seems that nvidia drivers in X11 doesn't like EGL context
I see zero benefit to using ANGLE on Linux, but it's good you found the offending line.
| gharchive/issue | 2022-05-01T10:26:38 | 2025-04-01T06:44:47.351248 | {
"authors": [
"Frosty-J",
"fgnm"
],
"repo": "libgdx/libgdx",
"url": "https://github.com/libgdx/libgdx/issues/6860",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2616056688 | Consider i2c without a timeout function
The p_timeout approach is silly. Yes, its helpful for clock stretching case, but its very rare to find devices that use clock stretching. If most devices supported clock stretching, then having the API take a timeout callback would be acceptable. But its so rare that most code is just passing hal::never_timeout(). p_timeout is also problematic for i2c devices that use DMA. Because this has to be polled, when should the DMA code be running? It could just spin while it waits for DMA to finish processing, but we loose out on the capability to put the device to sleep or switch tasks by using the polling option. Overall, it is likely in the future that we will either make hal::i2c_v2 with an hal::io_waiter in place of the p_timeout or eliminate the parameter all together and recommend that i2c implementations accept an io_waiter as an input parameter. Not sure
Remove timeout, don't use clock stretching devices
After discussing this with members of the libhal team, almost everyone agrees that we should move away from passing around time outs. And if you have to deal with a device that performs clock stretching, our advice would be "don't" and if you still need to use it, we consider that the bus is no longer deterministic and other external things will be necessary to provide some determinism such as a means to power off the device if it holds the bus down for too long. Like an output pin connected to the devices reset line.
| gharchive/issue | 2024-10-26T18:39:17 | 2025-04-01T06:44:47.362690 | {
"authors": [
"kammce"
],
"repo": "libhal/libhal",
"url": "https://github.com/libhal/libhal/issues/71",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1677846421 | 能否发布一个新的release
看issue中对redis 7的支持是2月份增加的,但是releas还是去年的,希望能发布一个新的release,方便go mod引用
好的, master 分支因为最近改动较多而且存在部分 api 变动,所以 release 会相对谨慎点。
感谢支持!
| gharchive/issue | 2023-04-21T05:39:03 | 2025-04-01T06:44:47.363811 | {
"authors": [
"libi",
"sniperking1234"
],
"repo": "libi/dcron",
"url": "https://github.com/libi/dcron/issues/45",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1309785963 | Suggest to switch to default effort 3
I bench-marked about 2k of my private digital cam pictures with d=1.0 and different efforts using "JPEG XL encoder v0.7.0 ab775e2".
First I checked some hundreds in an effort range from 2 to 8 and found that the resulting BPP for 3 is almost same as with effort 4. Effort 5 is near effort 6 in resulting BPP. Effort 2 is significant 'bad' in resulting compression size and 8 takes too long time. Therefore I skipped 2, 5, 6 and 8 for later tests (means using only 3, 5, 7).
I counted how often which effort results in lowest BPP and over all always -e 3 won. (e.g. a folder with 1714 jpg-files effort 3 won 1310 times. 5 with 373 and effort 7 with only 31). So I calculated the optimal lowest file size using for each file the best effort, setting this to 100%. Recompressing jpg-files with only the same effort: Effort 7 results in 106.7% -- effort 5 in 105,5% -- effort 3 in 101,3% in relation to optimum.
So even I choose only effort 3 for all files I'm very near to the optimum.
An other set (about 500 files) results in: e7=107,3% -- e5=104,9% -- e3 =101,7%
Question (has not to be answered but be considered in future development): Why the effort 7 which is more complex and should outperform lower efforts, is so bad?
With effort 8 it's a bit different. It wins more often against e 3 but not as much as I thought. It depends on selected jpg-files. I tried about 50 pictures. Effort 3 results in an over all file size of only 100.6% related to using effort 8.
I also switched for my set of 1714 files to "-d 0.5". Resulting in: e7=103.1% -- e5=103.2% -- e3 = 101.3% (like before comparing optimum to using only one effort). Now e7 is a bit better than e5 but still e3 wins.
I suggest to switch to default effort=3.
Best regards
Andreas
@A-Conrath I'm wondering if -e9 would lead to "always the best" compression, and if not, why not :)
@A-Conrath just to make sure, you're converting jpgs I guess, so you need to use the --jpeg_transcode flag to not just do a lossless jpg conversion of your files, which can be fully reverted.
I tried (granted with just one image) but I don't see any effect you're seeing with my file:
e9: 1016806 bytes (1.006 bpp)
e8: 1017526 bytes (1.007 bpp)
e7: 1101571 bytes (1.090 bpp)
e6: 1092293 bytes (1.081 bpp)
e5: 1081313 bytes (1.070 bpp)
e4: 1362793 bytes (1.349 bpp)
e3: 1368776 bytes (1.355 bpp)
e2: 1597923 bytes (1.582 bpp)
e1: 1598482 bytes (1.582 bpp)
While the encoding times are the following:
e9: 18.445s
e8: 11.041s
e7: 1.299s
e6: 0.695s
e5: 0.671s
e4: 0.472s
e3: 0.456s
e2: 0.448s
e1: 0.439s
So from my observation the jump from 4 to 5 is definitely worth the additionally processing time, while 7 tries a bit harder but doesn't seem to find anything worth compressing better for all the effort. 8/9 try even harder but find something worth compressing.
Overall, 3 looks like a bad choice as default for me, as most files are saved once and read and transferred many times.
@A-Conrath I'm wondering if -e9 would lead to "always the best" compression, and if not, why not :)
I tried to figure out a practicable way to convert (if encoder has version 1.0) all my pics. 9 is taking a way to much time. So I don't tried.
@A-Conrath just to make sure, you're converting jpgs I guess, so you need to use the --jpeg_transcode flag to not just do a lossless jpg conversion of your files, which can be fully reverted.
In my opinion "-j" is only needed if you not use the "-d" command. But the jxl-files are so much smaller than the original jpg-files, I have no doubt they are transcoded and not only recompressed. Additional the encoder always says "VarDCT".
Some of my files have also your straight behavior. I think you should benchmark some more.
A file size regression with higher effort should be considered a bug - the man page states "That is to say, all other parameters being equal, a higher effort should yield a file that is at least as dense and possibly denser, and with at least as high and possibly higher quality." https://github.com/libjxl/libjxl/blob/main/doc/man/cjxl.txt
Your dataset is original high resolution, low compression jpeg picture from camera? (with so many noise per pixel)
It's not at all representative to the real diversity of image and size encoded with an encoder, gain could be much higher.
Do the same test in -d 2 and resized denoised picture, you will see a bigger difference.
Also if you want to store your picture, time encoding is not a big deal.
Secondly, I can say that comparing file size without using the butteraugli benchmark is not good at all. Encoder is not doint an EXACT quality. You can have a two file @ -d 0.5 with different visual quality.
Use benchmark_xl.exe tool to compare compression efficiency.
Many people do tests on the discord, it's a best place to speak about tunning encoder.
After I have dealt a little more with this topic, I can understand the objection from the technical side. From the user side, however, this is completely useless. When I select a quality (d-Value), as a user I want to have exactly this quality and without the influence of the encoder effort.
So my original point should actually be "Effort-independent quality setting". In addition, my pictures are a cross-section of various (possibly not so good) digital cameras and smartphone shots. So they are quite representative (especially due to the noise that may be present).
If you want all efforts to make the same quality-affecting choices, then that leaves only room for differences in the entropy coding, not all the other encoder heuristics.
I don't think it works like that in any other modern encoder. They will all produce different pixels for the same quality setting at different effort settings.
The man page for cjxl is wrong: there is no guarantee that higher effort means smaller filesize. The only real guarantee is that it is slower. Generally it should also result in better consistency (less variance in visual quality for a given d setting, both within an image and across a corpus), and also in better overall (corpus-wide) compression in the 'bang for buck' sense (quality per bit). You can get better bang for buck by reducing bucks (bits) or by increasing bang (quality). It can certainly happen that lower effort results in a lower filesize and a lower quality. Whether that means it's "compressing better" depends on how much lower the filesize is and how much lower the quality is. The answer should typically be "no" though.
The man page for cjxl is wrong: there is no guarantee that higher effort means smaller filesize.
Thanks for the explanation (and your great work, again).
I have one last nitpick to add because I was one of those confused by how modern encoders work:
From users' perspective it would be nice if at least the "slowest" effort would be very likly, if not even guaranteed, to result in the best compression (aka smallest file size, speaking as a simple user).
It's reaonable that different efforts enable different encoder heuristics, but there could-should be a "best compression, no matter the cost" setting to get a baseline or to be on the safe side.
Looking at how slow effort 9 currently is that might be already covered though?
If not, maybe cjxl deserves a setting like optipng that tries all strategies and returns the most favorable result (again, looking at optipng no matter the cost).
For lossless compression, I think it's reasonable to have this expectation, and while we don't currently have this guarantee, if you find a counterexample we will treat it as a bug.
For lossy compression, it is much less clear what "best compression" means. You can always make a file smaller by reducing quality. But the point is not only to make the smallest possible file no matter what, it is also to make a file that looks 'good enough' according to the fidelity target the user chose. If the file is smaller, but it also looks worse than what you wanted to get, then I wouldn't call that "better compression".
Lower effort settings get much of their speed by using simpler/faster heuristics to make choices regarding quality. This means there will be more fluctuations in the actual quality you get: say for a d1 target, at e3 you will sometimes actually get d1.2, sometimes d0.9, sometimes d1.5, etc. At higher effort the amount of fluctuation is lower (thanks to spending more effort on 'getting it right'), so maybe those images now will be d1.05, d0.98, d1.07 instead at e8. In other words, the consistency of the results improves. But that means that for some images, the actual quality will be a bit lower and for others, it will be a bit higher. In the cases where it's higher, the file size will generally also be larger.
So the confusing thing is that most of the effort actually goes to improving consistency (which can lead to both smaller and larger files), not to improving compression itself (making files smaller for every single image).
| gharchive/issue | 2022-07-19T16:38:35 | 2025-04-01T06:44:47.390111 | {
"authors": [
"A-Conrath",
"RubenKelevra",
"damiencarrier",
"gitoss",
"jonsneyers"
],
"repo": "libjxl/libjxl",
"url": "https://github.com/libjxl/libjxl/issues/1632",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
998085956 | Support compiling GIMP plugin with glib < 2.62
glib 2.61 and earlier versions didn't have the helper function
g_clear_signal_handler now used by the GIMP plugin. This makes it
compile for older distributions like Debian buster.
@xiota FYI
I don't know if this is a hack or the proper way to fix it. You never know with glib.
Where is the plugin using g_clear_signal_handler? What is the compiler error/warning?
If this is needed, can it be put in a separate file, like gtk-compat.h/c? Or maybe there is a pre-existing compatibility package that can be used?
There are 9 calls to g_clear_signal_handler in the plugin. I don't know of any compat library for this. The point is to compile the plugin with an old glib version installed, so we would need to use a new compat library with the old glib. Given that this is a simple helper function I though the easiest was to include it here.
Here's an error message from the deploy pipeline in main (the one that creates the .deb files): https://github.com/libjxl/libjxl/runs/3621107836?check_suite_focus=true#step:13:2193
There is also a warning about JpegXlSaveOpts::SetQuality being unused BTW.
@deymo Sorry, I was grepping a folder with an older version of the plugin earlier. I was thinking about what if other parts of the plugin start using the same function. Then it would be nice if all the compatibility stuff were together. I suppose it can be split off later.
Right now it's setting quality/distance and updating distance/quality in separate steps when needed, but since quality and distance are linked, it should be using SetQuality and SetDistance to avoid introducing bugs by forgetting the update step. The change is in another branch I was working on.
Cool, I'll merge this now to get the build back to green then.
| gharchive/pull-request | 2021-09-16T11:10:53 | 2025-04-01T06:44:47.395884 | {
"authors": [
"deymo",
"xiota"
],
"repo": "libjxl/libjxl",
"url": "https://github.com/libjxl/libjxl/pull/595",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2275260112 | Error when compiling launcher with intrinsics
Trying to compile launcher code with intrinsics will cause error:
error: ‘void*’ is not a pointer-to-object type
in okl_memcpy_async
Fixed in #220
| gharchive/issue | 2024-05-02T11:07:19 | 2025-04-01T06:44:47.404267 | {
"authors": [
"IuriiKobein",
"ypanke"
],
"repo": "libocca/occa-transpiler",
"url": "https://github.com/libocca/occa-transpiler/issues/211",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1739275737 | PeerStore Merge Not Preserving Existing Protocols on Update
Version: libp2p@0.45.4
Platform: MacOS
Subsystem: Upgrader/PeerStore
Severity: High
Description:
While testing my code with upgraded libp2p@0.45.4., I encountered an issue where the protocol array in the peer store was not behaving as expected. While I expect that a new protocol should be appended to the existing list of protocols for a given peer, it appears that instead the new protocol is replacing the old entries.
Here is the test code:
nwaku = new NimGoNode(makeLogFileName(this));
await nwaku.start({
filter: true,
lightpush: true,
store: true,
relay: true,
});
waku = await createLightNode({
staticNoiseKey: NOISE_KEY_1,
libp2p: { addresses: { listen: ["/ip4/0.0.0.0/tcp/0/ws"] } },
});
await waku.start();
await waku.dial(await nwaku.getMultiaddrWithId());
await waitForRemotePeer(waku, [
Protocols.Filter,
Protocols.LightPush,
Protocols.Store,
]);
console.log((await waku.libp2p.peerStore.all()).map((p) => p.protocols));
This can be run in an environment through this commit message: https://github.com/waku-org/js-waku/pull/1385/commits/e0f7af3e06a41f47e8bd064699ce8b29e353d55b
This test code is throwing an error: "Error: Failed to find known peer that registers protocols: /vac/waku/filter-subscribe/2.0.0-beta1". It seems that only one protocol is registered at a time according to the output of console.log((await waku.libp2p.peerStore.all()).map((p) => p.protocols));.
However, I expect the peer to have multiple protocols registered: Filter, LightPush, and Store.
I found that changing the upstream libp2p code with upgrader.ts (https://github.com/libp2p/js-libp2p/blob/41641f1a7656aa654234f6f849b1749786867121/src/upgrader.ts#LL465C11-L467C13) to the following fixed this issue.
await this.components.peerStore.merge(remotePeer, {
protocols: [...protocols, protocol]
})
Steps to reproduce the error:
Run the above test code (through the commit message)
clone the repo
npm i && npm run build
cd packages/tests && npm run test:node
Observe the mentioned error
Check the registered protocols for a peer and observe that only the latest one is present.
Note: This is part of a larger migration PR for js-waku from v0.42 to 0.45.4: https://github.com/waku-org/js-waku/pull/1385
cc @achingbrain
Thanks for spotting this issue. .peerStore.merge does exactly that, it deep-merges all passed fields, see the test for this here.
I've dug into it a little and it seems the remote peer does not send a signed peer record during identify, this hits this branch and so the sent protocols are never persisted.
This is a bug as we should persist the identify information, but replace any addresses with those from the signed peer record if it's present.
TBH I thought all libp2p implementations sent signed peer records by now - what is the remote peer type in this test?
Yes, I'll get a fix out very soon.
TBH I thought all libp2p implementations sent signed peer records by now
I'm pretty sure this is disabled on most, if not all eth consensus clients (which use a very small/safe subset of libp2p features)
I was about to open an issue but I believe this covers my issue, I'll add the details here. Please let me know if it's unrelated.
I'm observing that rust-libp2p can no longer be used as a relay connection with js-libp2p since v0.45.0.
Steps to Reproduce/summary>
Run the rust-libp2p relay using any transport. Example using tcp will do https://github.com/libp2p/rust-libp2p/tree/master/examples/relay-server
Run the js-libp2p client which jus
import { multiaddr } from '@multiformats/multiaddr'
import { createLibp2p } from 'libp2p'
import { circuitRelayTransport } from 'libp2p/circuit-relay'
import { tcp } from '@libp2p/tcp'
import { mplex } from '@libp2p/mplex'
import { yamux } from '@chainsafe/libp2p-yamux'
import { noise } from '@chainsafe/libp2p-noise'
import { identifyService } from 'libp2p/identify'
export async function privateLibp2pNode () {
const node = await createLibp2p({
addresses: {
listen: ['/ip4/0.0.0.0/tcp/0']
},
transports: [
tcp(),
circuitRelayTransport({
discoverRelays: 1,
}),
],
streamMuxers: [yamux(), mplex()],
connectionEncryption: [noise()],
peerDiscovery: [],
services: {
identify: identifyService()
}
})
return node
}
(async () => {
const node = await privateLibp2pNode()
console.log(node.peerId)
node.addEventListener("peer:identify", (event) => {
const identify = event.detail;
console.log('peer:identify', identify.peerId, identify.protocols)
})
node.addEventListener("self:peer:update", (event) => {
console.log('self:peer:update', node.getMultiaddrs())
})
const ma = multiaddr('<addr of the rust-libp2p relay>')
await node.dial(ma)
})()
Actual Result
Relayed address is not listed on peer:update
self:peer:update [
Multiaddr(/ip4/127.0.0.1/tcp/58971/p2p/12D3KooWEcek3uKFuLasuAMcHTMzv5rxHqTm4W3YrQMjSjNFQtAs),
Multiaddr(/ip4/192.168.178.106/tcp/58971/p2p/12D3KooWEcek3uKFuLasuAMcHTMzv5rxHqTm4W3YrQMjSjNFQtAs)
]
Expected Result
Peer update includes the relayed address on peer:update
self:peer:update [
Multiaddr(/ip4/127.0.0.1/tcp/59802/p2p/12D3KooWFyhdyjdcxWyFeQtcVJCDW3mkja7jyRZ52cYNuhKPeM65),
Multiaddr(/ip4/192.168.178.106/tcp/59802/p2p/12D3KooWFyhdyjdcxWyFeQtcVJCDW3mkja7jyRZ52cYNuhKPeM65),
Multiaddr(/ip4/127.0.0.1/tcp/54336/p2p/12D3KooWCNibjstG4xDuh5ybRN8knVoYfGD7kuinJYEyaHVvob5d/p2p-circuit/p2p/12D3KooWFyhdyjdcxWyFeQtcVJCDW3mkja7jyRZ52cYNuhKPeM65),
Multiaddr(/ip4/192.168.178.106/tcp/54336/p2p/12D3KooWCNibjstG4xDuh5ybRN8knVoYfGD7kuinJYEyaHVvob5d/p2p-circuit/p2p/12D3KooWFyhdyjdcxWyFeQtcVJCDW3mkja7jyRZ52cYNuhKPeM65)
]
rust-libp2p identify protocol does not implement signedPeerRecord https://github.com/libp2p/rust-libp2p/issues/4017
There appears to be an undocumented breaking change between js-libp2p v0.44.0 & v0.45.0 which requires signedPeerRecord for this to work.
I was about to open an issue but I believe this covers my issue, I'll add the details here. Please let me know if it's unrelated.
I'm observing that rust-libp2p can no longer be used as a relay connection with js-libp2p since v0.45.0.
Steps to Reproduce
rust-libp2p identify protocol does not implement signedPeerRecord libp2p/rust-libp2p#4017 There appears to be an undocumented breaking change between js-libp2p v0.44.0 & v0.45.0 which requires signedPeerRecord for this to work.
Good catch @Marcel-G thanks for spotting this. I've merged @achingbrain 's fix https://github.com/libp2p/js-libp2p/pull/1790 which should be released in v0.45.5
v0.45.5 has been released so this should be fixed after deleting node_modules and any lock files and reinstalling your deps.
| gharchive/issue | 2023-06-03T08:15:39 | 2025-04-01T06:44:47.475026 | {
"authors": [
"Marcel-G",
"Menduist",
"achingbrain",
"danisharora099",
"maschad"
],
"repo": "libp2p/js-libp2p",
"url": "https://github.com/libp2p/js-libp2p/issues/1789",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2706667989 | fix: do not add peers to routing table during RPC handling
Description
I was having a issue where 2 docker dht nodes failed to dht ping each other on startup. The streams would abort due to a read timeout.
Also removed key from the ping tests as not required.
Notes & open questions
I don't think there needs to be a call to this.routingTable.add in onIncomingStream as it should be handled by the topology register. Although alternatively maybe it could be moved to after the pipe?
Couldn't replicate the issue on localhost between 2 non-dockerised instances although I can provide docker files / test repo / logs if needed
Change checklist
[x] I have performed a self-review of my own code
[ ] I have made corresponding changes to the documentation if necessary (this includes comments as well)
[ ] I have added tests that prove my fix is effective or that my feature works
Makes sense, thanks for opening this. I think we might also end up trying to add DHT clients to the routing table like this, which would later fail because they wouldn't respond to a ping so it's a waste of resources.
| gharchive/pull-request | 2024-11-30T04:47:09 | 2025-04-01T06:44:47.478755 | {
"authors": [
"achingbrain",
"dozyio"
],
"repo": "libp2p/js-libp2p",
"url": "https://github.com/libp2p/js-libp2p/pull/2866",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
459830910 | fix most of the typos
Motivation
fix typos
Have you read the Contributing Guidelines on pull requests?
yes
Test Plan
no new feature and 'cargo test' passed
Related PRs
none
👍
As a semi-related remark, I think line breaks before and after links in the readme markdowns help readability and could be added. I.e. writing
the text like this with
[some description with a link to some page](https://github.com/libra/libra/pull/219)
like this
instead of
the text like this with [some description with a link to some page](https://github.com/libra/libra/pull/219) like this
As displayed on github, those will appear exactly the same but the later kind of markdown is harder to read, especially if the files link to 5 different web pages in one paragraph.
E.g. kind of text wall is not necessary:
I rebased this branch from master to fix the conflict and modified the readme as suggested.
Resolve confit
I rebased it again.
@lhfbc
Thanks for the contribution! This looks like a good change, but needs a rebase.
However, due to some CI changes put in place this past week, to get CI to run we would need you to push against your branch again (e.g in a rebase), or to close this PR and make a new one. Going forward this won't be an issue, but unfortunately, we do not have a nice way to force CI to run for older PRs.
| gharchive/pull-request | 2019-06-24T11:10:44 | 2025-04-01T06:44:47.528404 | {
"authors": [
"Nikolaj-K",
"huitseeker",
"lhfbc"
],
"repo": "libra/libra",
"url": "https://github.com/libra/libra/pull/210",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1917837444 | Spotify Connect does not show up with zeroconf
Describe the bug
Spotify connect device does not show up on the network with librespot-java.
To Reproduce
Download latest librespot player jar file
Run with sudo java -jar librespot-player-1.6.3.jar
Make sure ZEROCONF is used in config.toml - restart if needed
Expected behavior
Zeroconf announcement should be visible on network and show up as Spotify Connect device in the app on devices connected to the same local network
Screenshots/Stracktraces/Logs
This is in an Ubuntu VM with bridged networking and no firewall on the Ubuntu machine. I'm running into this same issue with both Spocon and the librespot-java player. I decided to try the original librespot and it works fine with zeroconf, the connect item shows up in the Spotify app and the announcement is visible on the network. Below are the ports and protocols used by librespot-java (java) and librespot respectively
Attached are the outputs from Spocon and from librespot-java. I noticed that no addresses are listed in the announcement but I'm not sure if that's intended
INFO Zeroconf:500 - Announced Service{alias='librespot-java', service='spotify-connect', port=45337, text={VERSION=1.0, CPath=/, Stack=SP},
Version/Commit
librespot-player-1.6.3.jar
Logs:
librespot-java_output.txt
Spocon_log.txt
I dug a bit further into the zeroconf-java repository and it looks like the IP addresses should be added, so I think this is the problem. I don't understand how everything going on with the zeroconf and service code, but somewhere along the chain there must be an issue with adding the addresses from my NIC.
https://github.com/devgianlu/zeroconf-java/tree/1c98d74ecccc2909b70b46a6113d8ba13bb23f93
src/main/java/xyz/gianlu/zeroconf/Zeroconf.java
src/main/java/xyz/gianlu/zeroconf/Service.java
I've tried using all interfaces and entering specifically the "enp0s3" NIC into the config file with no luck. Here's the NIC info on my VM:
docker0: flags=4099<UP,BROADCAST,MULTICAST> mtu 1500
inet 172.17.0.1 netmask 255.255.0.0 broadcast 172.17.255.255
ether 02:42:ec:70:ca:eb txqueuelen 0 (Ethernet)
RX packets 0 bytes 0 (0.0 B)
RX errors 0 dropped 0 overruns 0 frame 0
TX packets 0 bytes 0 (0.0 B)
TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
enp0s3: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
inet 192.168.1.220 netmask 255.255.255.0 broadcast 192.168.1.255
inet6 fd25:3921:203e:41f5:4a53:e97a:4eab:1ecf prefixlen 64 scopeid 0x0<global>
inet6 fe80::bc80:5771:8d4f:c19 prefixlen 64 scopeid 0x20<link>
inet6 fd25:3921:203e:41f5:e692:c505:d205:f7cd prefixlen 64 scopeid 0x0<global>
ether 08:00:27:85:9b:a9 txqueuelen 1000 (Ethernet)
RX packets 1455773 bytes 2107724130 (2.1 GB)
RX errors 0 dropped 0 overruns 0 frame 0
TX packets 149006 bytes 14403116 (14.4 MB)
TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
lo: flags=73<UP,LOOPBACK,RUNNING> mtu 65536
inet 127.0.0.1 netmask 255.0.0.0
inet6 ::1 prefixlen 128 scopeid 0x10<host>
loop txqueuelen 1000 (Local Loopback)
RX packets 1781 bytes 197796 (197.7 KB)
RX errors 0 dropped 0 overruns 0 frame 0
TX packets 1781 bytes 197796 (197.7 KB)
TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
| gharchive/issue | 2023-09-28T15:58:12 | 2025-04-01T06:44:47.612524 | {
"authors": [
"dolanjp"
],
"repo": "librespot-org/librespot-java",
"url": "https://github.com/librespot-org/librespot-java/issues/716",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
644142058 | Use a monotonic clock on macOS
Resolves: https://github.com/libuv/libuv/issues/2891
Alternative with weak-linking trickery so macOS <10.12 can still be used: #2893
Fixed in 4685be2.
| gharchive/pull-request | 2020-06-23T20:58:20 | 2025-04-01T06:44:47.702513 | {
"authors": [
"WillerZ",
"bnoordhuis"
],
"repo": "libuv/libuv",
"url": "https://github.com/libuv/libuv/pull/2892",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1214971958 | New version fails on SparkyCI build
Hi guys.
This build fails - 463df4d - http://sparrowhub.io:2222/report/228
This build d28d6e6 (prior to 0.6.15 release) - succeeds - http://sparrowhub.io:2222/report/152
Commits relate to the fork - https://github.com/melezhik/LibXML-raku/commits/master , build started to fail after I merged in changes from the upstream - https://github.com/melezhik/LibXML-raku/commit/463df4d7b2e71c110b9fe62fb2ac072e033646d3
HTH
Aleksei
Sorry for the noise - now the same commit - succeeds - looks like SparkyCI quirks 😉- http://sparrowhub.io:2222/report/230
| gharchive/issue | 2022-04-25T19:53:00 | 2025-04-01T06:44:47.718358 | {
"authors": [
"melezhik"
],
"repo": "libxml-raku/LibXML-raku",
"url": "https://github.com/libxml-raku/LibXML-raku/issues/71",
"license": "Artistic-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2544027473 | 🛑 Landing page is down
In e41f614, Landing page (https://licenseware.io) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Landing page is back up in eaf82f8 after 1 hour, 33 minutes.
| gharchive/issue | 2024-09-24T00:53:37 | 2025-04-01T06:44:47.721128 | {
"authors": [
"meysam81"
],
"repo": "licenseware/uptime",
"url": "https://github.com/licenseware/uptime/issues/249",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
281287140 | URL constructor does not accept objects. The browsers that implement the spec do.
If passing an object to the URL constructor, in Firefox and Chrome it will use the toString() property of the object as the URL to process. This is most obvious (and useful) when passing a Location object such as window.location.
The polyfill does not do this; it simply checks if the parameter is typeof string, and if not it throws an exception.
Line 173 of url-polyfill.js should be either removed or accept objects and call their toString() method.
Before the line
if(typeof url !== 'string') throw new TypeError('Failed to construct \'URL\': Invalid URL');
we can add
if (typeof url === 'object') url = url.toString();
which will solve the problem
I made the changes. But take care, the URL polyfill is not an attempt to be 100% in concordance with the specs else all type checks, special char escape, etc... would result in a huge file size. The URL class expects a string as input you should not pass something else even if the browser correct it ;)
| gharchive/issue | 2017-12-12T07:45:19 | 2025-04-01T06:44:47.756983 | {
"authors": [
"lifaon74",
"llimos"
],
"repo": "lifaon74/url-polyfill",
"url": "https://github.com/lifaon74/url-polyfill/issues/12",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2303494405 | Make it a taskwarrior & timewarrior web client!
I've recently discovered that taskwarrior cli is fantastic and it a big community that standardized todo/task tracking on OS level. I'm currently learning timewarrior , it is not as intuitive to use however all the foundations/standardization is now there for operating systems.
This could be a great web client for taskwarrior and timewarrior ;)
Hi @izelnakri! Thank you for suggestion! Really neat tools! I think it's possible to at least add support for timewarrior data import.
To clarify scope of this app - it's kinda pet-projects to battle-test https://github.com/lifeart/glimmer-next and spot real-usage problems.
Here is a list of apps I made: https://github.com/lifeart/template-gxt/issues/1
| gharchive/issue | 2024-05-17T20:01:16 | 2025-04-01T06:44:47.766213 | {
"authors": [
"izelnakri",
"lifeart"
],
"repo": "lifeart/track-it",
"url": "https://github.com/lifeart/track-it/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
361140231 | Alphabetized Team Page
Currently the team members are rendered by their id — would be great if it was alphabetical by last name (from their photo or page url).
fixed by #367
| gharchive/issue | 2018-09-18T05:24:16 | 2025-04-01T06:44:48.175317 | {
"authors": [
"plhnk"
],
"repo": "liferay-design/liferay.design",
"url": "https://github.com/liferay-design/liferay.design/issues/72",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
562579970 | fix: make sort-import-destructures work with long lists
The old implementation happened to work with short lists but it made too many assumptions about the order in which pairs in the list would be compared. As soon as we tried running this on files with many names being destructured (like the one in the included regression test), then the assumption (that we'd only ever call element in list-order) was invalidated.
V8 actually uses Timsort under the covers (https://v8.dev/blog/array-sort), but that is besides the point. We need to make sure the comparator function passed to sort behaves identically regardless of comparison order.
| gharchive/pull-request | 2020-02-10T14:00:51 | 2025-04-01T06:44:48.749625 | {
"authors": [
"wincent"
],
"repo": "liferay/eslint-config-liferay",
"url": "https://github.com/liferay/eslint-config-liferay/pull/146",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
264601593 | adding csv suffix to known suffixes
Change: Add csv to the list of known suffixes.
Motive: @joescii @Bhashit and I was look into a bug that prevent us accessing a local csv resource from webapp. The request got 200 response, but the content is not csv content but base html page, so the request got redirected. Add csv to known suffix list fixed this issue.
Yeah, I was pretty surprised we didn't already have csv. @Bhashit even pointed out that we have tsv lol
If there are no objections to this PR, I'll hit the merge button later this week.
There have been a significant number of objections on the ML. I haven’t been following lately. What was the conclusion of that?
Thanks for pointing that out, @farmdawgnation. Even though I just caught up on the ML yesterday, I somehow missed the thread. Closing the PR per the discussion.
| gharchive/pull-request | 2017-10-11T14:09:51 | 2025-04-01T06:44:48.822964 | {
"authors": [
"farmdawgnation",
"joescii",
"snowf0x"
],
"repo": "lift/framework",
"url": "https://github.com/lift/framework/pull/1922",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
324421855 | Integration test coverage
We need to settle on some way to test NSM functionally.
Closing this now, the initial integration tests have merged, we'll open another PR if we need to add more.
| gharchive/issue | 2018-05-18T13:58:00 | 2025-04-01T06:44:48.828349 | {
"authors": [
"mestery"
],
"repo": "ligato/networkservicemesh",
"url": "https://github.com/ligato/networkservicemesh/issues/50",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1175205194 | Make ReadinessLevel's configurable
Removes Lightbend hardcoded values for readinessLevel's and allows them to be configurable. This is done by adding a new sbt key called readinessLevels which is a Map[String, ReadinessLevel] where the String is the config key in project-info.conf and the value is a Scala type that implements ReadinessLevel. I decided to make this confgurable via sbt settings instead of typesafe config because the typesafe config inheritance doesn't via nested projects doesn't work with sbt plugins (ontop of this defining the html rendering via name in hoconf is not that nice).
In order to test that this is actually working properly, I created a sample https://github.com/mdedetrich/sbt-paradox-lightbend-project-info and using sbt localPublish along with temporarily setting the version to -SNAPSHOT I republished https://github.com/akka/akka-paradox locally to test it with https://github.com/akka/alpakka docs/makeSite. Note that https://github.com/mdedetrich/sbt-paradox-lightbend-project-info implements the sbt-plugin to automatically trigger when you include it as a dependency (the alternative is to use noTrigger which means you manually have to enable it with .enablePlugins(LightbendParadoxInfoPlugin)).
This means that once this pull request and released the following things to need to be done
Create a new project such as https://github.com/mdedetrich/sbt-paradox-lightbend-project-info that uses the newly released sbt-paradox-project-info. Note that you can just copy https://github.com/mdedetrich/sbt-paradox-lightbend-project-info if you wish, its pretty much ready to go and just needs to be put under lightbend org + released
Update https://github.com/akka/akka-paradox to point to sbt-paradox-lightbend-project-info instead of sbt-paradox-project-info and make a new release
Gradually update akka/lightbend projects to use the newly released akka-paradox. This is very easy due to the fact that https://github.com/mdedetrich/sbt-paradox-lightbend-project-info automatically triggers when you include it as a dependency (even as a transitive one) so all you need to do is to just bump the value of a single dependency in project/plugins.sbt.
Resolves: https://github.com/lightbend/sbt-paradox-project-info/issues/17
Thanks for the merge, let me know when a a new release is made and what you plan to do regarding https://github.com/mdedetrich/sbt-paradox-lightbend-project-info
Releasing has started and I created and (so far) empty https://github.com/lightbend/sbt-paradox-lightbend-project-info.
@ennru So it turns out you cannot fork an empty repo https://stackoverflow.com/questions/39013301/how-do-i-fork-an-empty-repository-on-github . Can you just add a LICENSE file and then I can add the rest?
I pushed the contents from your repo there.
Perfect thanks, just created a PR on https://github.com/lightbend/sbt-paradox-lightbend-project-info/pull/1 that points to this new release.
| gharchive/pull-request | 2022-03-21T10:48:55 | 2025-04-01T06:44:48.838005 | {
"authors": [
"ennru",
"mdedetrich"
],
"repo": "lightbend/sbt-paradox-project-info",
"url": "https://github.com/lightbend/sbt-paradox-project-info/pull/34",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1346533167 | Update slf4j-api to 2.0.0
Updates org.slf4j:slf4j-api from 1.7.36 to 2.0.0.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.slf4j", artifactId = "slf4j-api" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "@monthly" },
dependency = { groupId = "org.slf4j", artifactId = "slf4j-api" }
}]
labels: library-update, early-semver-major, semver-spec-major, commit-count:1
Superseded by #366.
| gharchive/pull-request | 2022-08-22T14:41:35 | 2025-04-01T06:44:48.841435 | {
"authors": [
"scala-steward"
],
"repo": "lightbend/scala-logging",
"url": "https://github.com/lightbend/scala-logging/pull/361",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
73857741 | I want to use the latest ResponseFilter interface
According to the README and ResponseFilter.java, ResponseFilter.filterResponse method has 3 argument, but filterResponse method defined on my local 2.1.0-beta-1-SNAPSHOT jar of browsermob-proxy
( https://oss.sonatype.org/content/repositories/snapshots/net/lightbody/bmp/browsermob-core/
and https://oss.sonatype.org/content/repositories/snapshots/net/lightbody/bmp/browsermob-proxy/ ) has 2 argument.
I think these SNAPSHOT jars are not the latest, and I want to use the SNAPSHOT jar for the latest master..
Or should I see other SNAPSHOT repository?
I just uploaded the latest, so give it another try.
It works. Thank you!
| gharchive/issue | 2015-05-07T06:01:48 | 2025-04-01T06:44:48.846683 | {
"authors": [
"NozomiIto",
"jekh"
],
"repo": "lightbody/browsermob-proxy",
"url": "https://github.com/lightbody/browsermob-proxy/issues/222",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
189283523 | Too many connect failed
When I use BMP 2.1.1 and selenium to visit a website, there are a lot of connect failed.
But when I use BMP 2.1.0 it will playback succeed.
I use firebug to get the har and there are no connect.
My system info is:Red Hat Enterprise Linux Server release 7.0 3.10.0-123.el7.x86_64
My test URL is:https://www.mevsnet.net
My system info is:Red Hat Enterprise Linux Server release 7.0 3.10.0-123.el7.x86_64
What do the stack traces from the errors look like?
| gharchive/issue | 2016-11-15T02:31:51 | 2025-04-01T06:44:48.849091 | {
"authors": [
"jekh",
"leoforever"
],
"repo": "lightbody/browsermob-proxy",
"url": "https://github.com/lightbody/browsermob-proxy/issues/559",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2039063807 | Feature/create app 2
Added support for creating the Lightning 3 app.
Thanks for the PR so far Sandeep. I've left some comments inline.
I also noticed that when I cancel the creation flow in the middle (ctrl+c) it throws an error instead of gracefully handling it.
Hey @michielvandergeest , I have handled this case. Please have a look.
Linking to issue #43
| gharchive/pull-request | 2023-12-13T06:46:51 | 2025-04-01T06:44:48.854242 | {
"authors": [
"michielvandergeest",
"sandeep-vedam"
],
"repo": "lightning-js/blits",
"url": "https://github.com/lightning-js/blits/pull/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2091071045 | Feat/image worker
Add a web worker to load images.
@frank-weindel - Ideally the numWorkers should be passed into the stage and passed through. I wasn't sure the cleanest way of doing that part.
LGTM!
| gharchive/pull-request | 2024-01-19T18:27:37 | 2025-04-01T06:44:48.855285 | {
"authors": [
"chiefcll",
"frank-weindel"
],
"repo": "lightning-js/renderer",
"url": "https://github.com/lightning-js/renderer/pull/128",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1957364086 | universe: make additional details optional to speed up query
This makes the additional DB query for each grouped asset optional and hides it behind an additional gRPC/REST parameter.
Was this guided optimization via EXPLAIN ANALYZE, or something else? Do you have that dump we can use to add more detail to the commit body?
Ah reading more now, I see it's to just leave off some extra queries if/when only the roots are needed.
I think we'll want to have the REST interface display by default, or just make sure the new query param there is documented.
I think we'll want to have the REST interface display by default, or just make sure the new query param there is documented.
Yeah, I should've added more info here, sorry.
So this extra loop to fetch all assets within a group was a user experience improvement added in https://github.com/lightninglabs/taproot-assets/pull/492 to better understand the output of the roots call.
But because this does an extra DB call for each group, it is very expensive. So we definitely should NOT turn this on for universe sync and the REST call. I checked with @jamaljsr and we don't use the extra information on Terminal Web.
The extra info is only required in Polar where Jamal is going to add the query parameter (but there it will just query the local universe).
But I think we do want the CLI to remain at the previous behavior, so I'm going to fix that.
| gharchive/pull-request | 2023-10-23T14:51:54 | 2025-04-01T06:44:48.864706 | {
"authors": [
"Roasbeef",
"guggero"
],
"repo": "lightninglabs/taproot-assets",
"url": "https://github.com/lightninglabs/taproot-assets/pull/612",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1584190728 | Currencies getting lost when minting multiple currencies with skip_batch in one block
When I try to minting multiple currencies in one block currencies I submitted for minting get "lost" and never appear
for i in {1..20}; do tarocli assets mint --type=normal --name=NewCurrencyFromLoop$i --meta="test minting number $i" --skip_batch --supply=10000$i; done
This returns:
{
"batch_key": "036099a9139325a058325b1754967139ef50b21052cc3aef7a9dc95914981c4765"
}
{
"batch_key": "02a51a403ce9b5dfa70d9e1ade216a52d48eea551f4aa194699cd8f80eb6db0fd1"
}
{
"batch_key": "0300658b33d38d598c14f33305e2b905c1a5b45718bba010a9fada9f482a303cb2"
}
{
"batch_key": "02b139344bccb3e63ab1178560adb35fd31672f5ef555f40ec25d97de067dd282b"
}
...
Wait for a 5 blocks as that should be long enough even on testnet.
$ tarocli assets balance | grep NewCurrencyFromLoop
only gives
"name": "NewCurrencyFromLoop1"
So all of the 20 currencies that were minted got lost.
I know that there is the options to enable batches but still it happens to be a lot that I have to submit multiple mint commands in one block.
Summary
hey @snow884 if I understand your intention you'd like to mint assets NewCurrencyFromLoop1..20 in a single block.
Your issue may be described as:
Calling --skip_batch after a minting call, but before a minting txn is included in a block, the daemon doesn't rewrite a bitcoin transaction with the subsequent minting events.
Fixes
Quick
The short term amelioration would be: remove --skip_batch from your loop. Then all the minting transactions will be grouped together in a minting batch which ought to included NewCurrencyFromLoop1..20
Long Term
Longer term fix could be to either
Throw a warning when calling mint between a mint call but before a block which includes the first mint is reeived
Modify the minting function so that it rewrites the minting transaction if a subsequent mint --skip_batch is called to include the subsequent minting
Modify the minting function so that it uses a different UTXO to mint subsequent calls (I need to check on where the delta would be on this work)
Thank you for quick reply @dstadulis .
I will try minting the currencies without --skip_batch
When I try to minting multiple currencies in one block currencies I submitted for minting get "lost" and never appear
Yeah skip_batch means that you want to mine it asap, this won't make them all in a single block. You want to drop that, then let the batch "tick".
We're also working on a way to atomically create assets in a singel batch as well: https://github.com/lightninglabs/taro/issues/233 cc @jharveyb
Should be fixed with https://github.com/lightninglabs/taproot-assets/releases/tag/v0.2.0
| gharchive/issue | 2023-02-14T13:42:42 | 2025-04-01T06:44:48.873126 | {
"authors": [
"Roasbeef",
"dstadulis",
"jharveyb",
"snow884"
],
"repo": "lightninglabs/taro",
"url": "https://github.com/lightninglabs/taro/issues/266",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
867191661 | Documentation indicates lnd will run with bitcoind using cookie authentication for RPC, but it throws an error
Background
This is either an issue with the documentation or the implementation:
The lnd documentation states lnd can be run with bitcoind using cookie authentication here: https://github.com/lightningnetwork/lnd/blob/master/docs/INSTALL.md#using-bitcoind-or-litecoind "The auth parameters rpcuser and rpcpass parameters can typically be determined by lnd for a bitcoind instance running under the same user, including when using cookie auth. In this case, you can exclude them from the lnd options entirely."
When running lnd with bitcoind as a backend, if rpcuser and rpcpassword are not set in bitcoind environment, bitcoind uses cookie authentication for rpc as documented here: https://github.com/bitcoin/bitcoin/blob/7fcf53f7b4524572d1d0c9a5fdc388e87eb02416/doc/init.md
When bitcoind is running with cookie authentication for RPC, lnd checks explicity for rpcuser and rpcpassword and if they do not exist an error is generated. This is implemented on lines 1496-1503 in config.go ... here: https://github.com/bitcoin/bitcoin/blob/7fcf53f7b4524572d1d0c9a5fdc388e87eb02416/doc/init.md
Your environment
version of lnd: lnd version 0.12.1-beta commit=v0.12.1-beta
which operating system (uname -a on *Nix): Linux 5.8.0-1021-raspi #24-Ubuntu SMP PREEMPT Tue Apr 13 11:00:57 UTC 2021 aarch64 aarch64 aarch64 GNU/Linux
version of btcd, bitcoind, or other backend: bitcoind Bitcoin Core version v0.21.0
any other relevant environment details
Steps to reproduce
install run, and sync bitcoind Bitcoin Core version v0.21.0 with no rpcuser or rpcpassword set (forcing bitcoind to use cookie authentication)
install lnd, do not include rpcuser or rpcpassword in .lnd/lnd.conf
Expected behavior
config.go should not throw an error, unless it is not possible (or recommended) for lnd to use bitcoind with cookie authentication
Actual behaviour
Receive this error message: "unable to load RPC credentials for bitcoind: please set all or none of bitcoind.rpcuser, bitcoind.rpcpass, bitcoind.zmqpubrawblock, bitcoind.zmqpubrawtx"
Can you post your full lnd config here please? It sounds to me like you need to remove the ZMQ parameters from it as well as they are also read from the bitcoin.conf (and the example in the INSTALL.md file is wrong).
Your solution is correct and resolved the issue. Removing both ZMQ parameters from lnd.conf resulted in this response when running lnd:
Attempting automatic RPC configuration to bitcoind
Automatically obtained bitcoind's RPC credentials
I read the documentation again and it isn't wrong per se, just not very clear. The section could definitely be improved.
Though the error message is pretty clear in what needs to be done IMO. Closing for now, but feel free to add a PR to improve the documentation.
Thank you
| gharchive/issue | 2021-04-26T02:08:40 | 2025-04-01T06:44:48.915113 | {
"authors": [
"de6df1re",
"guggero"
],
"repo": "lightningnetwork/lnd",
"url": "https://github.com/lightningnetwork/lnd/issues/5241",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
467939211 | [feature] Information picker
We want to inspect vertex position/uv/normal, material id, weights value, etc of loaded glTF model.
[x] Pick bones/vertex/face using ray casting
[ ] Display information of selected bone/vertex/face(position, normal, uv, material/texture, etc)
infrastructure for ray cast picking works on my hand to find the face already. Currently working on adding the vertex and bone picking, will push this soon.
Could you take a look at this : https://github.com/lighttransport/nanort/pull/47
@syoyo Take a look at the GIF above.
I still need to refine it a tiny bit, but here's the gist of what's going on
clicking on view-port check which known sub-mesh is under the cursor
if you haven't clicked on a void space, we found what part of the geometry was just under the cursor, we shoot a ray to find a triangle (we get mesh data, transform it to world space, construct NanoRT geometry, shot ray from camera through cursor)
from finding triangle, we calculate (squared) distance from each vertex to the hit point. We choose the vertex that was closest, and it's marked as the "active vertex"
each vertex on the intersected polygon is flagged as the "active polygon"
active polygon indices are sent to shader, shader knows when current fragment was from the clicked polygon, if so, it will be rendered in bright orange
active vertex is drawn over with a big bright orange point
If mesh has skinning, skinning weights is checked. Joint that has the highest weight on vertex is set as current "active joint". This permit to control it with 3D gizmo
A window is created when valid selection is present, it contains info about the selected geometry (still need to add better display to that, and uv/normals are missing in the animation above). If mesh skinned, all the skinning data (weight/joint arrays for vertex) are shown.
Current quirks are :
As of now, this only works on "TRIANGLE_LIST" GLTF primitives. Code that interact with NanoRT has to take the TRIANGLE_{STRIP/FAN} case into account
Skinned mesh needs to have CPU-side vertex data available. Thus, selection of skinned mesh only works when "soft skinning" mode is turned on. Current soft-skinning implementation is slow, and currently, this mode is turned off by default. I'm probably going to make it "on" by default. CPU Skinning performance can be improved later easily (by threading it, and by not recomputing things if joint pose data did not change.)
Cat you take a look to this ?
https://github.com/lighttransport/nanort/pull/48
| gharchive/issue | 2019-07-15T04:47:21 | 2025-04-01T06:44:48.926174 | {
"authors": [
"Ybalrid",
"syoyo"
],
"repo": "lighttransport/gltf-insight",
"url": "https://github.com/lighttransport/gltf-insight/issues/29",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
199213180 | Tab completion in REPL on unknown objects acts un-shell-like
It returns completions of the existing parent instead of nothing, that I expected for non-existing objects.
Real life story: I wanted to access environment variables, and typed System.env. and pressed tab - which hinted me getProperty and thus I thought I may call System.env.getProperty() even though I misremembered and there isn't a System.env at all.
Loading...
Welcome to the Ammonite Repl 0.7.9
(Scala 2.11.8 Java 1.7.0_79)
@ System.
arraycopy err getProperty in loadLibrary runFinalization setProperties
clearProperty exit getSecurityManager inheritedChannel mapLibraryName setErr setProperty
console gc getenv lineSeparator nanoTime setIn setSecurityManager
currentTimeMillis getProperties identityHashCode load out setOut
@ System.foobar.
arraycopy err getProperty in loadLibrary runFinalization setProperties
clearProperty exit getSecurityManager inheritedChannel mapLibraryName setErr setProperty
console gc getenv lineSeparator nanoTime setIn setSecurityManager
currentTimeMillis getProperties identityHashCode load out setOut
@ System.gc.
!= == asInstanceOf equals getClass hashCode isInstanceOf toString |>
@ foobar
This probably belongs under https://github.com/lihaoyi/Ammonite/issues/275
| gharchive/issue | 2017-01-06T15:00:17 | 2025-04-01T06:44:49.076038 | {
"authors": [
"lihaoyi",
"wbazant"
],
"repo": "lihaoyi/Ammonite",
"url": "https://github.com/lihaoyi/Ammonite/issues/544",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
50063525 | Find Commands highlight
See here
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/6522212-find-commands-highlight?utm_campaign=plugin&utm_content=tracker%2F282001&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282001&utm_medium=issues&utm_source=github).
Migrated to limetext/lime-backend#3.
| gharchive/issue | 2014-11-25T18:07:02 | 2025-04-01T06:44:49.094767 | {
"authors": [
"erbridge",
"zoli"
],
"repo": "limetext/lime",
"url": "https://github.com/limetext/lime/issues/452",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
60557019 | Does not work
Hello @limonte , i appreciate your work and i have been trying to implement sweetalert2, but unfortunately the alerts do not popup at all in your current versions. I request you to please have a look at the code.
Many thanks!
@prateekvarma can you please show me some code. How you're using sweetalert2 and is there any errors in console?
Hello @limonte , thank you for the reply. Whats happening is - when i download the ZIP file from github, and after opening the index.html file from the "example" folder, none of the buttons show a sweetAlert. You can check it yourself, and for some strange reason it does not work...
I can confirm this, it is not an issue with sweetalert2 though. Its just sweetalert2.min.js isn't loaded.
Just alter the html to reflect the correct file paths, e.g.
<!-- This is what you need -->
<script src="../dist/sweetalert2.min.js"></script>
<link rel="stylesheet" href="../dist/sweetalert2.css">`
Ah dear lord LOL ... yes @andyharney , you're right.
To be exactly specific, the sweetalert2.min.js file, and the sweetalert2.css files both miss one 'dot'.
It'll be great if you guys could add it in the repository, so it won't confuse anyone.
Many thanks, you guys rock!
@andyharney thanks for helping!
@limonte No problem. Thanks for your fork of SweetAlert2.
| gharchive/issue | 2015-03-10T19:40:58 | 2025-04-01T06:44:49.104680 | {
"authors": [
"andyharney",
"limonte",
"prateekvarma"
],
"repo": "limonte/sweetalert2",
"url": "https://github.com/limonte/sweetalert2/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2233911671 | Autoreload the page
Feature request
Could you please add page auto-reloading for the dashboard? It would be nice to reload it every minute or so, so we can open the page in the "kiosk" mode and display it somewhere. I believe it may be done with a little JS code.
Before submitting
Please check the following:
[x] I have described the purpose of the suggested change, specifying what I need the enhancement to accomplish, i.e. what problem it solves.
[x] I have included any relevant links, screenshots, environment information, and data relevant to implementing the requested feature, as well as pseudocode for how I want to access the new functionality.
[x] If I have ideas for how the new feature could be implemented, I have provided explanations and/or pseudocode and/or task lists for the steps.
Closed by #43
| gharchive/issue | 2024-04-09T17:04:18 | 2025-04-01T06:44:49.112729 | {
"authors": [
"delucchi-cmu",
"hombit"
],
"repo": "lincc-frameworks/lf-workflow-dash",
"url": "https://github.com/lincc-frameworks/lf-workflow-dash/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
389589080 | Can't compile with NDK r19
* What went wrong:
Execution failed for task ':apng-drawable:externalNativeBuildDebug'.
> Build command failed.
Error while executing process /Users/user/Library/Android/sdk/cmake/3.6.4111459/bin/cmake with arguments {--build /Users/user/Projects/apng-drawable/apng-drawable/.externalNativeBuild/cmake/debug/x86_64 --target apng-drawable}
[1/1] Re-running CMake...
-- Configuring done
-- Generating done
-- Build files have been written to: /Users/user/Projects/apng-drawable/apng-drawable/.externalNativeBuild/cmake/debug/x86_64
[1/24] Building C object CMakeFiles/png.dir/libpng/arm/arm_init.c.o
[2/24] Building C object CMakeFiles/png.dir/libpng/arm/filter_neon_intrinsics.c.o
[3/24] Building C object CMakeFiles/png.dir/libpng/png.c.o
[4/24] Building C object CMakeFiles/png.dir/libpng/pngerror.c.o
[5/24] Building C object CMakeFiles/png.dir/libpng/pngget.c.o
[6/24] Building C object CMakeFiles/png.dir/libpng/pngmem.c.o
[7/24] Building C object CMakeFiles/png.dir/libpng/pngpread.c.o
[8/24] Building C object CMakeFiles/png.dir/libpng/pngread.c.o
[9/24] Building C object CMakeFiles/png.dir/libpng/pngrio.c.o
[10/24] Building C object CMakeFiles/png.dir/libpng/pngrtran.c.o
[11/24] Building C object CMakeFiles/png.dir/libpng/pngrutil.c.o
[12/24] Building C object CMakeFiles/png.dir/libpng/pngset.c.o
[13/24] Building C object CMakeFiles/png.dir/libpng/pngtrans.c.o
[14/24] Building C object CMakeFiles/png.dir/libpng/pngwio.c.o
[15/24] Building C object CMakeFiles/png.dir/libpng/pngwrite.c.o
[16/24] Building C object CMakeFiles/png.dir/libpng/pngwtran.c.o
[17/24] Building C object CMakeFiles/png.dir/libpng/pngwutil.c.o
[18/24] Linking C static library libpng.a
[19/24] Building CXX object CMakeFiles/apng-drawable.dir/apng-drawbale/ApngDecoder.cpp.o
[20/24] Building CXX object CMakeFiles/apng-drawable.dir/apng-drawbale/ApngDecoderJni.cpp.o
[21/24] Building CXX object CMakeFiles/apng-drawable.dir/apng-drawbale/ApngFrame.cpp.o
[22/24] Building CXX object CMakeFiles/apng-drawable.dir/apng-drawbale/ApngImage.cpp.o
[23/24] Building CXX object CMakeFiles/apng-drawable.dir/apng-drawbale/StreamSource.cpp.o
[24/24] Linking CXX shared library /Users/user/Projects/apng-drawable/apng-drawable/build/intermediates/cmake/debug/obj/x86_64/libapng-drawable.so
FAILED: : && /Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/bin/clang++ --target=x86_64-none-linux-android21 --gcc-toolchain=/Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64 -fPIC --sysroot /Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/sysroot -g -DANDROID -fdata-sections -ffunction-sections -funwind-tables -fstack-protector-strong -no-canonical-prefixes -Wa,--noexecstack -Wformat -Werror=format-security -stdlib=libc++ -std=c++17 -fno-rtti -fno-exceptions -DBUILD_DEBUG -O0 -fno-limit-debug-info -Wl,--exclude-libs,libgcc.a -Wl,--exclude-libs,libatomic.a -static-libstdc++ -Wl,--build-id -Wl,--warn-shared-textrel -Wl,--fatal-warnings -Wl,--no-undefined -Qunused-arguments -Wl,-z,noexecstack -Wl,-z,relro -Wl,-z,now -shared -Wl,-soname,libapng-drawable.so -o /Users/user/Projects/apng-drawable/apng-drawable/build/intermediates/cmake/debug/obj/x86_64/libapng-drawable.so CMakeFiles/apng-drawable.dir/apng-drawbale/ApngDecoder.cpp.o CMakeFiles/apng-drawable.dir/apng-drawbale/ApngDecoderJni.cpp.o CMakeFiles/apng-drawable.dir/apng-drawbale/ApngFrame.cpp.o CMakeFiles/apng-drawable.dir/apng-drawbale/ApngImage.cpp.o CMakeFiles/apng-drawable.dir/apng-drawbale/StreamSource.cpp.o -llog -ljnigraphics libpng.a -lz -latomic -lm "/Users/user/Library/Android/sdk/ndk-bundle/sources/cxx-stl/llvm-libc++/libs/x86_64/libc++_static.a" "/Users/user/Library/Android/sdk/ndk-bundle/sources/cxx-stl/llvm-libc++/libs/x86_64/libc++abi.a" && :
/Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/sysroot/usr/include/c++/v1/new:332: error: undefined reference to 'operator delete(void*, std::align_val_t)'
/Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/sysroot/usr/include/c++/v1/new:250: error: undefined reference to 'operator new(unsigned long, std::align_val_t)'
/Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/sysroot/usr/include/c++/v1/new:250: error: undefined reference to 'operator new(unsigned long, std::align_val_t)'
/Users/user/Library/Android/sdk/ndk-bundle/toolchains/llvm/prebuilt/darwin-x86_64/sysroot/usr/include/c++/v1/new:250: error: undefined reference to 'operator new(unsigned long, std::align_val_t)'
clang++: error: linker command failed with exit code 1 (use -v to see invocation)
ninja: build stopped: subcommand failed.
Related? https://github.com/android-ndk/ndk/issues/421
Yep, got the same issue on r19. It is surprising, because it was all ok on r18 with -std=c++17
| gharchive/issue | 2018-12-11T04:09:33 | 2025-04-01T06:44:49.115407 | {
"authors": [
"3damned",
"r-ralph"
],
"repo": "line/apng-drawable",
"url": "https://github.com/line/apng-drawable/issues/19",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
274722530 | [WIP] Add HTTP API version 1
Motivation:
The libthrift is not stable enough yet, we need two diffrent jars to compile.
If we use HTTP API, we can avoid that situation and have fewer depedencies for the client library.
Modifications:
Revise admin REST APIs
Add more APIs to cover all of the features in Thrift service
Result:
We are ready to transfer from Thrift to HTTP API
Todos:
Catch more exceptions and respond it with proper status
Add meaningful reponse body corresponding to the invalid requests
Gather scattered exception handling
Remove admin package or keep it minumun
Remove thrift service ultimately
Please ignore the *Dtos for now. Since the fields for the response is not defined, I just used POJO without much thinking about it.
Codecov Report
Merging #91 into master will increase coverage by 1.63%.
The diff coverage is 74.81%.
@@ Coverage Diff @@
## master #91 +/- ##
==========================================
+ Coverage 49.73% 51.36% +1.63%
==========================================
Files 219 231 +12
Lines 10077 10475 +398
Branches 1352 1388 +36
==========================================
+ Hits 5012 5381 +369
+ Misses 4302 4296 -6
- Partials 763 798 +35
Impacted Files
Coverage Δ
...ma/server/internal/admin/service/TokenService.java
4.54% <ø> (+4.54%)
:arrow_up:
...dogma/server/internal/httpapi/AbstractService.java
100% <ø> (ø)
.../server/internal/admin/service/RepositoryUtil.java
0% <ø> (ø)
:arrow_up:
...p/centraldogma/internal/httpapi/v1/ContentDto.java
0% <0%> (ø)
.../server/internal/admin/service/ProjectService.java
0% <0%> (ø)
:arrow_up:
...gma/server/internal/admin/service/UserService.java
0% <0%> (ø)
:arrow_up:
...ldogma/internal/httpapi/v1/HttpApiV1Constants.java
0% <0%> (ø)
...rver/internal/admin/service/RepositoryService.java
0% <0%> (ø)
:arrow_up:
...va/com/linecorp/centraldogma/internal/Jackson.java
73.77% <100%> (+4.8%)
:arrow_up:
...entraldogma/internal/httpapi/v1/RepositoryDto.java
100% <100%> (ø)
... and 39 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update f6bf9e7...d26fd79. Read the comment docs.
Thank you, @minwoox !
| gharchive/pull-request | 2017-11-17T01:31:52 | 2025-04-01T06:44:49.133578 | {
"authors": [
"codecov-io",
"minwoox",
"trustin"
],
"repo": "line/centraldogma",
"url": "https://github.com/line/centraldogma/pull/91",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
642619089 | Reconsider module structure and struct naming.
Our crate module is getting quite convoluted, maybe we should stop reexporting everything and start adhering to the Rust style guide for struct names (avoid module name as prefix).
This came up in #1039 regarding this line in lib.rs:
pub use alert::{AlertButton, AlertIcon, AlertOptions, AlertResponse, AlertToken};
According to the Rust style guide we should be calling these just Button or Options and then use them like alert::Button or alert::Options.
This will of course mean that the naming will collide with widget::Button if you were to import everything into the same scope, but I don't think typing alert::Button is worse than AlertButton.
This would make even more sense if we would not reexport everything in the crate root, because then you would have to use alert::AlertButton which is pretty redundant.
And splitting up our crate module into multiple modules would make it less overwhelming when looking at it on docs.rs, especially as it is going to grow ever further over time.
I do understand the convenience part of having 'everything' in one module when writing import statements by hand, but by today Rusts tooling is getting better quickly and auto imports in Analyzer are working really well for structs and enums.
Because of this I would propose doing what we did with lens for all modules: avoid c-style prefixes for struct names and maintain a similar module structure to what we have in druid internally.
As @xStrom mentioned rightfully:
If we decide to organize them differently then that would also mean a ton of broken docs links though. So I think even if we want to change that and move things into modules, we should wait until better doc linking is stabilized which looks like it might happen this year.
So for now this issue mostly a reminder and place for discussion.
The ever growing root module namespace seems like an issue indeed. Especially because things there aren't grouped together if they don't all start with the same prefix. Modules would make more sense for organizing.
Import-wise, there are several options:
// When just Button is fine in this scope
use alert::Button;
// When there would be a clash with Button
use alert; // alert::Button::new()
use alert::Button as AlertButton; // AlertButton::new()
I definitely see the value in cleaning up the root namespace. Renaming types is fine as well I guess, as it follows the std Rust style.
Would be nice to hear arguments for/against these changes. Does anyone think we should keep things in the root namespace and/or as AlertButton?
I would be in favor of the reexport-rename style (e.g. use alert::Button as AlertButton; ) , or in aiming to trim down what we reexport in the root, or both.
I would definitely prefer that we work to more closely follow the Rust style guide, unless we can give solid reasoning for not.
As to the docs issue: This is a good point, but I think it's just the cost of refactoring and maintaining docs. If we're likely to get improved linking in the fairly near-term, then excellent. Otherwise, I think we have to bite the bullet whenever someone is ready to make the PR with the changes. Though, as with other contributions, a significant amount of the basic docs work should be part of the PR - one can't just break things and hope somebody else cleans them up.
I think this is a fine line to walk; ideally we could have the commonly used types exported at the root of the crate, and then use modules for certain more specific and focused areas. I think that an alert module makes sense, for instance; so might a file module. What else? Maybe an event module for types that are only part of events (like KeyEvent and MouseButton and TimerToken)? Are there any other groupings that make sense?
| gharchive/issue | 2020-06-21T18:47:09 | 2025-04-01T06:44:49.150957 | {
"authors": [
"Finnerale",
"Zarenor",
"cmyr",
"xStrom"
],
"repo": "linebender/druid",
"url": "https://github.com/linebender/druid/issues/1051",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2353617764 | Differentiate pointer buttons in Button Widget and View
In most cases, you want a button which only actuates when the primary mouse button is pressed, so the easy case is still that. This is a short term hack, because e.g. the active state is still based on any button being pressed, not just those we are interested in.
That is, we probably need to represent a set of buttons we are interested in. However, this change minimally unblocks additional work with Xilem. In particular, see #xilem > Minesweeper converted from Iced to Xilem.
Sounds like this would (in subsequent iterations) benefit from my PointerButtons PR (#334)?
Just asking because if so, that's motivational and something I can look at more once I am well again.
Yeah, I think that's right.
| gharchive/pull-request | 2024-06-14T15:30:19 | 2025-04-01T06:44:49.153700 | {
"authors": [
"DJMcNab",
"waywardmonkeys"
],
"repo": "linebender/xilem",
"url": "https://github.com/linebender/xilem/pull/397",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2495946506 | Replace ApplicationDescription with Blobs
Motivation
Now that we have blobs, there are several follow ups we can do.
Proposal
This PR replaces ApplicationDescription with Blobs. This will clean up a lot of code and also make our blocks smaller when creating applications.
Fixes #2167
Fixes #2362
Test Plan
CI
[!WARNING]
This pull request is not mergeable via GitHub because a downstack PR is open. Once all requirements are satisfied, merge this PR as a stack on Graphite.
Learn more
#2426 👈
#2425
main
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @andresilva91 and the rest of your teammates on Graphite
I can't reproduce this test_wasm_end_to_end_social_user_pub_sub error locally 🤔
I can't reproduce this test_wasm_end_to_end_social_user_pub_sub error locally 🤔
This is fixed
Will split this
| gharchive/pull-request | 2024-08-30T01:34:05 | 2025-04-01T06:44:49.159842 | {
"authors": [
"andresilva91"
],
"repo": "linera-io/linera-protocol",
"url": "https://github.com/linera-io/linera-protocol/pull/2426",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
180421565 | avoid loop in LDAP org hierarchy
Sometimes there may be errors in LDAP, in case a user's manager is his/herself, the code to figure out the org hierarchy will go into infinite loop.
This patch will address such issue.
LGTM
| gharchive/pull-request | 2016-09-30T23:22:46 | 2025-04-01T06:44:49.171348 | {
"authors": [
"alyiwang",
"ericsun2"
],
"repo": "linkedin/WhereHows",
"url": "https://github.com/linkedin/WhereHows/pull/242",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
433932842 | Create ZkServerWrapper to speed up test.
Create ZkServerWrapper to replace ZKServer.
In Mac, NetworkUtil.getLocalHostNames() in ZKServer is time-consuming.
Code of ZkServerWrapper is copied from ZKServer startSingleZkServer and shutdown.
Codecov Report
Merging #1156 into master will increase coverage by 0.02%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #1156 +/- ##
============================================
+ Coverage 69.6% 69.63% +0.02%
+ Complexity 5248 5246 -2
============================================
Files 414 414
Lines 32323 32299 -24
Branches 4112 4113 +1
============================================
- Hits 22500 22491 -9
+ Misses 8704 8686 -18
- Partials 1119 1122 +3
Impacted Files
Coverage Δ
Complexity Δ
...mbry.cloud/azure/AzureCloudDestinationFactory.java
66.66% <0%> (-22.23%)
3% <0%> (ø)
.../main/java/com.github.ambry.router/EncryptJob.java
92.1% <0%> (-5.27%)
4% <0%> (-1%)
...b.ambry.network/BlockingChannelConnectionPool.java
70.17% <0%> (-2%)
9% <0%> (ø)
.../main/java/com.github.ambry.store/ScanResults.java
82.81% <0%> (-1.57%)
17% <0%> (-1%)
.../main/java/com.github.ambry.router/PutManager.java
87.42% <0%> (-0.63%)
29% <0%> (ø)
...in/java/com.github.ambry.store/BlobStoreStats.java
72.16% <0%> (-0.62%)
105% <0%> (ø)
...java/com.github.ambry.network/SSLTransmission.java
69.42% <0%> (-0.32%)
68% <0%> (-1%)
...rc/main/java/com.github.ambry.store/BlobStore.java
88.22% <0%> (-0.12%)
88% <0%> (ø)
...ain/java/com.github.ambry/store/DumpIndexTool.java
0% <0%> (ø)
0% <0%> (ø)
:arrow_down:
....github.ambry/tools/admin/ConcurrencyTestTool.java
0% <0%> (ø)
0% <0%> (ø)
:arrow_down:
... and 20 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 7769142...f7c51ad. Read the comment docs.
| gharchive/pull-request | 2019-04-16T18:50:34 | 2025-04-01T06:44:49.188009 | {
"authors": [
"codecov-io",
"zzmao"
],
"repo": "linkedin/ambry",
"url": "https://github.com/linkedin/ambry/pull/1156",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1322829551 | [Frontend] Cap number of metadata cache entries
This PR adds an ability to restrict the maximum number of entries
in the metadata cache. This allows us to control the amount of memory
consumed by the metadata cache.
Codecov Report
Merging #2167 (1583877) into master (cf70a80) will decrease coverage by 58.39%.
The diff coverage is 0.00%.
:exclamation: Current head 1583877 differs from pull request most recent head ff3e52f. Consider uploading reports for the commit ff3e52f to get more accurate results
@@ Coverage Diff @@
## master #2167 +/- ##
=============================================
- Coverage 73.21% 14.81% -58.40%
+ Complexity 10228 2284 -7944
=============================================
Files 735 735
Lines 55986 55988 +2
Branches 6897 6897
=============================================
- Hits 40990 8296 -32694
- Misses 12811 47209 +34398
+ Partials 2185 483 -1702
Impacted Files
Coverage Δ
...ain/java/com/github/ambry/config/RouterConfig.java
0.00% <0.00%> (-98.37%)
:arrow_down:
...main/java/com/github/ambry/commons/AmbryCache.java
0.00% <0.00%> (-80.00%)
:arrow_down:
.../github/ambry/router/NonBlockingRouterFactory.java
0.00% <ø> (-85.37%)
:arrow_down:
.../src/main/java/com/github/ambry/router/Router.java
0.00% <0.00%> (-100.00%)
:arrow_down:
.../src/main/java/com/github/ambry/store/Message.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...rc/main/java/com/github/ambry/named/PutResult.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...rc/main/java/com/github/ambry/quota/QuotaMode.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...rc/main/java/com/github/ambry/quota/QuotaName.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...rc/main/java/com/github/ambry/rest/RestMethod.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...c/main/java/com/github/ambry/config/KMSConfig.java
0.00% <0.00%> (-100.00%)
:arrow_down:
... and 574 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update cf70a80...ff3e52f. Read the comment docs.
This is my last patch for caching.
| gharchive/pull-request | 2022-07-30T01:25:47 | 2025-04-01T06:44:49.205570 | {
"authors": [
"codecov-commenter",
"snalli"
],
"repo": "linkedin/ambry",
"url": "https://github.com/linkedin/ambry/pull/2167",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
351485086 | unified architecture v0.1
This is pull request for the unified architecture
Review to move over to #430
| gharchive/pull-request | 2018-08-17T07:21:55 | 2025-04-01T06:44:49.242067 | {
"authors": [
"pralabhkumar",
"varunsaxena"
],
"repo": "linkedin/dr-elephant",
"url": "https://github.com/linkedin/dr-elephant/pull/418",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1268372859 | Fixing purview test issues and improve performance
Currently there are two issues:
https://github.com/linkedin/feathr/issues/349 where connecting to Purview always yield this error:
E requests.exceptions.ConnectionError: HTTPSConnectionPool(host='some-purview-name.catalog.purview.azure.com', port=443): Max retries exceeded with url: /api/atlas/v2/types/typedefs (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f90c3e7ea60>: Failed to establish a new connection: [Errno 111] Connection refused'))
Listing entities in purview is extremely slow.
This is caused by the fact that we are using the search_entities API in pyapacheatlas package. This API will try to get ALL the entities and return it back in a huge array, with paging internally implemented. This is:
not useful in most of the Feathr use cases
might be throttled backend,
will be slow (see implementation details here: https://github.com/wjohnson/pyapacheatlas/blob/master/pyapacheatlas/core/discovery/purview.py#L226)
In this case, instead of using search_entities API, we should use the query API which is just a single call to the backend. Note that this API might need the caller to page correctly, but usually we don't need that.
@YihuiGuo this should also fix your earlier issue
Could you please add more tech details and investigations about the bug? I'm not sure what happened actually.
To confirm that the server side throttling is the root cause, you may need try to change server side parameters to relax the restriction and check if it fixes the problem, otherwise the root cause could be somewhere else.
Could you please add more tech details and investigations about the bug? I'm not sure what happened actually. To confirm that the server side throttling is the root cause, you may need try to change server side parameters to relax the restriction and check if it fixes the problem, otherwise the root cause could be somewhere else.
As described in the PR, the implementation of the search_entities API in the underlying pyapacheatlas package is here (https://github.com/wjohnson/pyapacheatlas/blob/master/pyapacheatlas/core/discovery/purview.py#L226) where it's using a while True statement to keep querying the query API, and might got throttled.
More details are available in this issue: https://github.com/wjohnson/pyapacheatlas/issues/206 where I talked with Will offline, and he agrees to add some backoff in the search_entities API.
Also update the PR description to make it a bit more descriptive.
The canonical way to retrieve entity/entities from PurView is to fetch by GUID, also we already stored related GUIDs in our data model.
Querying/Searching is not the correct approach to load a project and may cause unwanted result if this PurView contains data other than what we've created by Feathr, it only fits for-eyes-only scenarios but not cases require accuracy and consistency such as Feathr client.
Please re-think and re-design the whole solution.
The canonical way to retrieve entity/entities from PurView is to fetch by GUID, also we already stored related GUIDs in our data model.
I agree, when CLI save data to purview, it does not specify any hits to partition data by project, that means search via startsWith might still experience perf issues when data volume grows. Since CLI already writes lineage relationship, for example, project contains feature/anchor/derived, using guid list to fetch registered features sounds more efficient and scalable.
The canonical way to retrieve entity/entities from PurView is to fetch by GUID, also we already stored related GUIDs in our data model. Querying/Searching is not the correct approach to load a project and may cause unwanted result if this PurView contains data other than what we've created by Feathr, it only fits for-eyes-only scenarios but not cases require accuracy and consistency such as Feathr client. Please re-think and re-design the whole solution.
Agree, but the goal of this PR is not to solve all those issues. I have a separate PR solving those issues and please take a look: https://github.com/linkedin/feathr/pull/368
| gharchive/pull-request | 2022-06-11T21:06:43 | 2025-04-01T06:44:49.252926 | {
"authors": [
"blrchen",
"windoze",
"xiaoyongzhu"
],
"repo": "linkedin/feathr",
"url": "https://github.com/linkedin/feathr/pull/350",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1111303279 | idna package conflict issue between iris & oncall
Hi All,
when am trying to install iris and oncall application in the same instance I'm getting conflicts with the idna package while starting iris & oncall.
For iris its required idna==2.7 version
For oncall it's required idna==2.10 version.
what is the solution for this?
Hey,
In some scenarios Iris and Oncall use different versions of the same libraries.
This is a result of them being developed independently using different Python virtualenvs.
You can fix this by either:
Instead of installing Iris/Oncall and their Python dependencies system-wide, install Iris/Oncall and their dependencies into their own independent virtualenvs, which will let them use their expected library versions.
This is the deployment strategy we recommend, as it keeps the system-wide Python installation clean and keeps both apps segregated from each other.
If you do want to install Iris/Oncall system-wide, do so by having each in a different container/VM so they don't share libraries.
Regards
Hello @jrgp somewhat related, so I didnt want to open new issue, is there reason for pinning version to 2.10? big change for 3.0 was that python2 support is dropped, and judging by setup.py its not a thing to support python2.x for this project
https://github.com/linkedin/oncall/blob/57e2ce4c7deedd67b5db43f00998be2697363be1/setup.py#L24
https://github.com/kjd/idna/blob/master/HISTORY.rst#30-2021-01-01
Hey @micwoj92
I'm not sure off hand.
@diegocepedaw should be able to provide more context here.
| gharchive/issue | 2022-01-22T06:57:58 | 2025-04-01T06:44:49.262796 | {
"authors": [
"jrgp",
"micwoj92",
"mohammad-sajid"
],
"repo": "linkedin/oncall",
"url": "https://github.com/linkedin/oncall/issues/364",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
215495013 | [TE] Email v2 copy change
Small copy changes per Tushar's feedback.
#1176
Closing this. Email layout was redesigned
| gharchive/pull-request | 2017-03-20T17:24:31 | 2025-04-01T06:44:49.264019 | {
"authors": [
"apucher",
"justYves"
],
"repo": "linkedin/pinot",
"url": "https://github.com/linkedin/pinot/pull/1175",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
340425429 | Add some messages to trace zk behaviors
This PR won't be merged.
It's used for testing zk in Travis.
Codecov Report
Merging #2946 into master will increase coverage by 0.03%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #2946 +/- ##
==========================================
+ Coverage 69.24% 69.27% +0.03%
==========================================
Files 914 914
Lines 43394 43403 +9
Branches 5919 5919
==========================================
+ Hits 30047 30069 +22
+ Misses 11334 11327 -7
+ Partials 2013 2007 -6
Impacted Files
Coverage Δ
...ava/com/linkedin/pinot/common/utils/ZkStarter.java
75% <100%> (+3.35%)
:arrow_up:
...lix/EmptyBrokerOnlineOfflineStateModelFactory.java
86.66% <0%> (-13.34%)
:arrow_down:
...ore/query/scheduler/resources/ResourceManager.java
87.09% <0%> (-9.68%)
:arrow_down:
...pinot/core/operator/docidsets/OrBlockDocIdSet.java
84.9% <0%> (-9.44%)
:arrow_down:
...e/operator/dociditerators/BitmapDocIdIterator.java
60.71% <0%> (-3.58%)
:arrow_down:
...not/broker/broker/helix/ClusterChangeMediator.java
63.33% <0%> (-3.34%)
:arrow_down:
.../pinot/core/query/scheduler/PriorityScheduler.java
78.08% <0%> (-2.74%)
:arrow_down:
...ata/manager/realtime/RealtimeTableDataManager.java
39.55% <0%> (-1.5%)
:arrow_down:
.../controller/helix/core/SegmentDeletionManager.java
76.72% <0%> (-0.87%)
:arrow_down:
...impl/dictionary/DoubleOnHeapMutableDictionary.java
75.55% <0%> (ø)
:arrow_up:
... and 14 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 0be1ec6...340f35e. Read the comment docs.
Since the issue hasn't reproduced recently, I think we should keep the sys.out.printlns to aid debugging when it happens next and merge the change.
| gharchive/pull-request | 2018-07-11T22:18:29 | 2025-04-01T06:44:49.279673 | {
"authors": [
"codecov-io",
"jackjlli",
"sunithabeeram"
],
"repo": "linkedin/pinot",
"url": "https://github.com/linkedin/pinot/pull/2946",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
683979590 | edges: Add new test suite for linkerd edges
This PR adds a new test suite for linkerd edges
Done! @Pothulapati
| gharchive/pull-request | 2020-08-22T11:21:26 | 2025-04-01T06:44:49.282260 | {
"authors": [
"mayankshah1607"
],
"repo": "linkerd/linkerd2-conformance",
"url": "https://github.com/linkerd/linkerd2-conformance/pull/22",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2567669402 | Download failed - Got status code 404 when trying to download https://github.com/linknlink/linknlink-local/archive/refs/tags/1e2336f.zip
Download failed - Got status code 404 when trying to download https://github.com/linknlink/linknlink-local/archive/refs/tags/1e2336f.zip
Traceback (most recent call last):
File "/config/custom_components/hacs/base.py", line 770, in async_download_file
raise HacsException(
custom_components.hacs.exceptions.HacsException: Got status code 404 when trying to download https://github.com/linknlink/linknlink-local/archive/refs/tags/1e2336f.zip
See this in the system logs when Download in HCAS takes forever. I don't think there is a package there and we could ever be successful.
Seems to be local network related. Download was successful later. Most likely related to switching to 2.4GHz only.
| gharchive/issue | 2024-10-05T05:26:01 | 2025-04-01T06:44:49.311140 | {
"authors": [
"dgzzhb"
],
"repo": "linknlink/linknlink-local",
"url": "https://github.com/linknlink/linknlink-local/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2543447079 | fix: [UIE-8007] - DBaaS Summary tab display correct user and readonly host
Description 📝
DBaaS Summary tab display correct user and readonly host
Changes 🔄
List any change relevant to the reviewer.
Display correct username for V2
Display correct read-only host for V2
Target release date 🗓️
9/30/24
Preview 📷
Before
After
How to test 🧪
Prerequisites
Managed Databases Beta account capability
Reproduction steps
Create new DB
Go to summary tab
Verification steps
Username is correct
readonly host displays
As an Author I have considered 🤔
Check all that apply
[x] 👀 Doing a self review
[x] ❔ Our contribution guidelines
[x] 🤏 Splitting feature into small PRs
[x] ➕ Adding a changeset
[] 🧪 Providing/Improving test coverage
[x] 🔐 Removing all sensitive information from the code and PR description
[x] 🚩 Using a feature flag to protect the release
[x] 👣 Providing comprehensive reproduction steps
[ ] 📑 Providing or updating our documentation
[ ] 🕛 Scheduling a pair reviewing session
[ ] 📱 Providing mobile support
[ ] ♿ Providing accessibility support
I'm pointing at the dev environment and have the Managed Databases Beta account capability, but I'm seeing the error below whenever I try to get on the DBaaS landing page -- is this happening for others as well?
Also running into issues with provisioning DBs in alpha at the moment, will try again in a bit
The BE is currently failing for me in alpha too
| gharchive/pull-request | 2024-09-23T19:14:10 | 2025-04-01T06:44:49.423869 | {
"authors": [
"corya-akamai"
],
"repo": "linode/manager",
"url": "https://github.com/linode/manager/pull/10989",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
426209736 | Update volume landing
Description
Regina had a call with a customer that was trying to create a Volume in ATL. This option is filtered out when creating a Volume through /volumes but it is still available from /linodes/$linodeID/volumes. When clicked nothing happens, no error or success message is shown. Console shows:
POST https://cloud.linode.com/api/v4/volumes 400 (BAD REQUEST)
This would present them with an option to open a support ticket to migrate to a data center that has block storage available, but it could easily be changed to just show an error message. I thought about filtering out the Volumes tab entirely when the Linode is in us-southeast but I think it's better that people know the reason it is not an option and it then gives them a next step which is to migrate in order to use block storage.
Type of Change
Non breaking change ('change')
Note to Reviewers
Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration.
Separate from this PR, if errors on volume creation requests aren't being shown to the user, we'll have to fix that.
Definitely. Regina was creating a PDI so it should be tracked.
| gharchive/pull-request | 2019-03-27T22:22:59 | 2025-04-01T06:44:49.427253 | {
"authors": [
"ajdeleon"
],
"repo": "linode/manager",
"url": "https://github.com/linode/manager/pull/4732",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1055677858 | M3-5190: React Query for Object Storage
Description
React Query for Object Storage
[x] Clusters
[x] Buckets (may need to rework to use old per-cluster logic)
[x] Bucket - Create
[x] Bucket - Delete
[x] Bucket Contents (will need to refactor to be a functional component)
[x] Search Buckets
How to test
Test all Object Storage Functionality
Use the MSW to see how /object-storage/buckets handles a cluster being offline
@HanaXu Would you be ok with holding off on those UI changes to keep the scope of this PR focused?
| gharchive/pull-request | 2021-11-17T04:01:15 | 2025-04-01T06:44:49.430475 | {
"authors": [
"bnussman"
],
"repo": "linode/manager",
"url": "https://github.com/linode/manager/pull/8046",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
724074950 | CI for Node
CI for Node
comment
| gharchive/pull-request | 2020-10-18T19:02:52 | 2025-04-01T06:44:49.484040 | {
"authors": [
"lintao-shen"
],
"repo": "lintao-shen/github-actions-for-ci",
"url": "https://github.com/lintao-shen/github-actions-for-ci/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
116423590 | Create New Address form - Alpha field - Case
Requirement: Alphas in the Alpha field (addressNumberSuffix) must be upper case when submitted to the database.
This requirement was not advised prior to development and was not developed.
Implemented - Please test in next version
Alphas entered in lower case in UI submitted to API in upper case.
| gharchive/issue | 2015-11-11T21:19:45 | 2025-04-01T06:44:49.709049 | {
"authors": [
"Deb-Jones",
"SPlanzer"
],
"repo": "linz/QGIS-AIMS-Plugin",
"url": "https://github.com/linz/QGIS-AIMS-Plugin/issues/15",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
859289810 | Spike: what do we need to build to support Elevation Data Maintenance?
Enabler
So that we can create user stories to support maintenance of Elevation data, we want to find out how Elevation Data Maintainers want to import and maintain their data and metadata.
Acceptance Criteria
[ ] The product team understands how Data Maintainers want to umport and maintain their data and metadata
[ ] The product team can work on the user stories to fill the gaps in Data Lake functionality
Additional context
Tasks
[x] Decide list of users to target
[x] Workshop current state with users
[x] Analyse / collate the results of current state
[x] Workshop future state with Topo architects and POs
[ ] Write user stories for future state
Definition of Ready
[x] This story is ready to work on, according to the team's definition
Definition of Done
[ ] This story is done, according to the team's definition
Next steps are to workshop a future state with Topo team (make sure we get info about what is changing with the Topo Processor). Then write user stories.
Current state and future state updated.
Existing user stories already cover the work needed. Mainly
https://github.com/linz/geospatial-data-lake/issues/528
https://github.com/linz/geospatial-data-lake/issues/496
https://confluence.linz.govt.nz/display/LGDL/User+Flows
| gharchive/issue | 2021-04-15T22:22:00 | 2025-04-01T06:44:49.715147 | {
"authors": [
"billgeo"
],
"repo": "linz/geospatial-data-lake",
"url": "https://github.com/linz/geospatial-data-lake/issues/558",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
560958598 | Incorrect scores for cosine similarity
Hi there, I'm trying to use the plugin with Elasticsearch 6.8.1 but getting strange document scores when doing cosine similarity queries.
I have the following query:
{
"query": {
"function_score": {
"boost_mode": "replace",
"script_score": {
"script": {
"source": "binary_vector_score",
"lang": "knn",
"params": {
"cosine": true,
"field": "embedding_vector",
"encoded_vector": "PVC0BjxrQug8VcbbvRqFmrz5KQy9aQ2ku7g0eL1AxVo+KQK6vGHZJz56bne9YwQNvoFb7r21yio8TIP4PZH8b71ksh+8NetEvh/vhL3+gL89pUxfPfdQ8D3j8j47/BHwvio1dL62+ak5LioAvgslND2Hy9O+BvLdvAepBD3P1fW+MjrwvaWz+D1mjWw9vSibvUqxBL09jYM+Nx6FPfTthr4safw9LmFMvT+ZCD5mid0+M9wxPUQ3VD2K8kK+Lrv2PdsAdr6FzIg9pDl3Pd+3/j2C0Q67Vw6gPiMK1b4V43o7RwUAPd8Q9b6ZqnE9wArBPY9d0rx5qUS9xDpQvdPjKj4BZSc9rvZMvb3B373G6eg+JhPOvjg0o72BcnU9/Up6vYwqzr5VTkK+oOwOPUIGXD6mofU9/95Kvkt2/DvSv0i8o6poPKxrKD3gzTA7JUFAvZRJdjnaugC9DVqJPQkg1D4VoQW9JIsQvYnK0D5oMQe9XsByPVlXdj1Hoxk866z4vgtqyjyUp/y+DROkuwS1QD3OKlS+GaOBPPq92D1Xq8i+WVwqPgsmPL2CFY2+AuraPV35uL1VjGC9+bikPKydSD5Rk+K+oYchPoe5ez5bYWE9wBhcPgm1GT1L43A8mKqIPA+NULw6xgC+T0Hova0b1jxpyYy8/RdsvKX5OLycdjA="
}
}
}
}
},
"size": 100
}
Which returns the following hit:
{
"_index": "test_index",
"_type": "_doc",
"_id": "wJUwGnABOtCXXTPakSP3",
"_score": 0.94501674,
"_source": {
...
"embedding_vector": "vRfgbD01DzG9RhYYvVYFdb0z/i69n1VGvEj+Qr3nw3g+PAt3vZZsHz3uAA48eLY0vmkUzb0Lrmy9U2W4PGLMdb4gdSi9lorsvWP/rL4hyg09s1xEPQQPfj2RYjI9FC7avgOK8r6EPzK95e7IvilbiDzxH3S9xA2EPYD4Sj1rkVG+VO7qvZL7Lr0JLYY8wf2yPMpmxDzN44A9+Is6O572hr29T7Q8mEXYPM7/xD6DwzU+SUVZPbrdizznYfG9q4yiPcuGaL5BUDo9PwzWPeRaNz2BUIw85y1sPX/2dL5SsfI82nFgPRG17L5Pi+Y92SyIPXEezr0gcn69L2QpPNajKD5FR3k9FlwLveHtVL2vyNY+E5KOvl6xXztcz4A9qd25vfodoL4y2a6+jZO8PRmFhD7Uxx4+B+lGvZVW4z2MvMC+GtDwPSvnhD2ADPS6/DgAvcmDiTx8j4C982oePbJ9gD4EQj+85ABIveBOKD4nGse9Ik2oPQCXojzcDsw9x9oeviKutbwopsi982++u5f1QD1JZIy97Ul5vBff4D2hK7O+F1woPgWhF7zE2+y9hXfCvbxKDzyM2Wi+AF2gO7XkYD4eYD6+jTnoPmmgqz5GX3477NuIPkzHJDw1ZzA9dUM6vT2p+rzXI4i+AppVvcY047w2xCg9AflKPaailDydrYg="
}
}
The score returned from the plugin is 0.94501674, however if you decode both vectors (using python code provided in the README) and compute the cosine similarity using the below function then the actual (correct) answer is 0.8900337068367593:
import numpy as np
from numpy.linalg import norm
def cosine_sim(vec1, vec2):
"""
Computes the cosine similarity between two vectors
:param vec1:
:param vec2:
:return:
"""
cos_sim = np.dot(vec1, vec2) / \
(norm(vec1) * norm(vec2))
return cos_sim
I've tried the same with Elasticsearch 7.5.0 (and appropriate version of the plugin) and get the same result. I'm using the following Dockerfile to build/install the plugin and run Elasticsearch:
FROM maven:3.5-jdk-8-alpine AS build
COPY fast-elasticsearch-vector-scoring /opt/fast-elasticsearch-vector-scoring
RUN cd /opt/fast-elasticsearch-vector-scoring && mvn package
FROM elasticsearch:6.8.1
COPY --from=build /opt/fast-elasticsearch-vector-scoring/target/releases/elasticsearch-binary-vector-scoring-6.8.1.zip /plugins/elasticsearch-binary-vector-scoring-6.8.1.zip
# Set development mode ENV variables
ENV xpack.security.enabled=false
ENV discovery.type=single-node
# Install the plugin
RUN /usr/share/elasticsearch/bin/elasticsearch-plugin install file:///plugins/elasticsearch-binary-vector-scoring-6.8.1.zip
Any ideas why the score is inconsistent with the calculated value? Any help is greatly appreciated!
This is by design. see why in the readme file:
Note for ElasticSearch 6 and 7 only:
Because scores produced by the script_score function must be non-negative on elasticsearch 7, We convert the dot product score and cosine similarity score by using these simple equations:
(changed dot product) = e^(original dot product)
(changed cosine similarity) = ((original cosine similarity) + 1) / 2
We can use these simple equation to convert them to original score.
(original dot product) = ln(changed dot product)
(original cosine similarity) = (changed cosine similarity) * 2 - 1
Ah, I'm sure the README said this was just Elasticsearch 7 initially so I tried 6.8.1. Thanks for clearing this up :)
| gharchive/issue | 2020-02-06T11:43:20 | 2025-04-01T06:44:49.747789 | {
"authors": [
"lior-k",
"sully90"
],
"repo": "lior-k/fast-elasticsearch-vector-scoring",
"url": "https://github.com/lior-k/fast-elasticsearch-vector-scoring/issues/40",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
304264957 | 之前的节操播放器是没有了是吧
之前就git clone了节操播放器,一直没看到更新,原来是换成了饺子播放器了,这其中是经历了啥啊,换了名字其他的东西应该没换吧
改名字了
| gharchive/issue | 2018-03-12T07:27:33 | 2025-04-01T06:44:49.749006 | {
"authors": [
"feer921",
"lipangit"
],
"repo": "lipangit/JiaoZiVideoPlayer",
"url": "https://github.com/lipangit/JiaoZiVideoPlayer/issues/1679",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
308786412 | 读取视频失败,禁音逻辑导致的疑问
从sd卡读取视频, 可能存在读取失败,在onerror方法中 重新调用了startvideo方法,
@Override
public void onError(int what, int extra) {
super.onError(what, extra);
MyLog.e(TAG, "android.media.MediaPlayer回调的onError" + what + "--extra:" + extra);
startVideo();//重调播放
}
来实现失败的话自动重新加载视频, 在onStatePlaying 做了静音逻辑的处理,
@Override
public void onStatePlaying() {
super.onStatePlaying();
if (current_state != 5) {
goOnPlayOnPause(); //直接进入暂停状态
}
current_state = 5;
JZMediaManager.instance().mediaPlayer.setVolume(0.0f, 0.0f);
}
MyLog.e(TAG, "preparing之后进入播放状态");
MyLog.e(TAG, "模拟暂停了" + current_state + "状态");
}
问题:如果读取正常很顺没事,,如果失败自动再加载,会报setVolume(静音)的异常, 这种错误的原因会是什么导致呢?? 猜测 按说在stateplay中禁音,视频应该正常加载了,去静音很合适啊,却出现了问题; 感谢回答的前辈
解决了吗
暂时没有
查看c++层,,好像mp为空 调用setvolume的时候出现的此异常 @lipangit
android_media_MediaPlayer_setVolume(JNIEnv *env, jobject thiz, jfloat leftVolume, jfloat rightVolume)
{
ALOGV("setVolume: left %f right %f", (float) leftVolume, (float) rightVolume);
sp mp = getMediaPlayer(env, thiz);
if (mp == NULL ) {
jniThrowException(env, "java/lang/IllegalStateException", NULL);
return;
}
process_media_player_call( env, thiz, mp->setVolume((float) leftVolume, (float) rightVolume), NULL, NULL );
}
断点所有meidaplayer.release()的地方,看看有没有蹊跷
@lipangit
补充 bug的出现是我快速切换视频的时候,因为左侧是listview 右侧是视频 快速点击listview的item切换视频,出现这两个bug 静音和暂停
哈哈,,我又来提交bug了,项目紧急那个静音bug还预留着,没解决 又发现了新的,如上复写的onStatePlaying() 里面执行了暂停goOnPlayOnPause(); //直接进入暂停状态 这个暂停方法里面会报非法异常
如下:
java.lang.IllegalStateException
at android.media.MediaPlayer._pause(Native Method)
at android.media.MediaPlayer.pause(MediaPlayer.java:1542)
at cn.jzvd.JZVideoPlayer.goOnPlayOnPause(JZVideoPlayer.java:289)
at cn.jucheng.heart.view.widget.JZVideoPlayerStandard.onStatePlaying(JZVideoPlayerStandard.java:317)
at cn.jzvd.JZVideoPlayer.onInfo(JZVideoPlayer.java:657)
保持关注,有点复杂
项目需求解决,以上所有的问题 在onStatePlaying() 调用暂停,静音 都可能非法异常(随机性),猜测原因是media实例的问题,项目结束有待研究!
项目解决 只要求上来直接禁音,所以在源码里初始化media的时候加上setVolume(0.0f, 0.0f); 就好了,这个当初没想到,笨死,哈哈,耽误不少时间,
经验:在build中依赖三方 无法满足需求 可尝试下载源码依赖
没问题,解决了就好,感谢你的分享
@lipangit
上面提到的设置静音会报IllegalStateException的问题,我也遇到了,我的情景是在播放错误后,去切换静音状态就会报错.查看代码之后发现是在JZVideoPlayer的onError方法中有这样处理:
`
if (isCurrentPlay()) {
JZMediaManager.instance().releaseMediaPlayer();
}
`
而这个时候,isCurrentPlay()判断是true的,所以在此时MediaPlayer被release掉了,这就是上面 @RXlung 提到setvolume报错的原因.希望可以考虑优化这个逻辑.
解决了就好
| gharchive/issue | 2018-03-27T01:06:07 | 2025-04-01T06:44:49.758689 | {
"authors": [
"MorganGeng",
"RXlung",
"lipangit"
],
"repo": "lipangit/JiaoZiVideoPlayer",
"url": "https://github.com/lipangit/JiaoZiVideoPlayer/issues/1735",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
41888479 | Sri Lankan flag issue
Current:
http://lipis.github.io/flag-icon-css/flags/4x3/lk.svg
Expected:
http://upload.wikimedia.org/wikipedia/commons/1/11/Flag_of_Sri_Lanka.svg
I think stretching the background height so the original design fits a 4x3 ratio is a good idea. The flag's original design should be first priority, with some stretching if needed. Can you please update the Sri Lankan flag?
Fixed in #69.. can you please check if it's ok?
@chris-gunawardena, correct me if I'm wrong, but I think stretching the yellow background is better than stretching the elements on the flag.
Looks good to me. The colours represents different ethnicities and this is closest to the original colour ratios.
Thanks for all the input and great work put into fixing this guys!
| gharchive/issue | 2014-09-04T02:22:46 | 2025-04-01T06:44:49.763238 | {
"authors": [
"chris-gunawardena",
"lipis",
"yaniv691"
],
"repo": "lipis/flag-icon-css",
"url": "https://github.com/lipis/flag-icon-css/issues/29",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
635426399 | As a user, I would like to have a command to test my database connection with Liquibase
Description
Currently if there is a need to verify the connection to the database, it can only be done with commands that are associated with a changeLog like "status" and "generateChangeLog".
This command will conveniently allow the user to verify that the connection to the database is successful.
For example:
liquibase test-connection
Console output:
Liquibase Pro 3.9.0 by Datical licensed to Liquibase Pro Customer until Tue Sep 22 19:00:00 CDT 2020
postgres@jdbc:postgresql://localhost:5432/MYDATABASE_DEV
Liquibase command 'test-connection' was executed successfully.
Now we have! https://docs.liquibase.com/commands/change-tracking/connect.html
| gharchive/issue | 2020-06-09T13:23:55 | 2025-04-01T06:44:49.768206 | {
"authors": [
"filipelautert",
"szandany"
],
"repo": "liquibase/liquibase",
"url": "https://github.com/liquibase/liquibase/issues/1202",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2047032729 | Sailee: test-DAT-16470
Impact
[ ] Bug fix (non-breaking change which fixes expected existing functionality)
[ ] Enhancement/New feature (adds functionality without impacting existing logic)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Description
Things to be aware of
Things to worry about
Additional Context
testing done
| gharchive/pull-request | 2023-12-18T16:33:44 | 2025-04-01T06:44:49.771154 | {
"authors": [
"sassasayu",
"sayaliM0412"
],
"repo": "liquibase/liquibase",
"url": "https://github.com/liquibase/liquibase/pull/5401",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
817761274 | Update scala-library to 2.13.5
Updates org.scala-lang:scala-library from 2.13.4 to 2.13.5.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Files still referring to the old version number
The following files still refer to the old version number (2.13.4).
You might want to review and update them manually.
.github/workflows/ci.yml
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scala-lang", artifactId = "scala-library" } ]
labels: library-update, semver-patch, old-version-remains
Superseded by #55.
| gharchive/pull-request | 2021-02-27T00:04:46 | 2025-04-01T06:44:49.774882 | {
"authors": [
"scala-steward"
],
"repo": "liquibase4s/liquibase4s",
"url": "https://github.com/liquibase4s/liquibase4s/pull/20",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
558913308 | 404 Not Found
After I use this command
code-push release-react android android -t "5.0.2" --des "热更新(添加:确定订单备注功能)" -m true -d Production
code-push-server
server display "404 Not Found
"
react-native-code-push version: 5.6.1
react-native version: 0.59.9
iOS/Android/Windows version: all
我也遇到的这个错误,访问登录界面之后,就出现了这个错误,不过不影响登录获取token;
但是我用code-push login命令的时候 会出现错误,不知道是不是这个错误引起的
服务器版本和插件版本不一致
解决方式:
1服务器升级最新的版本5.7.1
2或者降级codepush插件版本,1.11.*
我和你报的一样的错!都是87行,哪位大神解答下! 我是手机端的react-native version 0.60 react-native-code-push 6.0.0
| gharchive/issue | 2020-02-03T08:11:40 | 2025-04-01T06:44:49.801056 | {
"authors": [
"164580084",
"gaopeng1234567",
"shanzemin",
"wxp20032003"
],
"repo": "lisong/code-push-server",
"url": "https://github.com/lisong/code-push-server/issues/262",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
256411986 | QQ音乐无法播放
资源的地址改了
+1
@listen1 QQ音乐无法播放了,求更新。
桌面版 QQ音乐无法播放 +1
手动将 js/provider/qq.js 中的 qq_bootstrap_track 函数中的 http://cc.stream.qqmusic.qq.com 改为 http://dl.stream.qqmusic.qq.com 就可以了
chrome插件版已经更新,谢谢大家的反馈,接下来会重新打包桌面版
| gharchive/issue | 2017-09-09T03:47:08 | 2025-04-01T06:44:49.803293 | {
"authors": [
"Blade6",
"Chigogo",
"huangzhongxin",
"listen1",
"purplest"
],
"repo": "listen1/listen1_chrome_extension",
"url": "https://github.com/listen1/listen1_chrome_extension/issues/90",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2244651515 | [Bug]: Task returning object never completes; returning boolean it does
Replication of the Problem
No response
Problem
I was porting some code from listr to listr2 (thanks very much for this updated package!).
I was using a task like this with ssh2-sftp-client:
const tasks = new Listr([
{
title: `Connecting to ${passportUploadHost}`,
task: () =>
sftp.connect({
host: passportUploadHost,
username: passportUploadUser,
password: passportUploadPassword,
privateKey: sshKey,
})
},
… // more tasks
])
This is correctly-typed, but although the connect method completes, the task never does. The connect method returns Promise<SFTPWrapper>, meaning the return value of the task function is Promise<SFTPWrapper>, which seems fine (it's a Promise).
I eventually found a way to make it work:
const tasks = new Listr([
{
title: `Connecting to ${passportUploadHost}`,
task: async () => {
await sftp.connect({
host: passportUploadHost,
username: passportUploadUser,
password: passportUploadPassword,
privateKey: sshKey,
})
return true
},
},
… // more tasks
])
Here, I await the connection object, then return true. I cannot see from the documentation why this should be necessary. Apologies if I have overlooked something.
Environment
Operating system: GNU/Linux
Node.js version: 20.11.0
Terminal: VS Code
listr2 version: 8.2.1
Log Output
No response
Preflight
[X] Read/Search existing issues and the provided documentation.
[X] Be sure to not open an issue in the looking for contributions section.
[X] Checking the latest version of the software to confirm it has not been fixed already.
Dear @rrthomas,
Interesting issue, that I have never experienced yet.
The only reason I can think of is due to this part of recursively https://github.com/listr2/listr2/blob/b29f80c2dcde3d12a0385a9895e7e3033fc53e61/packages/listr2/src/lib/task.ts#L283 handling the return value.
If the .connect method always returns a promise where the SFTPWrapper is again a promise and so on it will recursively await it.
The behavior (i think) should be really similar to the original listr. Maybe it does not handle the promises recursively.
I would investigate this at a convient time possibly in a couple of days.
| gharchive/issue | 2024-04-15T21:24:35 | 2025-04-01T06:44:49.810528 | {
"authors": [
"cenk1cenk2",
"rrthomas"
],
"repo": "listr2/listr2",
"url": "https://github.com/listr2/listr2/issues/714",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
126967970 | Fetch room data in message source
Rather than just initializing a room object, this will ensure the room belonging to a message Source is fetching from the database. This provides access to room metadata that you would otherwise have to look up after the fact.
See discussion https://github.com/litaio/lita/commit/f507bb0b02940b35b1d9ce3218714a0a5046c460#commitcomment-15466157
cc @jimmycuadra
I would have liked to have implemented this without having to add another interface, but I didn't feel confident enough to change an existing interface. I'm happy to change this as needed to make it acceptable :)
:eyes: @chaeokay
Thanks for the PR! Let's just inline the logic instead of adding a new method to Lita::Room for now. If that operation proves necessary to repeat in more places later, then adding it as a method may be a good refactoring.
That makes sense :). I've updated the PR. Check out my comment on the second commit. I'm not a huge fan of the test, but it feels desirable to cover that case...
Thanks a lot! Nice improvement.
| gharchive/pull-request | 2016-01-15T22:07:28 | 2025-04-01T06:44:49.816319 | {
"authors": [
"iamvery",
"jimmycuadra",
"wolfbrain"
],
"repo": "litaio/lita",
"url": "https://github.com/litaio/lita/pull/167",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1789164957 | Add metadata to offer/bid/listing creation
Description
Add metadata to offer/bid/listing creation
Update @nft/api-graphql to latest
Checklist
[ ] Update related changelogs
[ ] Update docs if necessary
@antho1404 ready for review
@antho1404 updates dome
| gharchive/pull-request | 2023-07-05T09:34:36 | 2025-04-01T06:44:49.829059 | {
"authors": [
"NicolasMahe"
],
"repo": "liteflow-labs/liteflow-js",
"url": "https://github.com/liteflow-labs/liteflow-js/pull/165",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
178453511 | How to remove selected shape / drawing. Possibly using delete button.
Suppose I have 5 shapes drawn one after other. I imported few images and now I want to remove one of the shapes keeping rest as it is. I can use a button to perform this or delete button on the keyboard.
How do I achieve this.?
we have access to shape array. If we can get id of selected shape. we can remove that shape from array and trigger 'drawingChange' to redraw the canvas. I'm sure how to access the current selected shape.
There isn't actually a straightforward way to do this. You can modify
lc.shapes and then call lc.repaintLayer('main'). Note that this will
not affect the undo stack, so you might run into weird problems unless
you're just working with background shapes.
If you want to do better, you'll have to familiarize yourself with the
library's source code and implement a new action in actions.coffee,
and then write a lc.removeShape() method that mirrors
lc.saveShape().
Looks like its too complex for me. I'm not a java-script or coffee-script developer. I'm learning it as needed. After spending few days with the source, now I'm getting some sense about what's going on. Can you please provide more guidance?
Please correct my steps
Add new action class 'RemoveShapeAction' in actions.coffee and exported.
Add new method 'removeShape' in LiterallyCanvas.coffee
Get access to currently selected Shape object -> How?
Remove shape from Shapes array
Remove shape from undo stack (if not empty)
Remove shape from redo stack (if not empty)
Redraw canvas using lc.repaintLayer('main')
In later stages, I'll also need to modify the selected shape like move / rotate / flip and also crop for images.
Please let me know how feasible do you think it is?
I'm sorry, it isn't practical for me to provide that level of support to users. I am occasionally available for contract work but other than that I don't have time.
This is how I remove shapes from a canvas...
canvas is your LC instance returned by LC.init(targetElement);
deleteButton is a DOM element the user the clicks once a shape is selected using LC's built-in select tool
const deleteButton = document.getElementById('button-delete');
deleteButton.addEventListener('click', function() {
const selectedShape = canvas.tool.selectedShape;
/* Remove shape from shapes list*/
if (selectedShape) {
const selectedShapeIndex = canvas.shapes.indexOf(selectedShape);
canvas.shapes.splice(selectedShapeIndex, 1);
canvas.setShapesInProgress([]);
/*Redraw the canvas with the shape now removed */
canvas.repaintLayer('main');
}
});
@sir-dunxalot - Thanks a lot, it works like a charm.
@sir-dunxalot Thank you so much!
A little enhancement to @sir-dunxalot code snippet which append inline undo/redo support (work for me with LC v0.4.14 ) for the delete shape operation (certainly optimizable) :
const deleteButton = document.getElementById('button-delete');
deleteButton.addEventListener('click', function() {
const selectedShape = canvas.tool.selectedShape;
/* Remove shape from shapes list*/
if (selectedShape) {
const selectedShapeIndex = canvas.shapes.indexOf(selectedShape);
canvas.shapes.splice(selectedShapeIndex, 1);
canvas.setShapesInProgress([]); /* Also removes selection box */
/* Mimic actions of select tool:
https://github.com/literallycanvas/literallycanvas/blob/884fc422604d7cf6e4159fb9415e735ac19bfba3/src/tools/SelectShape.coffee#L59
*/
canvas.trigger('shapeMoved', { shape: selectedShape })
canvas.trigger('drawingChange', {})
/* Clear the selected shape (prevents second click on delete button deleting the shape now at the index of the old selected shape) */
//--------------------------------- Undo/ReDo Support here ------------------------------------------------------
var ss = canvas.tool.selectedShape;
var sid = canvas.tool.selectedShape.id;
canvas.tool.selectedShape = null;
/* Redraw the canvas with the shape now removed */
canvas.repaintLayer('main');
//Custom shapeDeleted event
canvas.trigger('shapeDeleted', ss);
//Add to undo/redo stack
canvas.execute({'do':function(){
//Del shape
for(var i=0;i!=canvas.shapes.length;i++){
if(canvas.shapes[i].id == sid){
canvas.shapes.splice(i, 1);
break;
}
}
canvas.repaintLayer('main');
},undo:function(){
//ReAdd shape
canvas.shapes.push(ss);
canvas.repaintLayer('main');
}});
//-------------------------------------------------END------------------------------------------------------
/* Instantiate a new instance of the select tool so the user can select another shape immediately */
canvas.setTool(new LC.tools.SelectShape(canvas));
}
});
Hi @TheWalkingCat and thanks for the example.
I noticed that when one calls canvas.execute, one does not need to call canvas.shapes.splice(selectedShapeIndex, 1);, as canvas.execute also executes the do function.
Hey Guys,
I know this is not the correct thread to post my question here, but I am really stuck at a point and would want some guidance to move forward.
I am working on a POC right now where in, user wants to upload an image from local directory and draw shapes on this image and then save the updated image to a local directory.
So basically they want to upload an image and drag, drop, resize and delete circles, squares, arrows, arrows with text and text area on top of the image to highlight certain areas in the image and at the end save the updated image back to local directory.
I started off with using canvas for this, but read few articles and figured out that svg+d3 would be best suited for this kind of problem.
What do you suggest?
| gharchive/issue | 2016-09-21T20:25:20 | 2025-04-01T06:44:49.851393 | {
"authors": [
"AbhijitParate",
"TheWalkingCat",
"anx-ckreuzberger",
"irskep",
"milisharma1",
"mraguraman3",
"sir-dunxalot"
],
"repo": "literallycanvas/literallycanvas",
"url": "https://github.com/literallycanvas/literallycanvas/issues/432",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
258774437 | Create Set and MPR function hangs UI
This should be moved to its own thread, with some indication of work in progress
Fixed in https://github.com/lithnet/miis-autosync/releases/tag/v1.1.6500
| gharchive/issue | 2017-09-19T10:38:25 | 2025-04-01T06:44:49.867722 | {
"authors": [
"ryannewington"
],
"repo": "lithnet/miis-autosync",
"url": "https://github.com/lithnet/miis-autosync/issues/97",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
956839817 | Question about Code Engine
What is the minimum docker image I need to run your applications inside CE? Now I see that you are based on python:3.8-slim-buster and then put a couple of system dependencies and a rather large list of python dependency? Are they all really needed? I am interested about packages for numeric and Compose Libs and also kafka, redis, twisted ...
Could you publish new image for ibmfunctions/lithops-ce-v385 with support for version 2.4.0 on Docker Hub?
@sergii-mamedov You can take this Dockerfile and add the python packages you need in the pip install (this dockerfile contains the minimum list of python libs required by lithops):
# Python 3.6
#FROM python:3.6-slim-buster
# Python 3.7
#FROM python:3.7-slim-buster
# Python 3.8
FROM python:3.8-slim-buster
# Python 3.9
#FROM python:3.9-slim-buster
RUN apt-get update \
# add some packages required for the pip install
&& apt-get install -y \
gcc \
zlib1g-dev \
libxslt-dev \
libxml2-dev \
zip \
unzip \
make \
# cleanup package lists, they are not used anymore in this image
&& rm -rf /var/lib/apt/lists/* \
&& apt-cache search linux-headers-generic
RUN pip install --upgrade pip setuptools six gevent \
&& pip install --no-cache-dir \
simplejson \
httplib2 \
kafka_python \
lxml \
python-dateutil \
pika \
flask \
gevent \
ibm-cos-sdk \
redis \
requests \
pandas \
numpy \
cloudpickle \
ps-mem \
tblib
ENV PORT 8080
ENV CONCURRENCY 4
ENV TIMEOUT 600
ENV PYTHONUNBUFFERED TRUE
# Copy Lithops proxy and lib to the container image.
ENV APP_HOME /lithops
WORKDIR $APP_HOME
COPY lithops_codeengine.zip .
RUN unzip lithops_codeengine.zip && rm lithops_codeengine.zip
CMD exec gunicorn --bind :$PORT --workers $CONCURRENCY --timeout $TIMEOUT lithopsentry:proxy
@gilv Could you create the ibmfunctions/lithops-ce-v385:240 runtime?
@JosepSampe yeap
@sergii-mamedov @JosepSampe I released new Lithops version
@gilv
I do not see new release here
@JosepSampe
I have seen the given code. My question was more about whether this minimal image really needs numpy, pandas ...?
| gharchive/issue | 2021-07-30T15:16:25 | 2025-04-01T06:44:49.873229 | {
"authors": [
"JosepSampe",
"gilv",
"sergii-mamedov"
],
"repo": "lithops-cloud/lithops",
"url": "https://github.com/lithops-cloud/lithops/issues/734",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
74337641 | Multiple requests failing
Hi guys,
I'm writting a library (myql) which use rauth , but everytime i run more than one request, i get this error even though i used header_auth=True
[2015-05-08 13:45:49,823 DEBUG] [root.yoauth.token_is_valid] ELAPSED TIME : 2245.65489984
[2015-05-08 13:45:49,824 DEBUG] [root.yoauth.token_is_valid] TOKEN IS STILL VALID
[2015-05-08 13:45:49,827 DEBUG] [root.yoauth.token_is_valid] ELAPSED TIME : 2245.65891886
[2015-05-08 13:45:49,828 DEBUG] [root.yoauth.token_is_valid] TOKEN IS STILL VALID
[2015-05-08 13:45:49,833 INFO] [requests.packages.urllib3.connectionpool.connectionpool._new_conn] Starting new HTTP connection (1): query.yahooapis.com
[2015-05-08 13:45:50,196 DEBUG] [requests.packages.urllib3.connectionpool.connectionpool._make_request] "GET /v1/yql?format=json&q=select+%2A+from+fantasysports.teams.roster++where+team_key%3D%27mlb.l.1328.t.1%27+and+date%3D%272015-05-05%27&callback=&jsonCompact=False&diagnostics=False&debug=False HTTP/1.1" 200 None
1 Y! - Funston 0 11
[2015-05-08 13:45:50,287 DEBUG] [root.yoauth.token_is_valid] ELAPSED TIME : 2246.11847591
[2015-05-08 13:45:50,287 DEBUG] [root.yoauth.token_is_valid] TOKEN IS STILL VALID
Traceback (most recent call last):
File "test.py", line 25, in <module>
response = yql.select('fantasysports.teams.roster').where(['team_key','=',team],['date','=',year])
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/myql/myql.py", line 216, in where
response = self.executeQuery(payload)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/myql/myql.py", line 89, in executeQuery
response = self.oauth.session.get(self.url, params=payload, header_auth=True)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/requests/sessions.py", line 477, in get
return self.request('GET', url, **kwargs)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/rauth/session.py", line 210, in request
return super(OAuth1Session, self).request(method, url, **req_kwargs)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/requests/sessions.py", line 465, in request
resp = self.send(prep, **send_kwargs)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/requests/sessions.py", line 573, in send
r = adapter.send(request, **kwargs)
File "/home/yosuke/.virtualenvs/test/local/lib/python2.7/site-packages/requests/adapters.py", line 415, in send
raise ConnectionError(err, request=request)
requests.exceptions.ConnectionError: ('Connection aborted.', ResponseNotReady())
This is my code
import sys
import myql
from myql.contrib.auth import YOAuth
if '__main__' == __name__:
oauth = YOAuth(None, None, from_file='credentials.json')
yql = myql.MYQL(format='json',oauth=oauth)
teams = ('mlb.l.1328.t.1','mlb.l.1328.t.2')
year = '2015-05-05'
for team in teams :
response = yql.select('fantasysports.teams.roster').where(['team_key','=',team],['date','=',year])
if not response.status_code == 200:
sys.exit(1)
data = response.json()
current_team = data['query']['results']['team']
print current_team['team_id'],current_team['name'],current_team['number_of_trades'],current_team['number_of_moves']
I know this one has already been fixed, i can't figure out what's going wrong.
What version of rauth are you using?
Hello @maxcountryman here they are
rauth==0.7.1
requests==2.6.2
Thanks for the tip @maxcountryman. I just updated the version of requests
rauth==0.7.1
requests==2.7.0
And everything is working. I need to run more tests though before closing the issue if you don't mind.
Thanks
Thanks @maxcountryman the issue can be closed
| gharchive/issue | 2015-05-08T11:50:53 | 2025-04-01T06:44:49.878504 | {
"authors": [
"josuebrunel",
"maxcountryman"
],
"repo": "litl/rauth",
"url": "https://github.com/litl/rauth/issues/180",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
906612413 | Evs二进制代码文件结构设计 草案
文件头:Evs byte[3] char
版本:1.0 byte[4] float
备用文件信息 表示这个二进制文件处于avg游戏的第几个章节第几段的意思
章节:int
段落:int
块大小:20 int (根据数据自己生成)
每块数据:byte[30] char 人物名称15个汉字 utf-8 不足30字节用00补足 //对应结构[人物],[句子],[代码]
句子字节总数:int (自动生成)
句子: byte[] char utf-8 长度是上一个int指定的
代码字节总数:int 自动生成
js代码:char utf-8
用于编译源码Evs到二进制数据结构文件 data file 而不是excute file 没有语法解析器 token解析等 复杂编译原理 。
这是一种自创的数据文件,用于减少源码占用的空间
| gharchive/issue | 2021-05-30T03:00:36 | 2025-04-01T06:44:49.881105 | {
"authors": [
"2439905184"
],
"repo": "little-game-and-simple-software/EasyAvgFrameWork",
"url": "https://github.com/little-game-and-simple-software/EasyAvgFrameWork/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
244631314 | 同一台机器,有的微信可以上传文件,有的不行,会报10061错误
版本是最新版本
mediaid = self._itchat_bot.upload_file(local_pic_url, isPicture=True)
执行上面语句 ,偶尔会报 错,错误如下,错误概率 60% ,但是其他的微信号,就是 100% 正确,不会报错
HTTPSConnectionPool(host='file.wx8.qq.com', port=443): Max retries exceeded with url: /cgi-bin/mmwebwx-bin/webwxuploadmedia?f=json (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x0000025AA982F5C0>: Failed to establish a new connection: [Errno 10061] [WinError 10061] 由于目标计算机积极拒绝,无法连接。',))
原因找到了,WX8 ,网络不稳定,号的问题
| gharchive/issue | 2017-07-21T10:36:25 | 2025-04-01T06:44:49.883342 | {
"authors": [
"teli5200"
],
"repo": "littlecodersh/ItChat",
"url": "https://github.com/littlecodersh/ItChat/issues/453",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1479651863 | LFS_ERR_NOSPC returned by lfs_dir_compact() when there is still room on the disk
Hello all,
I am running into an issue where littlefs fails to close a file once a certain number of files are present in the root directory. Specifically, the check
if (commit->off + dsize > commit->end) { return LFS_ERR_NOSPC; }
in lfs_dir_commitattr will return LFS_ERR_NOSPC with the following call-stack:
lfs_file_close ->
lfs_file_rawclose ->
lfs_file_rawsync ->
lfs_dir_commit ->
lfs_dir_orphaningcommit ->
lfs_dir_relocatingcommit ->
lfs_dir_splittingcompact ->
lfs_dir_compact ->
lfs_dir_traverse ->
lfs_dir_commit_commit ->
lfs_dir_commitattr
The littlefs configuration is typically as follows:
read_size = 4096
prog_size = 4096
block_size = 262144
block_count = 2032
block_cycles = 500
cache_size = 4096
lookahead_size = 64
name_max = 255
file_max = default
attr_max = 1022
metadata_max = 4096
When this error occurs, littlefs is using less than 1% of the total space allocated for it (508 MB). The number of files required to cause this issue changes with the configured metadata_max. If metadata_max is 4 kB, the issue happens when closing the 124th file. If metadata_max is 512 B, the issue happens when closing the 15th file. It is my understanding that littlefs should move metadata around so that many files can be stored in a directory, and that metadata_max should not have an effect on the total number of files which can be placed in a directory. Are these assumptions correct?
I've taken a look at https://github.com/littlefs-project/littlefs/issues/478, and the files only have a small custom attribute containing an application level checksum. Additionally, the read_size and prog_size are not the same size as block_size, which was another suggested fix in the linked ticket.
I'm having this exact same issue with basically the same configuration. One of the things that seems to be happening is that metadata_max is less than the block size, which seems to cause it to artificially run out of space.
The backend here is directly to an unmanaged NAND, so changing the block size is not possible. Also in this case I see the performance of lfs_file_opencfg() gradually degrades as the number of files. Increasing the size of metadata_max will increase the # of files I can open, but also severely degrades lfs_file_opencfg() performance.
I know @geky has previously mentioned using a B-tree instead of a CTZ list; I think this is possibly the only thing that will improve performance in this situation.
I'm having the same issue.
Hi all, sorry for the late response.
I've been digging around to try to figure out these LFS_ERR_NOSPC errors and found one bug that may be related: https://github.com/littlefs-project/littlefs/pull/1031
It would be interesting to know if this fixed @ajheck's issue, since it looks like you're using the same prog_size and metadata_max.
I'm not aware of an issue when metadata_max > prog_size, though it's certainly possible and would be good to know about.
Just some other -- late -- comments:
It is my understanding that littlefs should move metadata around so that many files can be stored in a directory, and that metadata_max should not have an effect on the total number of files which can be placed in a directory. Are these assumptions correct?
Yes, this is how it's supposed to work normally. Directories are linked-lists of metadata blocks, and metadata_max just limits how much metadata is stored in each block. If you have many files in a directory -- and no fs bugs -- these can end up in multiple metadata blocks.
The one catch is that a single file's metadata can't span multiple metadata blocks. So if you have, say, a large filename or many custom attributes, it can cause problems.
I know @geky has previously mentioned using a B-tree instead of a CTZ list; I think this is possibly the only thing that will improve performance in this situation.
I don't think this specific feature will help with lfs_file_opencfg speed, since directories in littlefs are not stored in files, but I've been looking at a few other ideas such as red-black trees in logs.
| gharchive/issue | 2022-12-06T17:09:04 | 2025-04-01T06:44:49.905296 | {
"authors": [
"ajheck",
"dpkristensen",
"geky",
"pedromarinho94"
],
"repo": "littlefs-project/littlefs",
"url": "https://github.com/littlefs-project/littlefs/issues/755",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1142865277 | 🛑 USTCLUG is down
In 8346e88, USTCLUG (https://lug.ustc.edu.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: USTCLUG is back up in 631628e.
| gharchive/issue | 2022-02-18T11:13:31 | 2025-04-01T06:44:49.908095 | {
"authors": [
"littlekud"
],
"repo": "littlekud/sites-status",
"url": "https://github.com/littlekud/sites-status/issues/507",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
418010100 | littlevgl site down
https://littlevgl.com is down :(
Maybe the host did some maintenance. Seems working now.
Anyway, thank you for letting us know!
| gharchive/issue | 2019-03-06T20:33:02 | 2025-04-01T06:44:49.909422 | {
"authors": [
"codehackr",
"kisvegabor"
],
"repo": "littlevgl/lvgl",
"url": "https://github.com/littlevgl/lvgl/issues/931",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
688792734 | vim-clap: on_event:['Error: No such file or directory (os error 2)', '']
Environment:
OS: macOS
(Neo)Vim version: 0.4.4
vim-clap version: e690bde9e7838894b68f9d3d4d1a131e86c13ffb
Have you reproduced with a minimal vimrc: yes
Have you updated to the latest plugin version: yes
Have you upgraded to/compiled the latest Rust binary: ???
Describe the bug
:Clap files and other providers stopped working when current buffer is a :term between v0.19 and v0.20.
Clap debug
~ NVIM v0.4.4
has cargo: 1
has maple: /Users/stefan/.dotfiles/_vim/plugged/vim-clap/target/release/maple
maple info: version 0.1.20, built for x86_64-apple-darwin by rustc 1.45.2.
has +python3: 1
has py dynamic module: 1
Current FileType:
Third Party Providers: []
Global Options:
let g:clap#autoload_dir = '/Users/stefan/.dotfiles/_vim/plugged/vim-clap/autoload'
let g:clap#provider_alias = {'gfiles': 'git_files', 'hist:': 'command_history', 'hist/': 'search_history'}
let g:clap_current_selection_sign = {'texthl': 'WarningMsg', 'linehl': 'ClapCurrentSelection', 'text': ' ➤'}
let g:clap_disable_bottom_top = 0
let g:clap_disable_matches_indicator = v:false
let g:clap_disable_optional_async = v:false
let g:clap_disable_run_rooter = v:false
let g:clap_enable_icon = 0
let g:clap_forerunner_status_sign = {'done': '•', 'running': '!', 'using_cache': '*'}
let g:clap_insert_mode_only = v:false
let g:clap_layout = {'relative': 'editor'}
let g:clap_multi_selection_warning_silent = 0
let g:clap_no_matches_msg = 'NO MATCHES FOUND'
let g:clap_open_action = {'ctrl-v': 'vsplit', 'ctrl-x': 'split', 'ctrl-t': 'tab split'}
let g:clap_popup_border = 'rounded'
let g:clap_preview_size = 5
let g:clap_providers_relaunch_code = '@@'
let g:clap_search_box_border_style = 'nil'
let g:clap_search_box_border_symbols = {'nil': ['', ''], 'curve': ['', ''], 'arrow': ['', '']}
let g:clap_selected_sign = {'texthl': 'WarningMsg', 'linehl': 'ClapSelected', 'text': ' ✔︎'}
Provider Variables: []
thx for the quick fix :)
| gharchive/issue | 2020-08-30T21:15:09 | 2025-04-01T06:44:49.914375 | {
"authors": [
"sscherfke"
],
"repo": "liuchengxu/vim-clap",
"url": "https://github.com/liuchengxu/vim-clap/issues/513",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
811212183 | Preview on filer is limited to 10 lines
Environment (please complete the following information):
OS: MacOS
Vim version: 8.2.2500
vim-clap version: master
Have you reproduced with a minimal vimrc: No
Have you updated to the latest plugin version: Yes
Have you upgraded to/compiled the latest Rust binary: Yes
Describe the bug
It seems like the preview pane in filer only shows the first 10 lines of the file when hovering on them.
Clap debug
has cargo: 1
has maple: /Users/amousavi/.vim/plugged/vim-clap/target/release/maple
maple info: version 0.1.23 (git v0.23), built for x86_64-apple-darwin by rustc 1.50.0 (cb75ad5db 2021-02-10).
has +python3: 1
has py dynamic module: 1
has ctags: ctags with JSON output support
Current sync impl: Lua
Current FileType:
Third Party Providers: []
Global Options:
let g:clap#autoload_dir = '/Users/amousavi/.vim/plugged/vim-clap/autoload'
let g:clap#popup#display = {'shrink': function('164'), 'open': function('<SNR>112_create_display'), 'width': 81, 'shrink_if_undersize': function('163')}
let g:clap#popup#preview = {'line_count': function('<SNR>111__line_count'), 'show': function('165'), 'hide': function('166'), 'clear': function('167'), 'add_highlight': function('168'), 'get_lines': f
unction('<SNR>111__get_lines'), 'getbufvar': function('<SNR>111__getbufvar'), 'setbufvar_batch': function('<SNR>111__setbufvar_batch'), 'setbufvar': function('<SNR>111__setbufvar'), 'winid': 1018, 'win_is
_valid': function('<SNR>111__win_is_valid'), 'goto_win': function('<SNR>111__goto_win'), 'set_syntax': function('169'), 'bufnr': 6}
let g:clap#provider_alias = {'gfiles': 'git_files', 'hist:': 'command_history', 'hist/': 'search_history'}
let g:clap_background_shadow_blend = 50
let g:clap_disable_bottom_top = 0
let g:clap_disable_matches_indicator = v:false
let g:clap_disable_optional_async = v:false
let g:clap_disable_run_rooter = v:true
let g:clap_enable_background_shadow = v:true
let g:clap_enable_debug = v:false
let g:clap_enable_icon = 1
let g:clap_forerunner_status_sign = {'done': '•', 'running': '!', 'using_cache': '*'}
let g:clap_indicator_winid = 1019
let g:clap_insert_mode_only = v:false
let g:clap_layout = {'relative': 'editor'}
let g:clap_multi_selection_warning_silent = 0
let g:clap_no_matches_msg = 'NO MATCHES FOUND'
let g:clap_open_action = {'ctrl-v': 'vsplit', 'ctrl-x': 'split', 'ctrl-t': 'tab split'}
let g:clap_open_preview = 'always'
let g:clap_popup_border = 'rounded'
let g:clap_preview_direction = 'LR'
let g:clap_preview_size = 5
let g:clap_providers_relaunch_code = '@@'
let g:clap_search_box_border_style = 'nil'
let g:clap_search_box_border_symbols = {'nil': ['', ''], 'curve': ['', ''], 'arrow': ['', '']}
let g:clap_spinner_winid = 1021
let g:clap_theme = 'dogrun'
Provider Variables:
**[]**
Increase the preview size to fit your screen size, e.g., let g:clap_preview_size = 10 will show 20 lines in the preview window.
In https://github.com/liuchengxu/vim-clap/commit/b6d82bc748a80577a31f2a40eb36947d70197a67 the default behaviour has been changed to fulfil the whole preview window, please pull the latest commit and compile the Rust binary.
Works great! thanks 👍
I just realized I have the same issue with buffers. Could be cool to have defaulted to window window height as well.
Should be resolved in https://github.com/liuchengxu/vim-clap/commit/9bb839bd9c5a5130a1794583741cbcf6406eb074 @arashm
Works perfectly 👍 Thanks
| gharchive/issue | 2021-02-18T15:37:11 | 2025-04-01T06:44:49.920627 | {
"authors": [
"arashm",
"liuchengxu"
],
"repo": "liuchengxu/vim-clap",
"url": "https://github.com/liuchengxu/vim-clap/issues/641",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1321902130 | added gr20.1 to main after meeting
Not sure if it's ok here.
merged gr20
| gharchive/pull-request | 2022-07-29T08:05:35 | 2025-04-01T06:44:49.924785 | {
"authors": [
"liuquan18"
],
"repo": "liuquan18/3rdPanel",
"url": "https://github.com/liuquan18/3rdPanel/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1320536947 | 🛑 shop.worldofwarships.asia is down
In 182dbdd, shop.worldofwarships.asia (https://shop.worldofwarships.asia/) was down:
HTTP code: 500
Response time: 5826 ms
Resolved: shop.worldofwarships.asia is back up in 9975cf6.
| gharchive/issue | 2022-07-28T07:50:33 | 2025-04-01T06:44:49.939193 | {
"authors": [
"nonamenix"
],
"repo": "live4dev/uptime.live4.dev",
"url": "https://github.com/live4dev/uptime.live4.dev/issues/126",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1512623082 | Update to Angular v15
Type of change
[ ] Docs
[ ] Bug fix
[x] Feature
Packages
[ ] @livechat/widget-core
[x] @livechat/widget-react
[ ] @livechat/widget-vue
[x] @livechat/widget-angular
Issue
Update to Angular v15.
Description
Previous Pull Request introduced update to Angular v14. This one updates the package and the example app to v15. Also updated jest, jest-preset-angular. Introduced config changes for jest.
Jest update was needed because of typescript update to v4.8 which is required by Angular v15
One thing I mentioned is the first run of tests performance that takes about 30s in my local machine. Please let me know if you faced the same issue.
Hello @walaszczykm! Can you please take a look at the PR please. Thank you!
Hi @ectuser, thanks for the PR. I will take a look at that and let you know 👍
@ectuser changes look good, and tests pass ✅ I have checked the updated example App, and it works great 👍 I am going to verify the backward compatibility of the updated Angular package with previously supported versions to make sure we will not introduce any breaking changes with this update 🙂
Hello @ectuser. I have run regression tests, and unfortunately, it looks like the upgraded Angular version used for building the package causes breaking compatibility changes 😢
I have run through projects with Angular: 12, 13, 14, and 15, and it looks like upgrading the package Angular version 15 causes incompatibility with Angular 12 and 13.
On the other hand, I have run the same regression tests with the @livechat/widget-angualr package built using Angular 13 a surprise it worked well without any issues on projects with Angular 12, 13, 14, 15. The only required thing was to extend the supported peer dependency on Angular, but despite that, nothing more was required.
Considering those results IMO, the Angular version for the package needs to be downgraded to Angular 13, and only peerDependecies inside its package.json should be updated to mark support for Angular 15.
We can move forward with the Angular version upgrade for the package when we are ready to drop support for Angular 12 and 13, but we want to do something other than that.
That means I would close this PR and prepare an appropriate downgrade on myself while keeping Angular 14 and 15 s supported peer dependency.
| gharchive/pull-request | 2022-12-28T09:59:34 | 2025-04-01T06:44:49.945368 | {
"authors": [
"ectuser",
"walaszczykm"
],
"repo": "livechat/chat-widget-adapters",
"url": "https://github.com/livechat/chat-widget-adapters/pull/64",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2275326378 | Check write permission for output locations
Annoyingly this relies on having delete permission (to delete the temp file once created) but I couldn't find any other way to achieve this check. Pretty sure there are cases where delete is already needed anyway when we overwrite certain output files.
Why don't we just call it metadata.json, put some metadata about the video in it and then leave it there
@thomshutt could you take another look? thanks
| gharchive/pull-request | 2024-05-02T11:38:54 | 2025-04-01T06:44:49.949896 | {
"authors": [
"mjh1",
"thomshutt"
],
"repo": "livepeer/catalyst-api",
"url": "https://github.com/livepeer/catalyst-api/pull/1247",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
345950259 | The status bar is not aligned with the boarder lines for smaller screen sizes.
See attached.
Describe the bug (required)
A clear and concise description of what the bug is.
Expected behavior (required)
A clear and concise description of what you expected to happen.
To Reproduce (required)
Steps to reproduce the behavior:
Go to '...'
Click on '....'
Scroll down to '....'
See error
Device (required):
Device: [e.g. MacBook Pro, iPhone6]
OS: [e.g. iOS, Linux]
Browser [e.g. Chrome, Safari]
Version [e.g. 22]
Screenshots (optional)
If applicable, add screenshots to help explain your problem.
Console logs (optional)
Paste related logs and/or error messages from the console here.
Network issues, such as the cURL output from failing requests, can also be pasted here.
Additional context (optional)
Add any other context about the problem here.
@bijanshahrokhi This is caused by the button failing to resize based upon default styling placed on it. I will change this soon.
@bijanshahrokhi This is done.
| gharchive/issue | 2018-07-30T21:54:13 | 2025-04-01T06:44:49.971035 | {
"authors": [
"Randy1Burrell",
"bijanshahrokhi"
],
"repo": "livepeer/livepeerjs",
"url": "https://github.com/livepeer/livepeerjs/issues/161",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
62511975 | Master
Make the call to LoginAsync easier.
Before the change, user has to create an IEnumerable to passing in the scope like:
LoginAsync(new List(){"wl.basic"});
Now, they can pass in the string directly:
LoginAsync("wl.basic", ...)
Thanks for the contribution!
| gharchive/pull-request | 2015-03-17T20:38:29 | 2025-04-01T06:44:49.972747 | {
"authors": [
"peternied",
"xiaomi7732"
],
"repo": "liveservices/LiveSDK-for-Windows",
"url": "https://github.com/liveservices/LiveSDK-for-Windows/pull/39",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
718810741 | [error] Error: file not exist
{
"name": "flash",
"host": "97.64.31.24",
"protocol": "ftp",
"port": 21,
"username": "admin",
"password": "123456",
"remotePath": "/www/wwwroot/www",
"uploadOnSave": true,
"connectTimeout": 30000
}
[10-11 18:54:25] [error] Error: file not exist
at l. (c:\Users\Administrator.vscode\extensions\liximomo.sftp-1.12.9\dist\extension.js:70:98188)
at Generator.next ()
at s (c:\Users\Administrator.vscode\extensions\liximomo.sftp-1.12.9\dist\extension.js:70:96928)
at processTicksAndRejections (internal/process/task_queues.js:94:5)
I have the exact same problem, a few days ago it was working fine, today when I created a new config on a new folder to have multiple connections, each with its own context I start to have this problem, even my old folders don't work anymore.
OS: Ubuntu 20.04
VSCode: Latest Version
SFTP: Latest Version
I saw an old post from 2018 where someone also had this issue but I couldn't see what the fix was for it.
Have you solved the same problem? Can you tell me about the solution
| gharchive/issue | 2020-10-11T10:55:55 | 2025-04-01T06:44:49.986414 | {
"authors": [
"RomaneeSean",
"criticaldany",
"miyaohua"
],
"repo": "liximomo/vscode-sftp",
"url": "https://github.com/liximomo/vscode-sftp/issues/837",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2000543831 | Netflix最近上线的新剧集字幕会变得特别小,且无法调整字幕的大小
如题,好像是因为插件会把中文字幕强制设定成旧样式的关系,使用edge观看时,中文字幕会变得特别小,且无法透过帐户的偏好设定调整。切换成其他浏览器观看后,便能显示正常大小的新版字幕。希望您可以帮忙解决此问题,谢谢。
网飞的原因,老版是图片字幕,新的字幕会变小,以后不知道会不会变回来,因为王冠第六季有些集数是正常的,只有第二集不正常,考虑到网飞已经强推新字幕格式的情况,所以昨天更新了版本对新版字幕样式进行适当优化(需重置网飞账号内的字幕个性化设置)
后续我也会增加选项给老版的字幕,目前能实现普通放大(比老版的略微小点,但是清晰),略损放大(比老版的略大,但是略微模糊)
网页上看上面的图1可能也有点糊,但是实际上在观影的时候是清晰的。
差不多就这么个意思了,会跟随下一次core更新一起发
| gharchive/issue | 2023-11-18T19:11:19 | 2025-04-01T06:44:49.998933 | {
"authors": [
"lkmvip",
"panggino"
],
"repo": "lkmvip/netflix-4K-DDplus",
"url": "https://github.com/lkmvip/netflix-4K-DDplus/issues/70",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2272629292 | 🛑 까사미아 www.guud.com is down
In 5095958, 까사미아 www.guud.com (http://www.guud.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 까사미아 www.guud.com is back up in 39ef785 after 14 minutes.
| gharchive/issue | 2024-04-30T23:12:10 | 2025-04-01T06:44:50.002339 | {
"authors": [
"lksjames"
],
"repo": "lksjames/monitoring",
"url": "https://github.com/lksjames/monitoring/issues/1569",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1613954442 | QR code Activity: Textbar unaligned
Hey @llaske
I have found a UI issue in the QR code activity the text bar is unaligned if it would be aligned it would look much better.
For reference I have attached a screenshot for the same, please look and review them and give me your thoughts.
Before
After
No we don't need a such alignment.
| gharchive/issue | 2023-03-07T18:03:26 | 2025-04-01T06:44:50.004606 | {
"authors": [
"Ydvaaman",
"llaske"
],
"repo": "llaske/sugarizer",
"url": "https://github.com/llaske/sugarizer/issues/1306",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2438664431 | OpenAIError: The api_key client option must be set
while running basic example I get this error
`
import os
from routellm.controller import Controller
os.environ["OPENAI_API_KEY"] = 'my api'
client = Controller(
routers=["mf"],
strong_model="gpt-4o",
weak_model="gpt-4o-mini",
)
`
'---------------------------------------------------------------------------
OpenAIError Traceback (most recent call last)
Cell In[1], line 2
1 import os
----> 2 from routellm.controller import Controller
4 os.environ["OPENAI_API_KEY"] = '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'
7 client = Controller(
8 routers=["mf"],
9 strong_model="gpt-4o",
10 weak_model="gpt-4o-mini",
11 )
File ~/playground/agent_2906/0507/RouteLLM/routellm/controller.py:10
7 from litellm import acompletion, completion
8 from tqdm import tqdm
---> 10 from routellm.routers.routers import ROUTER_CLS
12 # Default config for routers augmented using golden label data from GPT-4.
13 # This is exactly the same as config.example.yaml.
14 GPT_4_AUGMENTED_CONFIG = {
15 "sw_ranking": {
16 "arena_battle_datasets": [
(...)
27 "mf": {"checkpoint_path": "routellm/mf_gpt4_augmented"},
28 }
File ~/playground/agent_2906/0507/RouteLLM/routellm/routers/routers.py:17
12 from routellm.routers.causal_llm.llm_utils import (
13 load_prompt_format,
14 to_openai_api_messages,
15 )
16 from routellm.routers.causal_llm.model import CausalLLMClassifier
---> 17 from routellm.routers.matrix_factorization.model import MODEL_IDS, MFModel
18 from routellm.routers.similarity_weighted.utils import (
19 OPENAI_CLIENT,
20 compute_elo_mle_with_tie,
21 compute_tiers,
22 preprocess_battles,
23 )
26 def no_parallel(cls):
File ~/playground/agent_2906/0507/RouteLLM/routellm/routers/matrix_factorization/model.py:4
1 import torch
2 from huggingface_hub import PyTorchModelHubMixin
----> 4 from routellm.routers.similarity_weighted.utils import OPENAI_CLIENT
6 MODEL_IDS = {
7 "RWKV-4-Raven-14B": 0,
8 "alpaca-13b": 1,
(...)
70 "zephyr-7b-beta": 63,
71 }
74 class MFModel(torch.nn.Module, PyTorchModelHubMixin):
File ~/playground/agent_2906/0507/RouteLLM/routellm/routers/similarity_weighted/utils.py:11
8 from sklearn.linear_model import LogisticRegression
10 choices = ["A", "B", "C", "D"]
---> 11 OPENAI_CLIENT = OpenAI()
14 def compute_tiers(model_ratings, num_tiers):
15 n = len(model_ratings)
File ~/anaconda3/envs/langchain/lib/python3.11/site-packages/openai/_client.py:105, in OpenAI.init(self, api_key, organization, project, base_url, timeout, max_retries, default_headers, default_query, http_client, _strict_response_validation)
103 api_key = os.environ.get("OPENAI_API_KEY")
104 if api_key is None:
--> 105 raise OpenAIError(
106 "The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable"
107 )
108 self.api_key = api_key
110 if organization is None:
OpenAIError: The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable'
I tried with clean conda environment, same result.
Are you using any IDEs?
try config the env OPENAI_API_KEY in IDE's running config
set it to environment variable by running this in your terminal:
export OPENAI_API_KEY='your_api_key_here'
Yes, please try setting the OpenAI API key before running.
@DmitriyG228 , you should revoke the key that you pasted above.
| gharchive/issue | 2024-07-30T20:56:09 | 2025-04-01T06:44:50.206783 | {
"authors": [
"DmitriyG228",
"fengshichen",
"hmoghimifam",
"iojw",
"lee-b"
],
"repo": "lm-sys/RouteLLM",
"url": "https://github.com/lm-sys/RouteLLM/issues/38",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1394739358 | fix: cmd comment
take example has zip test command
Thanks a lot @James1x0 🙏
| gharchive/pull-request | 2022-10-03T13:30:18 | 2025-04-01T06:44:50.207927 | {
"authors": [
"James1x0",
"lmammino"
],
"repo": "lmammino/iteration-protocols-workshop",
"url": "https://github.com/lmammino/iteration-protocols-workshop/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
213149321 | Cerebro shows only 3 out of 5 ES nodes
We have a setup with 5 ES nodes in the same cluster, 3 of which are load-balanced behind nginx. The 2 last ones are not part of the load-balanced setup.
Using cerebro and pointing it to the nginx shows only 3 out of 5 nodes. Not sure if that's because those are the only ones accessuble thru nginx or if there's something else going on.
Since Cerebro performs ES queries using its backend and not simple js json calls against ES, its hard for me to say what's going on.
So I guess:
Is there a requirement for cerebro to be able to directly connect to every single node in the cluster?
If not, what should be my first stop in troubleshooting this?
@trondhindenes I imagine you mean you can only see 3 out of the 5 nodes in the overview, right?
If thats the case, could it be that 2 of your 5 nodes are not data nodes? In this case this would be correct, as the overview shows shard allocation and non master nodes cannot host shards.
On the latest version there is also a nodes page, there you should be able to see all nodes. If this is not the case, please re-open the issue :)
Thanks for replying. Turns out I wasn't on the recent version so I updated but I'm still seeing the same. Here's my "nodes" view, which shows all nodes:
And here's the overview where nodes are missing. All my nodes are data nodes.
Can you paste the output of running
"/_cat/nodes?v" on your cluster?
The nodes page shows your nodes "es11" and "es12" are just master nodes. Check your configuration for those two nodes and make sure that node.data isn't set to false if you do mean for them to be data nodes. Their CPU usage is quite low compared to the other three which does make me believe that node.data is set to false for those two.
They definetely have data on them.
cat output:
host ip heap.percent ram.percent load node.role master name
10.244.13.164 10.244.13.164 46 99 -1.00 d m es05
10.244.13.165 10.244.13.165 46 94 -1.00 d m es06
10.244.13.167 10.244.13.167 60 94 -1.00 d m es07
10.244.13.169 10.244.13.169 31 48 -1.00 d m es11
10.244.13.170 10.244.13.170 30 38 -1.00 d * es12
The 11 and 12 nodes are "cold data" nodes using spinnig disks rather than ssds for long-term storage of old logstash indexes, but they definetely have data.
@trondhindenes could I have the output of /_nodes ? Feel free to obscure any private date from it :)
@trondhindenes is this still an issue?
it is. Here's the node info from my cluster:
{
"cluster_name": "stuff-prod-es",
"nodes": {
"OTfpV0vVT_GqaPRkAwg7ww": {
"name": "zzz-prod-es07",
"transport_address": "10.244.13.167:9300",
"host": "10.244.13.167",
"ip": "10.244.13.167",
"version": "2.4.1",
"build": "c67dc32",
"http_address": "10.244.13.167:9200",
"attributes": {
"node.type": "hot"
},
"settings": {
"cluster": {
"name": "stuff-prod-es"
},
"node": {
"name": "zzz-prod-es07",
"node": {
"type": "hot"
}
},
"path": {
"conf": "F:\\ElasticSearch\\elasticsearch-2.4.1\\config",
"data": "F:/Elasticsearch/data",
"logs": "F:/ElasticSearch/elasticsearch-2.4.1/logs",
"home": "F:\\ElasticSearch\\elasticsearch-2.4.1"
},
"indices": {
"memory": {
"index_buffer_size": "50%"
}
},
"discovery": {
"zen": {
"minimum_master_nodes": "2",
"ping": {
"unicast": {
"hosts": [
"zzz-prod-es05",
"zzz-prod-es06",
"zzz-prod-es07"
]
}
}
}
},
"name": "zzz-prod-es07",
"client": {
"type": "node"
},
"config": {
"ignore_system_properties": "true"
},
"network": {
"host": "_site_",
"bind_host": "0.0.0.0"
}
},
"os": {
"refresh_interval_in_millis": 1000,
"name": "Windows Server 2012 R2",
"arch": "amd64",
"version": "6.3",
"available_processors": 2,
"allocated_processors": 2
},
"process": {
"refresh_interval_in_millis": 1000,
"id": 1240,
"mlockall": false
},
"jvm": {
"pid": 1240,
"version": "1.8.0_73",
"vm_name": "Java HotSpot(TM) 64-Bit Server VM",
"vm_version": "25.73-b02",
"vm_vendor": "Oracle Corporation",
"start_time_in_millis": 1489624004077,
"mem": {
"heap_init_in_bytes": 9663676416,
"heap_max_in_bytes": 9646243840,
"non_heap_init_in_bytes": 2555904,
"non_heap_max_in_bytes": 0,
"direct_max_in_bytes": 9646243840
},
"gc_collectors": [
"ParNew",
"ConcurrentMarkSweep"
],
"memory_pools": [
"Code Cache",
"Metaspace",
"Compressed Class Space",
"Par Eden Space",
"Par Survivor Space",
"CMS Old Gen"
],
"using_compressed_ordinary_object_pointers": "true"
},
"thread_pool": {
"force_merge": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"percolate": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"fetch_shard_started": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"listener": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"index": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 200
},
"refresh": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"suggest": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"generic": {
"type": "cached",
"keep_alive": "30s",
"queue_size": -1
},
"warmer": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"search": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"flush": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"fetch_shard_store": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"management": {
"type": "scaling",
"min": 1,
"max": 5,
"keep_alive": "5m",
"queue_size": -1
},
"get": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"bulk": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 50
},
"snapshot": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
}
},
"transport": {
"bound_address": [
"[::]:9300"
],
"publish_address": "10.244.13.167:9300",
"profiles": {}
},
"http": {
"bound_address": [
"[::]:9200"
],
"publish_address": "10.244.13.167:9200",
"max_content_length_in_bytes": 104857600
},
"plugins": [
{
"name": "cloud-aws",
"version": "2.4.1",
"description": "The Amazon Web Service (AWS) Cloud plugin allows to use AWS API for the unicast discovery mechanism and add S3 repositories.",
"jvm": true,
"classname": "org.elasticsearch.plugin.cloud.aws.CloudAwsPlugin",
"isolated": true,
"site": false
},
{
"name": "head",
"version": "master",
"description": "head - A web front end for an elastic search cluster",
"url": "/_plugin/head/",
"jvm": false,
"site": true
},
{
"name": "hq",
"version": "1.0.0",
"description": "Elastic-HQ",
"url": "/_plugin/hq/",
"jvm": false,
"site": true
},
{
"name": "mapper-attachments",
"version": "2.4.1",
"description": "The mapper attachments plugin adds the attachment type to Elasticsearch using Apache Tika.",
"jvm": true,
"classname": "org.elasticsearch.mapper.attachments.MapperAttachmentsPlugin",
"isolated": true,
"site": false
}
],
"modules": [
{
"name": "lang-expression",
"version": "2.4.1",
"description": "Lucene expressions integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.expression.ExpressionPlugin",
"isolated": true,
"site": false
},
{
"name": "lang-groovy",
"version": "2.4.1",
"description": "Groovy scripting integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.groovy.GroovyPlugin",
"isolated": true,
"site": false
},
{
"name": "reindex",
"version": "2.4.1",
"description": "_reindex and _update_by_query APIs",
"jvm": true,
"classname": "org.elasticsearch.index.reindex.ReindexPlugin",
"isolated": true,
"site": false
}
]
},
"t6FJD_WOQc6RJUFgmWIGeg": {
"name": "zzz-prod-es12",
"transport_address": "10.244.13.170:9300",
"host": "10.244.13.170",
"ip": "10.244.13.170",
"version": "2.4.1",
"build": "c67dc32",
"http_address": "10.244.13.170:9200",
"attributes": {
"data": "yes",
"node.type": "cold"
},
"settings": {
"cloud": {
"aws": {
"region": "eu-central"
}
},
"cluster": {
"name": "stuff-prod-es"
},
"node": {
"data": "yes",
"name": "zzz-prod-es12",
"node": {
"type": "cold"
}
},
"path": {
"data": "F:/apps/elasticsearch/data",
"logs": "F:/apps/elasticsearch/elasticsearch-2.4.1/logs",
"home": "F:\\apps\\elasticsearch\\elasticsearch-2.4.1"
},
"indices": {
"memory": {
"index_buffer_size": "50%"
}
},
"discovery": {
"zen": {
"minimum_master_nodes": "2",
"ping": {
"unicast": {
"hosts": [
"zzz-prod-es05",
"zzz-prod-es06",
"zzz-prod-es07"
]
}
}
}
},
"name": "zzz-prod-es12",
"client": {
"type": "node"
},
"config": {
"ignore_system_properties": "true"
},
"network": {
"host": "_site_",
"bind_host": "0.0.0.0"
}
},
"os": {
"refresh_interval_in_millis": 1000,
"name": "Windows NT (unknown)",
"arch": "amd64",
"version": "10.0",
"available_processors": 4,
"allocated_processors": 4
},
"process": {
"refresh_interval_in_millis": 1000,
"id": 2716,
"mlockall": false
},
"jvm": {
"pid": 2716,
"version": "1.8.0_73",
"vm_name": "Java HotSpot(TM) 64-Bit Server VM",
"vm_version": "25.73-b02",
"vm_vendor": "Oracle Corporation",
"start_time_in_millis": 1489686484805,
"mem": {
"heap_init_in_bytes": 17179869184,
"heap_max_in_bytes": 17145004032,
"non_heap_init_in_bytes": 2555904,
"non_heap_max_in_bytes": 0,
"direct_max_in_bytes": 17145004032
},
"gc_collectors": [
"ParNew",
"ConcurrentMarkSweep"
],
"memory_pools": [
"Code Cache",
"Metaspace",
"Compressed Class Space",
"Par Eden Space",
"Par Survivor Space",
"CMS Old Gen"
],
"using_compressed_ordinary_object_pointers": "true"
},
"thread_pool": {
"force_merge": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"percolate": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"fetch_shard_started": {
"type": "scaling",
"min": 1,
"max": 8,
"keep_alive": "5m",
"queue_size": -1
},
"listener": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": -1
},
"index": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 200
},
"refresh": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"suggest": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"generic": {
"type": "cached",
"keep_alive": "30s",
"queue_size": -1
},
"warmer": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"search": {
"type": "fixed",
"min": 7,
"max": 7,
"queue_size": 1000
},
"flush": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"fetch_shard_store": {
"type": "scaling",
"min": 1,
"max": 8,
"keep_alive": "5m",
"queue_size": -1
},
"management": {
"type": "scaling",
"min": 1,
"max": 5,
"keep_alive": "5m",
"queue_size": -1
},
"get": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"bulk": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 50
},
"snapshot": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
}
},
"transport": {
"bound_address": [
"[::]:9300"
],
"publish_address": "10.244.13.170:9300",
"profiles": {}
},
"http": {
"bound_address": [
"[::]:9200"
],
"publish_address": "10.244.13.170:9200",
"max_content_length_in_bytes": 104857600
},
"plugins": [
{
"name": "cloud-aws",
"version": "2.4.1",
"description": "The Amazon Web Service (AWS) Cloud plugin allows to use AWS API for the unicast discovery mechanism and add S3 repositories.",
"jvm": true,
"classname": "org.elasticsearch.plugin.cloud.aws.CloudAwsPlugin",
"isolated": true,
"site": false
},
{
"name": "head",
"version": "master",
"description": "head - A web front end for an elastic search cluster",
"url": "/_plugin/head/",
"jvm": false,
"site": true
},
{
"name": "hq",
"version": "1.0.0",
"description": "Elastic-HQ",
"url": "/_plugin/hq/",
"jvm": false,
"site": true
},
{
"name": "mapper-attachments",
"version": "2.4.1",
"description": "The mapper attachments plugin adds the attachment type to Elasticsearch using Apache Tika.",
"jvm": true,
"classname": "org.elasticsearch.mapper.attachments.MapperAttachmentsPlugin",
"isolated": true,
"site": false
}
],
"modules": [
{
"name": "lang-expression",
"version": "2.4.1",
"description": "Lucene expressions integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.expression.ExpressionPlugin",
"isolated": true,
"site": false
},
{
"name": "lang-groovy",
"version": "2.4.1",
"description": "Groovy scripting integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.groovy.GroovyPlugin",
"isolated": true,
"site": false
},
{
"name": "reindex",
"version": "2.4.1",
"description": "_reindex and _update_by_query APIs",
"jvm": true,
"classname": "org.elasticsearch.index.reindex.ReindexPlugin",
"isolated": true,
"site": false
}
]
},
"LtPYSHGJQt6zxEqozp-ofg": {
"name": "zzz-prod-es05",
"transport_address": "10.244.13.164:9300",
"host": "10.244.13.164",
"ip": "10.244.13.164",
"version": "2.4.1",
"build": "c67dc32",
"http_address": "10.244.13.164:9200",
"attributes": {
"node.type": "hot"
},
"settings": {
"cluster": {
"name": "stuff-prod-es"
},
"node": {
"name": "zzz-prod-es05",
"node": {
"type": "hot"
}
},
"path": {
"conf": "F:\\ElasticSearch\\elasticsearch-2.4.1\\config",
"data": "F:/Elasticsearch/data",
"logs": "F:/ElasticSearch/elasticsearch-2.4.1/logs",
"home": "F:\\ElasticSearch\\elasticsearch-2.4.1"
},
"discovery": {
"zen": {
"minimum_master_nodes": "2",
"ping": {
"unicast": {
"hosts": [
"zzz-prod-es05",
"zzz-prod-es06",
"zzz-prod-es07"
]
}
}
}
},
"name": "zzz-prod-es05",
"client": {
"type": "node"
},
"config": {
"ignore_system_properties": "true"
},
"network": {
"host": "_site_",
"bind_host": "0.0.0.0"
}
},
"os": {
"refresh_interval_in_millis": 1000,
"name": "Windows Server 2012 R2",
"arch": "amd64",
"version": "6.3",
"available_processors": 2,
"allocated_processors": 2
},
"process": {
"refresh_interval_in_millis": 1000,
"id": 1232,
"mlockall": false
},
"jvm": {
"pid": 1232,
"version": "1.8.0_73",
"vm_name": "Java HotSpot(TM) 64-Bit Server VM",
"vm_version": "25.73-b02",
"vm_vendor": "Oracle Corporation",
"start_time_in_millis": 1489557533368,
"mem": {
"heap_init_in_bytes": 9663676416,
"heap_max_in_bytes": 9646243840,
"non_heap_init_in_bytes": 2555904,
"non_heap_max_in_bytes": 0,
"direct_max_in_bytes": 9646243840
},
"gc_collectors": [
"ParNew",
"ConcurrentMarkSweep"
],
"memory_pools": [
"Code Cache",
"Metaspace",
"Compressed Class Space",
"Par Eden Space",
"Par Survivor Space",
"CMS Old Gen"
],
"using_compressed_ordinary_object_pointers": "true"
},
"thread_pool": {
"force_merge": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"percolate": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"fetch_shard_started": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"listener": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"index": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 200
},
"refresh": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"suggest": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"generic": {
"type": "cached",
"keep_alive": "30s",
"queue_size": -1
},
"warmer": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"search": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"flush": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"fetch_shard_store": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"management": {
"type": "scaling",
"min": 1,
"max": 5,
"keep_alive": "5m",
"queue_size": -1
},
"get": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"bulk": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 50
},
"snapshot": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
}
},
"transport": {
"bound_address": [
"[::]:9300"
],
"publish_address": "10.244.13.164:9300",
"profiles": {}
},
"http": {
"bound_address": [
"[::]:9200"
],
"publish_address": "10.244.13.164:9200",
"max_content_length_in_bytes": 104857600
},
"plugins": [
{
"name": "cloud-aws",
"version": "2.4.1",
"description": "The Amazon Web Service (AWS) Cloud plugin allows to use AWS API for the unicast discovery mechanism and add S3 repositories.",
"jvm": true,
"classname": "org.elasticsearch.plugin.cloud.aws.CloudAwsPlugin",
"isolated": true,
"site": false
},
{
"name": "elasticsearch-migration",
"version": "2.0.4",
"description": "Elasticsearch Migration Helper",
"url": "/_plugin/elasticsearch-migration/",
"jvm": false,
"site": true
},
{
"name": "head",
"version": "master",
"description": "head - A web front end for an elastic search cluster",
"url": "/_plugin/head/",
"jvm": false,
"site": true
},
{
"name": "hq",
"version": "1.0.0",
"description": "Elastic-HQ",
"url": "/_plugin/hq/",
"jvm": false,
"site": true
},
{
"name": "mapper-attachments",
"version": "2.4.1",
"description": "The mapper attachments plugin adds the attachment type to Elasticsearch using Apache Tika.",
"jvm": true,
"classname": "org.elasticsearch.mapper.attachments.MapperAttachmentsPlugin",
"isolated": true,
"site": false
}
],
"modules": [
{
"name": "lang-expression",
"version": "2.4.1",
"description": "Lucene expressions integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.expression.ExpressionPlugin",
"isolated": true,
"site": false
},
{
"name": "lang-groovy",
"version": "2.4.1",
"description": "Groovy scripting integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.groovy.GroovyPlugin",
"isolated": true,
"site": false
},
{
"name": "reindex",
"version": "2.4.1",
"description": "_reindex and _update_by_query APIs",
"jvm": true,
"classname": "org.elasticsearch.index.reindex.ReindexPlugin",
"isolated": true,
"site": false
}
]
},
"nxtyqYFGQESeMqHvoMRs1Q": {
"name": "zzz-prod-es06",
"transport_address": "10.244.13.165:9300",
"host": "10.244.13.165",
"ip": "10.244.13.165",
"version": "2.4.1",
"build": "c67dc32",
"http_address": "10.244.13.165:9200",
"attributes": {
"node.type": "hot"
},
"settings": {
"cluster": {
"name": "stuff-prod-es"
},
"node": {
"name": "zzz-prod-es06",
"node": {
"type": "hot"
}
},
"path": {
"data": "F:/Elasticsearch/data",
"logs": "F:/ElasticSearch/elasticsearch-2.4.1/logs",
"home": "F:\\ElasticSearch\\elasticsearch-2.4.1"
},
"indices": {
"memory": {
"index_buffer_size": "50%"
}
},
"discovery": {
"zen": {
"minimum_master_nodes": "2",
"ping": {
"unicast": {
"hosts": [
"zzz-prod-es05",
"zzz-prod-es06",
"zzz-prod-es07"
]
}
}
}
},
"name": "zzz-prod-es06",
"client": {
"type": "node"
},
"config": {
"ignore_system_properties": "true"
},
"network": {
"host": "_site_",
"bind_host": "0.0.0.0"
}
},
"os": {
"refresh_interval_in_millis": 1000,
"name": "Windows Server 2012 R2",
"arch": "amd64",
"version": "6.3",
"available_processors": 2,
"allocated_processors": 2
},
"process": {
"refresh_interval_in_millis": 1000,
"id": 1320,
"mlockall": false
},
"jvm": {
"pid": 1320,
"version": "1.8.0_73",
"vm_name": "Java HotSpot(TM) 64-Bit Server VM",
"vm_version": "25.73-b02",
"vm_vendor": "Oracle Corporation",
"start_time_in_millis": 1489632821868,
"mem": {
"heap_init_in_bytes": 7516192768,
"heap_max_in_bytes": 7498760192,
"non_heap_init_in_bytes": 2555904,
"non_heap_max_in_bytes": 0,
"direct_max_in_bytes": 7498760192
},
"gc_collectors": [
"ParNew",
"ConcurrentMarkSweep"
],
"memory_pools": [
"Code Cache",
"Metaspace",
"Compressed Class Space",
"Par Eden Space",
"Par Survivor Space",
"CMS Old Gen"
],
"using_compressed_ordinary_object_pointers": "true"
},
"thread_pool": {
"force_merge": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"percolate": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"fetch_shard_started": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"listener": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"index": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 200
},
"refresh": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"suggest": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"generic": {
"type": "cached",
"keep_alive": "30s",
"queue_size": -1
},
"warmer": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"search": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"flush": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
},
"fetch_shard_store": {
"type": "scaling",
"min": 1,
"max": 4,
"keep_alive": "5m",
"queue_size": -1
},
"management": {
"type": "scaling",
"min": 1,
"max": 5,
"keep_alive": "5m",
"queue_size": -1
},
"get": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 1000
},
"bulk": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": 50
},
"snapshot": {
"type": "scaling",
"min": 1,
"max": 1,
"keep_alive": "5m",
"queue_size": -1
}
},
"transport": {
"bound_address": [
"[::]:9300"
],
"publish_address": "10.244.13.165:9300",
"profiles": {}
},
"http": {
"bound_address": [
"[::]:9200"
],
"publish_address": "10.244.13.165:9200",
"max_content_length_in_bytes": 104857600
},
"plugins": [
{
"name": "cloud-aws",
"version": "2.4.1",
"description": "The Amazon Web Service (AWS) Cloud plugin allows to use AWS API for the unicast discovery mechanism and add S3 repositories.",
"jvm": true,
"classname": "org.elasticsearch.plugin.cloud.aws.CloudAwsPlugin",
"isolated": true,
"site": false
},
{
"name": "head",
"version": "master",
"description": "head - A web front end for an elastic search cluster",
"url": "/_plugin/head/",
"jvm": false,
"site": true
},
{
"name": "hq",
"version": "1.0.0",
"description": "Elastic-HQ",
"url": "/_plugin/hq/",
"jvm": false,
"site": true
},
{
"name": "mapper-attachments",
"version": "2.4.1",
"description": "The mapper attachments plugin adds the attachment type to Elasticsearch using Apache Tika.",
"jvm": true,
"classname": "org.elasticsearch.mapper.attachments.MapperAttachmentsPlugin",
"isolated": true,
"site": false
}
],
"modules": [
{
"name": "lang-expression",
"version": "2.4.1",
"description": "Lucene expressions integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.expression.ExpressionPlugin",
"isolated": true,
"site": false
},
{
"name": "lang-groovy",
"version": "2.4.1",
"description": "Groovy scripting integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.groovy.GroovyPlugin",
"isolated": true,
"site": false
},
{
"name": "reindex",
"version": "2.4.1",
"description": "_reindex and _update_by_query APIs",
"jvm": true,
"classname": "org.elasticsearch.index.reindex.ReindexPlugin",
"isolated": true,
"site": false
}
]
},
"rtI_0d7sSd2k6-E-Bz6Rhg": {
"name": "zzz-prod-es11",
"transport_address": "10.244.13.169:9300",
"host": "10.244.13.169",
"ip": "10.244.13.169",
"version": "2.4.1",
"build": "c67dc32",
"http_address": "10.244.13.169:9200",
"attributes": {
"data": "yes",
"node.type": "cold"
},
"settings": {
"cloud": {
"aws": {
"region": "eu-central"
}
},
"cluster": {
"name": "stuff-prod-es"
},
"node": {
"data": "yes",
"name": "zzz-prod-es11",
"node": {
"type": "cold"
}
},
"path": {
"data": "F:/apps/elasticsearch/data",
"logs": "F:/apps/elasticsearch/elasticsearch-2.4.1/logs",
"home": "F:\\apps\\elasticsearch\\elasticsearch-2.4.1"
},
"indices": {
"memory": {
"index_buffer_size": "50%"
}
},
"discovery": {
"zen": {
"minimum_master_nodes": "2",
"ping": {
"unicast": {
"hosts": [
"zzz-prod-es05",
"zzz-prod-es06",
"zzz-prod-es07"
]
}
}
}
},
"name": "zzz-prod-es11",
"client": {
"type": "node"
},
"config": {
"ignore_system_properties": "true"
},
"network": {
"host": "_site_",
"bind_host": "0.0.0.0"
}
},
"os": {
"refresh_interval_in_millis": 1000,
"name": "Windows NT (unknown)",
"arch": "amd64",
"version": "10.0",
"available_processors": 4,
"allocated_processors": 4
},
"process": {
"refresh_interval_in_millis": 1000,
"id": 3188,
"mlockall": false
},
"jvm": {
"pid": 3188,
"version": "1.8.0_73",
"vm_name": "Java HotSpot(TM) 64-Bit Server VM",
"vm_version": "25.73-b02",
"vm_vendor": "Oracle Corporation",
"start_time_in_millis": 1488917587728,
"mem": {
"heap_init_in_bytes": 17179869184,
"heap_max_in_bytes": 17145004032,
"non_heap_init_in_bytes": 2555904,
"non_heap_max_in_bytes": 0,
"direct_max_in_bytes": 17145004032
},
"gc_collectors": [
"ParNew",
"ConcurrentMarkSweep"
],
"memory_pools": [
"Code Cache",
"Metaspace",
"Compressed Class Space",
"Par Eden Space",
"Par Survivor Space",
"CMS Old Gen"
],
"using_compressed_ordinary_object_pointers": "true"
},
"thread_pool": {
"force_merge": {
"type": "fixed",
"min": 1,
"max": 1,
"queue_size": -1
},
"percolate": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"fetch_shard_started": {
"type": "scaling",
"min": 1,
"max": 8,
"keep_alive": "5m",
"queue_size": -1
},
"listener": {
"type": "fixed",
"min": 2,
"max": 2,
"queue_size": -1
},
"index": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 200
},
"refresh": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"suggest": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"generic": {
"type": "cached",
"keep_alive": "30s",
"queue_size": -1
},
"warmer": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"search": {
"type": "fixed",
"min": 7,
"max": 7,
"queue_size": 1000
},
"flush": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
},
"fetch_shard_store": {
"type": "scaling",
"min": 1,
"max": 8,
"keep_alive": "5m",
"queue_size": -1
},
"management": {
"type": "scaling",
"min": 1,
"max": 5,
"keep_alive": "5m",
"queue_size": -1
},
"get": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 1000
},
"bulk": {
"type": "fixed",
"min": 4,
"max": 4,
"queue_size": 50
},
"snapshot": {
"type": "scaling",
"min": 1,
"max": 2,
"keep_alive": "5m",
"queue_size": -1
}
},
"transport": {
"bound_address": [
"[::]:9300"
],
"publish_address": "10.244.13.169:9300",
"profiles": {}
},
"http": {
"bound_address": [
"[::]:9200"
],
"publish_address": "10.244.13.169:9200",
"max_content_length_in_bytes": 104857600
},
"plugins": [
{
"name": "cloud-aws",
"version": "2.4.1",
"description": "The Amazon Web Service (AWS) Cloud plugin allows to use AWS API for the unicast discovery mechanism and add S3 repositories.",
"jvm": true,
"classname": "org.elasticsearch.plugin.cloud.aws.CloudAwsPlugin",
"isolated": true,
"site": false
},
{
"name": "head",
"version": "master",
"description": "head - A web front end for an elastic search cluster",
"url": "/_plugin/head/",
"jvm": false,
"site": true
},
{
"name": "hq",
"version": "1.0.0",
"description": "Elastic-HQ",
"url": "/_plugin/hq/",
"jvm": false,
"site": true
},
{
"name": "mapper-attachments",
"version": "2.4.1",
"description": "The mapper attachments plugin adds the attachment type to Elasticsearch using Apache Tika.",
"jvm": true,
"classname": "org.elasticsearch.mapper.attachments.MapperAttachmentsPlugin",
"isolated": true,
"site": false
}
],
"modules": [
{
"name": "lang-expression",
"version": "2.4.1",
"description": "Lucene expressions integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.expression.ExpressionPlugin",
"isolated": true,
"site": false
},
{
"name": "lang-groovy",
"version": "2.4.1",
"description": "Groovy scripting integration for Elasticsearch",
"jvm": true,
"classname": "org.elasticsearch.script.groovy.GroovyPlugin",
"isolated": true,
"site": false
},
{
"name": "reindex",
"version": "2.4.1",
"description": "_reindex and _update_by_query APIs",
"jvm": true,
"classname": "org.elasticsearch.index.reindex.ReindexPlugin",
"isolated": true,
"site": false
}
]
}
}
}
Ok, I see the problem. The nodes you do not see listed have the data property set to "yes". I have actually just found out this was possible, and hence, only accepted a boolean as a valid value for this. I will fix this to also support yes/no as value.
thanks for the feedback :)
ah, sweet. Yep, those are some huge json's - and even Elastic themselves don't always follow their own spec documentation. Anyways, glad you figured it out, its a fantastic app!
| gharchive/issue | 2017-03-09T20:05:26 | 2025-04-01T06:44:50.254538 | {
"authors": [
"lmenezes",
"marke72",
"trondhindenes"
],
"repo": "lmenezes/cerebro",
"url": "https://github.com/lmenezes/cerebro/issues/91",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2029020530 | Improvement: Unmarshal withdrawals from RPC blocks
This adds Withdrawals to blocks retrieved via RPC.
Thanks for merging (and also for the improvements!) 🙏
| gharchive/pull-request | 2023-12-06T17:11:19 | 2025-04-01T06:44:50.260783 | {
"authors": [
"joel-u410"
],
"repo": "lmittmann/w3",
"url": "https://github.com/lmittmann/w3/pull/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2488131995 | How to dynamically obtain the current leader
How to dynamically obtain the current leader
Please read all docs carefully before making any further issues.
Please see GetLeaderID() for what you want.
| gharchive/issue | 2024-08-27T02:35:09 | 2025-04-01T06:44:50.285879 | {
"authors": [
"lni",
"xiatianqiuyu"
],
"repo": "lni/dragonboat",
"url": "https://github.com/lni/dragonboat/issues/367",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
856410238 | [WIP] extract form factor base models [DXP-112]
I'm still getting the address models to work the way we need them to in all their glory, but the form factor and lob base work is done.
The core of this PR is:
lob_base.yml All non-AV Lob resources (*) have these properties, whether Form Factors or not
shared/models/form_factor Properties common to all Form Factors
editable.yml Properties which customers can customize
generated.yml Properties which we generate
to.yml The to address, which has rules sufficiently complex that it is its own model. Thus far, to is the same across Form Factors.
from.yml The from address has different rules sufficiently complex that it is own model.
from_us.yml Most Form Factors only allow from addresses in the US
And then there is:
shared/models/address Address models, US and International, editable and stored, inline vs not. The address models are driven by a combination of how polymorphism and inheritance work in OpenAPI v3 and the fact that we require a 2-digit country code in the input model which we then convert into the full length name for the printed address. Because the possible values for the discriminator for a polymorphic class must be enumerated, the full list of country codes has to be listed in shared/models/address/address.yml and shared/models/address/address_editable.yml.
(*) Certificates do not :sob:
replaced by #71
| gharchive/pull-request | 2021-04-12T22:34:07 | 2025-04-01T06:44:50.297621 | {
"authors": [
"hilary"
],
"repo": "lob/lob-openapi",
"url": "https://github.com/lob/lob-openapi/pull/62",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
94312785 | GEOMESA-819 Including size of query box in z3 range calculation
Start recursion at the longest common prefix of the min/max bounds
Depth of recursion depends on the size of our query region
Compared to fixed recursion depth, increases ranges for small query regions and decreases ranges for large query regions
Most queries will create less than 10k ranges - usually on the order of 2-4k for moderate queries and 30-500 for small queries
Signed-off-by: Emilio Lahr-Vivaz elahrvivaz@ccri.com
5789033
| gharchive/pull-request | 2015-07-10T14:06:47 | 2025-04-01T06:44:50.326203 | {
"authors": [
"anthonyccri",
"elahrvivaz"
],
"repo": "locationtech/geomesa",
"url": "https://github.com/locationtech/geomesa/pull/638",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
524922619 | Save computation time in DDP and FDDP solvers + fixed bug in DDP - [merged]
In GitLab by @cmastalli on Apr 5, 2019, 17:44
Merges topic/solvers-improvement -> devel
This is a list of changes in this PR:
Stored the $V_{xx}$ at the rollout state (not the guess one) in the SolverFDDP.
Changed the gap definition in SolverFDDP to make it compatible with PDF.
Allocated onces the following data: xs, us, xs_try, us_try and gaps. This was done for all solvers.
Created a computeGains function inside SolverFDDP which allows us to use SolverBoxDDP with SolverFDDP.
Fixed a bug in the SolverDDP, wrong computation of $Q_x$ and $Q_u$. With this, I have to change few log files (see 63155a6ab2828bae7d2ca2c65 for more details).
Little improvement in the documentation of SolverDDP and SolverFDDP.
In GitLab by @cmastalli on Apr 8, 2019, 14:50
changed title from Save computation{-t-} time in DDP and FDDP solvers to Save computation time in DDP and FDDP solvers
In GitLab by @cmastalli on Apr 9, 2019, 00:02
added 1 commit
b3eaade7 - [solvers] Stored Vx+, the gradient computed at the end of interval (ddp/fddp)
Compare with previous version
In GitLab by @cmastalli on Apr 9, 2019, 00:13
added 1 commit
190d9e73 - [solvers] Stored Vx+, the gradient computed at the end of interval (ddp/fddp)
Compare with previous version
In GitLab by @cmastalli on Apr 9, 2019, 00:19
added 1 commit
bc3be056 - [format] Fixed an issue with the format
Compare with previous version
In GitLab by @cmastalli on Apr 11, 2019, 18:09
added 1 commit
6ad2aad1 - [logs] These changes are due to the imposition of Vxx symmetric
Compare with previous version
In GitLab by @cmastalli on Apr 11, 2019, 18:11
added 12 commits
6ad2aad1...5ce975fc - 6 commits from branch loco-3d:develab2a3b2f - [fddp] Store the gradient value at gap point.63ccaca3 - [fddp] Fixed the code format6e5683d7 - [fddp] Changed the definition of the gaps to make it compatible with PDF9f38e437 - [fddp] self.gaps are allocated once164217e1 - [fddp] Created computeGains function which allows us to use BoxDDPc75234a3 - [fddp] Cleaned up a bit the code
Compare with previous version
In GitLab by @cmastalli on Apr 11, 2019, 18:12
marked as a Work In Progress
In GitLab by @cmastalli on Apr 12, 2019, 14:20
added 5 commits
ef746c8e - [ddp] self.gaps is allocated only once time63155a6a - [bugfix][ddp] Fixed wrong computation of Qx and Quaf50af44 - [fddp] Relinearized after Vxx regularization + cleaned up the code3ff64160 - [solvers] xs, us are allocated only onces, therefore setCandidate always copy data14be665f - [solvers] xs_try and us_try are allocated only onces, this save computation time
Compare with previous version
In GitLab by @cmastalli on Apr 12, 2019, 14:22
added 1 commit
e0612d21 - [cleanup] Removed useless variable in both solvers: ddp and fddp
Compare with previous version
In GitLab by @cmastalli on Apr 12, 2019, 14:27
added 1 commit
d1b99fa0 - [format] Added desired space in asserts
Compare with previous version
In GitLab by @cmastalli on Apr 12, 2019, 14:34
changed title from WIP: Save computation time in DDP and FDDP solvers to WIP: Save computation time in DDP and FDDP solvers{+ + fixed bug in DDP+}
In GitLab by @cmastalli on Apr 12, 2019, 14:34
changed the description
In GitLab by @cmastalli on Apr 12, 2019, 14:35
unmarked as a Work In Progress
In GitLab by @cmastalli on Apr 14, 2019, 23:30
changed the description
In GitLab by @nmansard on Apr 15, 2019, 08:51
I don't have time to handle that seriously today. I can do it tomorrow, or someone else can peer-review it now.
In GitLab by @nmansard on Apr 17, 2019, 10:27
why did you remove the copy=True? I think this is not safe and does not save much.
In GitLab by @nmansard on Apr 17, 2019, 10:31
same here: I think we better keep copy=True.
In GitLab by @nmansard on Apr 17, 2019, 10:32
Are you sure that copy=False by default?
In GitLab by @nmansard on Apr 17, 2019, 10:35
this choice is done to fit to the doc, is it not?
In GitLab by @nmansard on Apr 17, 2019, 10:46
so you made the choice to always copy. i think this is ok.
In GitLab by @cmastalli on Apr 17, 2019, 11:37
Yes, it is. This also fits the gap definition for both solvers: DDP and FDDP. This choice was done here: 6e5683d7ffd1fc4abf77aa65d4107bde351f45d0. Do you agree with it?
In GitLab by @cmastalli on Apr 17, 2019, 11:38
Yes, that's right. I will resolve the other discussion since you agree with this choice.
In GitLab by @cmastalli on Apr 17, 2019, 11:39
Explained above
In GitLab by @cmastalli on Apr 17, 2019, 11:39
Explained above
In GitLab by @cmastalli on Apr 17, 2019, 11:39
Explained above
In GitLab by @cmastalli on Apr 17, 2019, 11:39
resolved all discussions
In GitLab by @nmansard on Apr 17, 2019, 13:42
resolved all discussions
In GitLab by @nmansard on Apr 17, 2019, 13:43
merged
In GitLab by @nmansard on Apr 17, 2019, 13:43
mentioned in commit 04c8bd3b9f5e7b9e47e23ce1a3389c46d6b27c1c
In GitLab by @nmansard on Apr 17, 2019, 13:45
mentioned in merge request !159
| gharchive/issue | 2019-11-19T10:39:20 | 2025-04-01T06:44:50.374155 | {
"authors": [
"wxmerkt"
],
"repo": "loco-3d/crocoddyl",
"url": "https://github.com/loco-3d/crocoddyl/issues/500",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
217265078 | How to stop once a user finishes it's set of Tasks.
Hi, I am using locust for load testing of my app. I am using it with Selenium. The Flow of test is :
User opens the web app, which takes him to login page.
Web-driver finds elements for username, password, and Sign in button. Enters respective values, and when clicks sign in buttons.
then navigates to a page, lets call it page X. (Page X is protected by login).
When page X load, it triggers 5 queries in the back-end. I want to get the number-of-queries/Sec.
Everything goes fine till step 3, but after opening, page X, it gives me 'NoSuchElmentException'. Reason: after finishing the task, Locust fires the task again for that particular user. Now browser is on page X, the test started again, so it searches username and password field on page X, which triggers, aforementioned exception. I thought to get around this problem by closing driver, but then I get no such window exception.
So, I want to stop the test for that particular user, after he gets into page X.
try setting --num-request
but usually it's not very useful to just run one iteration
Won't this work: http://docs.locust.io/en/latest/writing-a-locustfile.html#the-on-start-function ?
or am I missing something as to what you are trying to do?
Okay if anyone can help me with this:
I want to remove phantomjs as it is heavy, and I can't run multiple instances of it (more than 60-70). I tried with distributed locust but I am still not able to send more than 2-3 requests per second. I am planning to use HtmlUnitWithJS, but can't get it work with Locust. What I need to pass in LocustUser class? Thanks in advance
@thekgt please open a new issue rather than hijacking this one
Did this every get resolved?
I'm building a script that will run as a synthetic monitoring script that will run once every 15 minutes. If it fails it'll kick off an alert via Grafana. How can I set the script to run just once? I tried the --num-request and it doesn't seem to be part of locust.
| gharchive/issue | 2017-03-27T14:30:41 | 2025-04-01T06:44:50.382881 | {
"authors": [
"aldenpeterson-wf",
"cgoldberg",
"mjohnson767",
"thekgt"
],
"repo": "locustio/locust",
"url": "https://github.com/locustio/locust/issues/567",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
247808720 | Issue with indendation on a PUT request (Newbie)
Hi All,
Really struggling, I need to make a simple PUT request to our API endpoint. I've written what I think would work? But honestly I have little to no python experience.
Can someone please help me out to get a working request? Attached is the error im currently getting... I'll paste my python file..
error.txt
from locust import HttpLocust, TaskSet, task, json
class loadParcel(TaskSet):
def on_start(self):
self.scanParcel()
@task
def scanParcel(self):
headers = {'content-type': 'application/json' , 'dpdsession':'06784db0-784a-11e7-a1fa-85ad1b51d266', 'highlander':'true', 'cache-control':'no-cache', 'dpdclient':'application/json'}
payload = {"driverCode":"0056*FD8608", "labelNumber":"%0ML11XR15500748427135832826", "dateTime":"2017-08-03 03:00:41" }
self.client.put("/depot/0056/route/796/?action=loadParcel", data=json.dumps(payload) , headers=headers)
class WebsiteUser(HttpLocust):
task_set = loadParcel
min_wait = 1000
max_wait = 1000
@ashleigh-robinson In python you must indent method definitions 4 spaces from the beginning of the namespace, and statements must be indented 4 spaces from the declaration of the method block. Also, in Python method names use snake_case. Here is your file re-indented. You can install Flake8 Python linter sublime package to show you these errors and more.
from locust import HttpLocust, TaskSet, json, task
class LoadParcel(TaskSet):
def on_start(self):
self.scanParcel()
@task(1)
def scan_parcel(self):
headers = {'content-type': 'application/json', 'dpdsession': '06784db0-784a-11e7-a1fa-85ad1b51d266', 'highlander': 'true', 'cache-control': 'no-cache', 'dpdclient': 'application/json'}
payload = {"driverCode": "0056*FD8608", "labelNumber": "%0ML11XR15500748427135832826", "dateTime": "2017-08-03 03:00:41"}
self.client.put("/depot/0056/route/796/?action=loadParcel", data=json.dumps(payload), headers=headers)
class WebsiteUser(HttpLocust):
task_set = LoadParcel
min_wait = 1000
max_wait = 1000
@bgroupe thank you so much for your help, I understand now, I ran the above, it failed as I wrongly named the definition, after altering this it now succeeds!
One last question for everyone, and I promise I'll leave you all alone, how do i log the request response body?
Thanks
Depends on what you mean by log -- if you want to log your requests and responses for every call you can just import the Python logging module:
import logging as Logger
class LoadParcel(TaskSet):
...
def scan_parcel(self):
...
response = self.client.put("/resource", headers=headers, json=payload)
Logger.info(response.request.headers)
Logger.info(response.request.body)
...
If you want to view the request body in the locust web interface on failure, I would look at the response context manager built into Locust, which will output to the Failures tab.
| gharchive/issue | 2017-08-03T19:19:17 | 2025-04-01T06:44:50.389183 | {
"authors": [
"ashleigh-robinson",
"bgroupe"
],
"repo": "locustio/locust",
"url": "https://github.com/locustio/locust/issues/641",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2138230535 | [CUDA out of memory] ResNet18 & Cifar100
Hi logancyang, I have successfully run your code using ResNet18 on Cifar100 with bs=16, epochs=200. However, during the 4th epoch, the GPU VRAM overflowed. Upon reviewing the code, I found that this line is causing the GPU RAM to accumulate at each step. Do you have a solution for this problem?
https://github.com/logancyang/loss-landscape-anim/blob/a8d5610ff433b94aca68af42952f0bad8e0e34b3/loss_landscape_anim/model.py#L180
hi @lhchau, thanks for checking this out. This project is not actively maintained at the moment, but feel free to make your own fork if you have time to implement a solution.
I also had the same issue. any solution @lhchau
| gharchive/issue | 2024-02-16T09:46:09 | 2025-04-01T06:44:50.424225 | {
"authors": [
"alqurri77",
"lhchau",
"logancyang"
],
"repo": "logancyang/loss-landscape-anim",
"url": "https://github.com/logancyang/loss-landscape-anim/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
125579483 | Update README.md
changed to match logentries.com/doc/android
Hey @ConallH note there seems to be conflicts. Maybe your local repo is not up-to-date?
| gharchive/pull-request | 2016-01-08T10:01:59 | 2025-04-01T06:44:50.436220 | {
"authors": [
"ConallH",
"vilda"
],
"repo": "logentries/le_android",
"url": "https://github.com/logentries/le_android/pull/34",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
66718785 | Logging MDC with Log4j
I'm using the le_java appender but my MDC values are not being passed to Logentries.
Is this a supported feature at le_java?
https://github.com/logentries/le_java/issues/19
Can you check if the above solves this issue?
Thanks
| gharchive/issue | 2015-04-06T22:19:52 | 2025-04-01T06:44:50.438008 | {
"authors": [
"nykolaslima",
"samjamesobrien"
],
"repo": "logentries/le_java",
"url": "https://github.com/logentries/le_java/issues/46",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.