id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
102412945
Facet sort *** Note: requires search-facets-stub branch of Honeycomb *** I created separate react components for each of the main widgets: searchbox, facets, sort dropdown. Create search box widget that can be displayed on any page. Create sort and facets widgets based on JSON data returned from ajax request. Only show them if they are returned. Pass the search page's url's query strings back to the ajax search request. :+1:
gharchive/pull-request
2015-08-21T16:13:38
2025-04-01T06:45:06.711970
{ "authors": [ "danthewolfe", "jgondron" ], "repo": "ndlib/beehive", "url": "https://github.com/ndlib/beehive/pull/179", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
476349904
No colors GHC 8.6.5, ghcid 0.7.5 and no colors in gnome-terminal (Ubuntu 16.04). I seem to recall that they were there. Is it a known problem? Okay, this is the problem only until first reload. Still, fills like a bug. I get this when running :Ghcid; might be related? Unknown flag: --color See https://github.com/ndmitchell/ghcid/blob/master/README.md#i-only-see-source-spans-or-colors-on-errorswarnings-after-the-first-load and https://twitter.com/ndm_haskell/status/1159207825194856449?s=20 which gives the background. Known issue, and working as intended. I see, thanks!
gharchive/issue
2019-08-02T20:46:23
2025-04-01T06:45:06.714749
{ "authors": [ "DestyNova", "ndmitchell", "ulysses4ever" ], "repo": "ndmitchell/ghcid", "url": "https://github.com/ndmitchell/ghcid/issues/269", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
866796155
Time travel simulation In some cases we want to test things happen only after a specific amount of time, for example: in staking pool contract, staking reward is distributed after one epoch in berry farm, reward is distributed after a minute One way of doing this is let near-sdk-sim runtime process N empty blocks. It currently have way of process block but not exposed Actually it's already doable let root = init_simulator(...) root.borrow_runtime_mut().produce_blocks(N) Practically it doesn't work, for example, simulate produce_blocks(86400) takes hours I addressed time simulation problem with https://github.com/near/near-sdk-rs/pull/378, this SIGSEGV: invalid memory reference I'll open a different issue
gharchive/issue
2021-04-24T17:24:36
2025-04-01T06:45:06.730418
{ "authors": [ "ailisp" ], "repo": "near/near-sdk-rs", "url": "https://github.com/near/near-sdk-rs/issues/377", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
85700259
Please upgrade to node:0.12 +1 Updated!
gharchive/issue
2015-06-06T05:12:41
2025-04-01T06:45:06.737104
{ "authors": [ "jlchereau", "mcollina" ], "repo": "nearform/docker-logentries", "url": "https://github.com/nearform/docker-logentries/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
182810725
As a Labs developer I'd like access to a CI environment... ...so that (CI) the Authorization module can be automatically tested every time there is a commit to the master branch Part of Epic #29 Part of this will be implemented in https://github.com/nearform/labs-authorization/issues/285 I'm not sure if we need to provide "deployment scripting" for it. I fear it may end up too generic or too specific. Thoughts? @nherment @paolochiodi @mihaidma @dberesford @p16 The initial discussion was that we'll set a private hand made CI system. Now that the project is public, tools like Travis are doing most of this work. We should still wait for feedback before closing it. maybe @dgonzalez knows more on this. /cc @p16 I think we can close this. We do have a CI in place
gharchive/issue
2016-10-13T14:47:09
2025-04-01T06:45:06.745683
{ "authors": [ "floridemai", "irelandm", "mihaidma", "p16", "paolochiodi" ], "repo": "nearform/labs-authorization", "url": "https://github.com/nearform/labs-authorization/issues/37", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2069333348
Integrate JupyterHub App Launcher into Nebari Reference Issues or PRs #2011 Basic Installation docs: https://github.com/nebari-dev/nebari-docs/pull/391 [x] Integrate JHub Apps (make it optional) [x] JupyterHub helm chart Upgrade to 3.2.1 (JupyterHub: 4.0.2) [x] Fetch conda environments dynamically from conda-store API [x] Update kindest/node to v1.23.13 as JupyterHub helm chart didn't supported k8s < 1.23 [x] Create jwt token secret for JHub Apps Authentication (mounted on hub pod) [x] Add ingress route for jhub apps service What does this implement/fix? Put a x in the boxes that apply [ ] Bug fix (non-breaking change which fixes an issue) [x] New feature (non-breaking change which adds a feature) [ ] Breaking change (fix or feature that would cause existing features not to work as expected) [ ] Documentation Update [ ] Code style update (formatting, renaming) [ ] Refactoring (no functional changes, no API changes) [ ] Build related changes [ ] Other (please describe): Testing [x] Did you test the pull request locally? [ ] Did you add new tests? Any other comments? Yes, there is one here: https://github.com/nebari-dev/nebari-docs/pull/391 with minimal stuff required to get it up and running.
gharchive/pull-request
2024-01-07T23:06:26
2025-04-01T06:45:06.756839
{ "authors": [ "aktech" ], "repo": "nebari-dev/nebari", "url": "https://github.com/nebari-dev/nebari/pull/2185", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
322565443
test functionality through chat by another contact how to create anther contact and test implemented functionality? or I would thankful if you share a scenario in order to test development functionality regist contact setting contact's username search contact by username regist contact ------------> do it by add manually in user table in mysql or add in android client? setting contact's username ------------> do it in android client? search contact by username ------------> do it in android client?
gharchive/issue
2018-05-13T05:14:26
2025-04-01T06:45:06.768338
{ "authors": [ "wubenqi", "yumcoder-platform" ], "repo": "nebulaim/telegramd", "url": "https://github.com/nebulaim/telegramd/issues/64", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1597666588
Local S3 support Hello! Great work with this! Do you have plans to allow local S3 ops? Would be great for local development! For e.g., we use LocalStack for mocking S3 (and other AWS services) locally as we develop. I suspect a relatively low lift solution could be to expose --endpoint-url as a configurable param? Sure, I can add a feature to provide your custom endpoint. I added this feature to my to-do list. I don't have localstack in my local machine to test it out. Can you help me with this test? download https://github.com/necatiarslan/aws-s3/raw/main/vsix/aws-s3-vscode-extension-1.1.2.vsix and install Run "Update Aws EndPoint" command and provide your localstack URL Pls let me know if it worked or you had an error That worked just as expected! thanks
gharchive/issue
2023-02-23T23:31:14
2025-04-01T06:45:06.773568
{ "authors": [ "mhonnatti", "necatiarslan" ], "repo": "necatiarslan/aws-s3", "url": "https://github.com/necatiarslan/aws-s3/issues/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1932793918
0347-top-k-frequent-elements.dart File(s) Modified: 0347-top-k-frequent-elements.dart Language(s) Used: Dart Submission URL: _https://leetcode.com/problems/top-k-frequent-elements/submissions/1070846902/_ Important Please make sure the file name is lowercase and a duplicate file does not already exist before merging. @AkifhanIlgaz Thanks for your contribution 👍
gharchive/pull-request
2023-10-09T10:49:35
2025-04-01T06:45:06.787424
{ "authors": [ "AkifhanIlgaz", "Ykhan799" ], "repo": "neetcode-gh/leetcode", "url": "https://github.com/neetcode-gh/leetcode/pull/3054", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2545467725
Create 1456-maximum-number-of-vowels-in-a-substring-of-given-length.c File(s) Modified: 1456-maximum-number-of-vowels-in-a-substring-of-given-length Language(s) Used: c Submission URL: _https://leetcode.com/problems/maximum-number-of-vowels-in-a-substring-of-given-length/submissions/1400672305/_ @adnanqidwai Thank you for your contribution 👍
gharchive/pull-request
2024-09-24T13:47:18
2025-04-01T06:45:06.789930
{ "authors": [ "Ykhan799", "adnanqidwai" ], "repo": "neetcode-gh/leetcode", "url": "https://github.com/neetcode-gh/leetcode/pull/3637", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
71132537
Make install more configurable Some distros needs more flexibility. For example Exherbo uses /usr/x86_64-pc-linux-gnu as a PREFIX on x86 64bit platforms while man-pages still resides in /usr/share/man. Also some tools have target-specific prefix like x86_64-pc-linux-gnu-pkg-config. allow overriding BINDIR and MANDIR allow overriding pkg-config Sounds good, thanks!
gharchive/pull-request
2015-04-26T23:03:55
2025-04-01T06:45:06.806296
{ "authors": [ "nelhage", "ony" ], "repo": "nelhage/reptyr", "url": "https://github.com/nelhage/reptyr/pull/55", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
748012011
Unable to add json Address Book after attaching invalid format file Steps to reproduce: Make sure there is no contacts Go to 'My contacts' tab and click 'Import address book' Select or drag a file with invalid format (any but not json file) Select a valid format file with an Address book (json) Expected result: A user is able to add a json address book Actual result: A user is unable to add a json address book file after once attaching the invalid one (not json) From issue https://github.com/nemgrouplimited/symbol-desktop-wallet/issues/736 When you click the import address book , upload the address book (json file) on first go , it is working, but the scenario i tried is I uploaded a invalid address file , it gave invalid address, but when i uploaded the correct one, it is still showing invalid address,not moving further Solved.
gharchive/issue
2020-11-21T12:40:00
2025-04-01T06:45:06.819230
{ "authors": [ "cryptoBeliever" ], "repo": "nemgrouplimited/symbol-desktop-wallet", "url": "https://github.com/nemgrouplimited/symbol-desktop-wallet/issues/724", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
887544771
Explorer graphic doesn't display the most important tx info about a tx easily enough, which is what was actually sent. The Transaction Graphics are great. But when a person is sending XYM, which is the majority of the txs on the network, it doesn't say how much was sent. Instead a person has to scroll down. I've used a lot of explorers and this is the only one that I had to scroll to get to an amount. Or when there is an aggregate, I have to click to get more details which is definitely outside of a normal users expectations when using an explorer IMO. Almost all users I guess look in the "Transaction Info" section to try to find how much was sent but it isn't there so that is a bit confusing for them I guess. So basically a person checks the graphic, and its not there. Then they check the Transaction Info, and it isn't there. Then they scroll down and get the info from Transaction Detail. @jabo38 thank you! If you hover cursor on the purple Connector icon, it will show you the transferred amount. The same works with the message and other icons on the arrow. But I agree with you that it might be unclear and could be designed better. We will have a look how to improve it Oh. That’s a nice feature. I’ll use it now that I know. I’m a bit worried there might be more people like me because up until now I’ve used the explorer a lot and didn’t discover that or if I did I forgot about it and went about doing it the other way. I now see hovering over the people gives the full address too which is nice and there are other hover options too which are nice. Get Outlook for iOShttps://aka.ms/o0ukef From: OlegMakarenko @.> Sent: Thursday, May 13, 2021 11:24:15 AM To: nemgrouplimited/symbol-explorer @.> Cc: jabo38 @.>; Mention @.> Subject: Re: [nemgrouplimited/symbol-explorer] Explorer graphic doesn't display the most important tx info about a tx easily enough, which is what was actually sent. (#887) @jabo38https://github.com/jabo38 thank you! If you hover cursor on the purple Connector icon, it will show you the transferred amount. The same works with the message and other icons on the arrow. But I agree with you that it might be unclear and could be designed better. We will have a look how to improve it — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHubhttps://github.com/nemgrouplimited/symbol-explorer/issues/887#issuecomment-840636135, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ACPKCA7CY74ZUWRBNVW6HRDTNPVJ7ANCNFSM44VX5FYQ.
gharchive/issue
2021-05-11T14:57:29
2025-04-01T06:45:06.825550
{ "authors": [ "OlegMakarenko", "jabo38" ], "repo": "nemgrouplimited/symbol-explorer", "url": "https://github.com/nemgrouplimited/symbol-explorer/issues/887", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
729121450
Port Awareness in healthCheck Describe the bug If the ports in docker-compose.yml are rewritten, a portion of the IP address is recognized as a port. To Reproduce Edit target/docker/docker-compose.yml api-node: ports: - '127.0.0.1:7900:7900' api-broker: ports: - '127.0.0.1:7902:7902' rest-gateway: ports: - '127.0.0.1:3000:3000' Expected behavior I think it's better to be recognized with 127.0.0.1:3000 instead of 127. Logs $ symbol-bootstrap healthCheck _ _ _ _ _ ___ _ _ _ __ ___ | |__ ___ | | | |__ ___ ___ | |_ ___ | |_ _ __ __ _ _ __ / __| | | | | | '_ ` _ \ | '_ \ / _ \ | | _____ | '_ \ / _ \ / _ \ | __| / __| | __| | '__| / _` | | '_ \ \__ \ | |_| | | | | | | | | |_) | | (_) | | | |_____| | |_) | | (_) | | (_) | | |_ \__ \ | |_ | | | (_| | | |_) | |___/ \__, | |_| |_| |_| |_.__/ \___/ |_| |_.__/ \___/ \___/ \__| |___/ \__| |_| \__,_| | .__/ |___/ |_| 2020-10-25T19:42:12.976Z info Container db is running 2020-10-25T19:42:12.977Z info Container api-node is running 2020-10-25T19:42:12.981Z info Container api-broker is running 2020-10-25T19:42:12.981Z info Container rest-gateway is running 2020-10-25T19:42:12.983Z warn Container api-node port 127 -> 7900 is NOT openYET. 2020-10-25T19:42:12.984Z warn Container api-broker port 127 -> 7902 is NOT open YET. 2020-10-25T19:42:12.984Z warn Container rest-gateway port 127 -> 3000 is NOT open YET. 2020-10-25T19:42:12.984Z info Retrying in 10 seconds. Polling will stop in 49.934 seconds ...... $ curl http://127.0.0.1:3000/node/health {"status":{"apiNode":"up","db":"up"}} Desktop (please complete the following information): OS: Ubuntu x64 Tool Version: symbol-bootstrap/0.2.0 Docker Version: Docker version 19.03.13, build 4484c46d9d Node Version: node-v14.14.0 Additional context I'm building a tor node(3br3ynmu7amoiclaetx4iu4v43qqfu5p4hcy3upvsdthaviz5gfpjbad.onion). This has been fixed in the current version of bootstrap
gharchive/issue
2020-10-25T21:44:36
2025-04-01T06:45:06.844322
{ "authors": [ "fboucquez", "wakiyamap" ], "repo": "nemtech/symbol-bootstrap", "url": "https://github.com/nemtech/symbol-bootstrap/issues/52", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
306138355
Update for getting-started-python.md (fixes #323) The updated file contains updated documentation to match readthedocs, most noticeably changing from Python 3.5 to Python 3.6. The relative links under "Further Reading" were not functional. The links are now fixed. 👍 This is a better PR than #448. Just wondering -- does it make sense to duplicate the neo-python docs into these docs, or should wes imply link to the neo-python docs? @localhuman @metachris You are right. I updated the file to directly point to neo-python docs. Please review it. https://github.com/neo-project/docs/pull/450 @Celia18305 why was this closed? The neo-python docs on the homepage are outdated, and this PR fixes them. Hi @metachris , becasue in the later PR #450 the file was updated to directly point to neo-python docs. I approved that update. Thanks! Didn't notice...
gharchive/pull-request
2018-03-17T08:14:43
2025-04-01T06:45:06.864735
{ "authors": [ "BSathvik", "Celia18305", "metachris" ], "repo": "neo-project/docs", "url": "https://github.com/neo-project/docs/pull/449", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
176945874
Main class, logging dependencies and authentication in URL, unit test: I was not able to run this example and these are the fixes I had to make: Main class exists in a different package Adding slf4j and log4j dependencies to get logging. Connection URL withuser:password authentication (enabled by default in a new Neo4j Installation). Updated Readme accordingly. Converted DocTest into a unit test that uses the MovieService. Really useful sample project to get started! Thanks. 👍 please merge these fixes Thanks a lot !
gharchive/pull-request
2016-09-14T15:52:48
2025-04-01T06:45:06.894519
{ "authors": [ "jexp", "netrebel", "roybailey" ], "repo": "neo4j-examples/neo4j-movies-java-bolt", "url": "https://github.com/neo4j-examples/neo4j-movies-java-bolt/pull/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
55820436
Could not find command activerecord Executing neo4apis activerecord --help from the terminal on Mac 10.10.1 results in message "Could not find command activerecord" neo4apis-activerecord gem installed with RVM. Somebody else was having this problem, and I just realized how to debug it. Looking at it now Ok, I just released a 0.5.1 version of neo4apis which I think will fix the problem. A couple of notes: Make sure to update the version of the base neo4apis gem. You may need to uninstall / reinstall the gems to make sure This fix will only work for rubygems, not any other gem package managers. I wanted to get something out there, but if you use something else let me know and we can see if we can find a solution. Last, when you run the command you're likely to get the following message: NOTE: Specification.all called from /Users/brian/.rvm/gems/ruby-2.2.0/gems/neo4apis-0.5.1/bin/neo4apis:8:in `<top (required)>' I'm not sure why that's outputting, but I'll look at it later. Thanks! Let me know if that worked for you Thanks for looking at it. The specification warning note comes from here: https://github.com/rubygems/rubygems/blob/master/lib/rubygems/specification.rb#L820 When I run the command now I get the warning message followed by this: WARN: Unresolved specs during Gem::Specification.reset: minitest (~> 5.1) WARN: Clearing out unresolved specs. Please report a bug if this causes problems. Importing tables: [table list] /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday_middleware-0.9.1/lib/faraday_middleware.rb:25:in block in <module:FaradayMiddleware>': cannot load such file -- faraday_middleware/request/encode_json (LoadError) from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:215:in call' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:215:in block in load_middleware' from /Users/harry/.rvm/rubies/ruby-2.2.0/lib/ruby/2.2.0/monitor.rb:211:in mon_synchronize' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:197:in middleware_mutex' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:214:in load_middleware' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:188:in lookup_middleware' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday/rack_builder.rb:203:in use_symbol' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday/rack_builder.rb:92:in request' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j-server/cypher_session.rb:30:in block in create_connection' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday/connection.rb:91:in initialize' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:70:in new' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/faraday-0.9.1/lib/faraday.rb:70:in new' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j-server/cypher_session.rb:28:in create_connection' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j-server/cypher_session.rb:48:in open' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j-server/cypher_session.rb:5:in block in module:Server' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j/session.rb:111:in call' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j/session.rb:111:in create_session' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4j-core-4.0.1/lib/neo4j/session.rb:98:in open' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:94:in neo4apis_client' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:66:in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:52:in all_models' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:115:in invoke' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:235:in block in subcommand' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/base.rb:440:in start' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/bin/neo4apis:14:in <top (required)>' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in load' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in ' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in eval' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in ' Interesting, it looks like you ran the all_models command. Is this in an existing ruby app with ActiveRecord models? What was the full command that you ran? Correct. neo4apis activerecord all_models --import-all-associations Oh, right, forgot to ask: What port is your neo4j server running on? If it's something other than 7474 than you need to use the --neo4j-url argument. I doubt that's it though... @subvertallchris Could you have a look since you've worked with Faraday more than me (and it's way past my bedtime ;)? The last line in our world is this: https://github.com/neo4jrb/neo4j-core/blob/master/lib/neo4j-server/cypher_session.rb#L30 If it helps, here's the Gemfile.lock GEM remote: https://rubygems.org/ specs: actionmailer (4.2.0) actionpack (= 4.2.0) actionview (= 4.2.0) activejob (= 4.2.0) mail (~> 2.5, >= 2.5.4) rails-dom-testing (~> 1.0, >= 1.0.5) actionpack (4.2.0) actionview (= 4.2.0) activesupport (= 4.2.0) rack (~> 1.6.0) rack-test (~> 0.6.2) rails-dom-testing (~> 1.0, >= 1.0.5) rails-html-sanitizer (~> 1.0, >= 1.0.1) actionview (4.2.0) activesupport (= 4.2.0) builder (~> 3.1) erubis (~> 2.7.0) rails-dom-testing (~> 1.0, >= 1.0.5) rails-html-sanitizer (~> 1.0, >= 1.0.1) activejob (4.2.0) activesupport (= 4.2.0) globalid (>= 0.3.0) activemodel (4.2.0) activesupport (= 4.2.0) builder (~> 3.1) activerecord (4.2.0) activemodel (= 4.2.0) activesupport (= 4.2.0) arel (~> 6.0) activesupport (4.2.0) i18n (~> 0.7) json (~> 1.7, >= 1.7.7) minitest (~> 5.1) thread_safe (~> 0.3, >= 0.3.4) tzinfo (~> 1.1) arel (6.0.0) awesome_print (1.6.1) binding_of_caller (0.7.2) debug_inspector (>= 0.0.1) builder (3.2.2) byebug (3.5.1) columnize (~> 0.8) debugger-linecache (~> 1.2) slop (~> 3.6) celluloid (0.16.0) timers (~> 4.0.0) choice (0.1.7) coffee-rails (4.1.0) coffee-script (>= 2.2.0) railties (>= 4.0.0, < 5.0) coffee-script (2.3.0) coffee-script-source execjs coffee-script-source (1.8.0) columnize (0.9.0) connection_pool (2.1.1) debug_inspector (0.0.2) debugger-linecache (1.2.0) erubis (2.7.0) execjs (2.2.2) faker (1.4.3) i18n (~> 0.5) font-awesome-rails (4.3.0.0) railties (>= 3.2, < 5.0) globalid (0.3.0) activesupport (>= 4.1.0) guard-compat (1.2.1) guard-minitest (2.4.2) guard-compat (~> 1.2) minitest (>= 3.0) haml (4.0.6) tilt hike (1.2.3) hitimes (1.2.2) i18n (0.7.0) jbuilder (2.2.6) activesupport (>= 3.0.0, < 5) multi_json (~> 1.2) jquery-rails (4.0.3) rails-dom-testing (~> 1.0) railties (>= 4.2.0) thor (>= 0.14, < 2.0) jquery-ui-rails (5.0.3) railties (>= 3.2.16) json (1.8.2) kaminari (0.16.2) actionpack (>= 3.0.0) activesupport (>= 3.0.0) loofah (2.0.1) nokogiri (>= 1.5.9) mail (2.6.3) mime-types (>= 1.16, < 3) mime-types (2.4.3) mini_portile (0.6.2) minitest (5.5.1) multi_json (1.10.1) nested_form (0.3.2) newrelic_rpm (3.9.9.275) nokogiri (1.6.6.2) mini_portile (~> 0.6.0) pg (0.18.1) rack (1.6.0) rack-pjax (0.8.0) nokogiri (~> 1.5) rack (~> 1.1) rack-protection (1.5.3) rack rack-test (0.6.3) rack (>= 1.0) rails (4.2.0) actionmailer (= 4.2.0) actionpack (= 4.2.0) actionview (= 4.2.0) activejob (= 4.2.0) activemodel (= 4.2.0) activerecord (= 4.2.0) activesupport (= 4.2.0) bundler (>= 1.3.0, < 2.0) railties (= 4.2.0) sprockets-rails rails-deprecated_sanitizer (1.0.3) activesupport (>= 4.2.0.alpha) rails-dom-testing (1.0.5) activesupport (>= 4.2.0.beta, < 5.0) nokogiri (~> 1.6.0) rails-deprecated_sanitizer (>= 1.0.1) rails-erd (1.1.0) activerecord (>= 3.0) activesupport (>= 3.0) choice (~> 0.1.6) ruby-graphviz (~> 1.0.4) rails-html-sanitizer (1.0.1) loofah (~> 2.0) rails_admin (0.6.6) builder (~> 3.1) coffee-rails (~> 4.0) font-awesome-rails (>= 3.0, < 5) haml (~> 4.0) jquery-rails (>= 3.0, < 5) jquery-ui-rails (~> 5.0) kaminari (~> 0.14) nested_form (~> 0.3) rack-pjax (~> 0.7) rails (~> 4.0) remotipart (~> 1.0) safe_yaml (~> 1.0) sass-rails (>= 4.0, < 6) railties (4.2.0) actionpack (= 4.2.0) activesupport (= 4.2.0) rake (>= 0.8.7) thor (>= 0.18.1, < 2.0) rake (10.4.2) rdoc (4.2.0) redis (3.2.0) redis-namespace (1.5.1) redis (~> 3.0, >= 3.0.4) remotipart (1.2.1) ruby-graphviz (1.0.9) safe_yaml (1.0.4) sass (3.4.10) sass-rails (5.0.1) railties (>= 4.0.0, < 5.0) sass (~> 3.1) sprockets (>= 2.8, < 4.0) sprockets-rails (>= 2.0, < 4.0) tilt (~> 1.1) sdoc (0.4.1) json (~> 1.7, >= 1.7.7) rdoc (~> 4.0) sidekiq (3.3.1) celluloid (>= 0.16.0) connection_pool (>= 2.1.1) json redis (>= 3.0.6) redis-namespace (>= 1.3.1) sinatra (1.4.5) rack (~> 1.4) rack-protection (~> 1.4) tilt (~> 1.3, >= 1.3.4) slop (3.6.0) spring (1.2.0) sprockets (2.12.3) hike (~> 1.2) multi_json (~> 1.0) rack (~> 1.0) tilt (~> 1.1, != 1.3.0) sprockets-rails (2.2.4) actionpack (>= 3.0) activesupport (>= 3.0) sprockets (>= 2.8, < 4.0) thor (0.19.1) thread_safe (0.3.4) tilt (1.4.1) timers (4.0.1) hitimes turbolinks (2.5.3) coffee-rails tzinfo (1.2.2) thread_safe (~> 0.1) uglifier (2.7.0) execjs (>= 0.3.0) json (>= 1.8.0) web-console (2.0.0) activemodel (~> 4.0) binding_of_caller (>= 0.7.2) railties (~> 4.0) sprockets-rails (>= 2.0, < 4.0) PLATFORMS ruby DEPENDENCIES awesome_print byebug faker guard-minitest jbuilder jquery-rails newrelic_rpm pg rails (= 4.2.0) rails-erd rails_admin sass-rails sdoc sidekiq sinatra (>= 1.3.0) spring turbolinks uglifier web-console I use the default 7474 as the neo4j port. I'm looking, but I can't quite figure it out. It seems to be something in the faraday_middleware gem, but as far as I can see it should work fine. I'm also testing it and it's working for me. How many tables were in that table list, BTW? Did you add neo4apis-activerecord to your Gemfile or did you use gem install neo4apis-activerecord? There are about 20 tables. I installed the gem via RubyGems. Trying again with the gem in my Gemfile, I get a new error. NOTE: Specification.all called from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/bin/neo4apis:8:in `<top (required)>' Importing tables: [table list] /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/association.rb:107:in `association_key_type': undefined method `type_for_attribute' for :part:Symbol (NoMethodError) from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/association.rb:103:in `key_conversion_required?' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/association.rb:60:in `owners_by_key' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/association.rb:78:in `associated_records_by_owner' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/collection_association.rb:13:in `preload' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader/association.rb:20:in `run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:146:in `block (2 levels) in preloaders_for_one' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:144:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:144:in `map' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:144:in `block in preloaders_for_one' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:143:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:143:in `flat_map' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:143:in `preloaders_for_one' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:115:in `preloaders_on' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:103:in `block in preload' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:102:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:102:in `flat_map' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/associations/preloader.rb:102:in `preload' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation.rb:644:in `block in exec_queries' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation.rb:643:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation.rb:643:in `exec_queries' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation.rb:514:in `load' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation.rb:243:in `to_a' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:117:in `find_in_batches' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:50:in `find_each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:76:in `block (2 levels) in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:67:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:67:in `block in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/lib/neo4apis/base.rb:54:in `batch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:66:in `import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:52:in `all_models' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in `run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in `invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in `dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:115:in `invoke' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:235:in `block in subcommand' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in `run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in `invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in `dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/base.rb:440:in `start' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/bin/neo4apis:14:in `<top (required)>' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in `load' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in `<main>' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in `eval' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in `<main>' I also tried importing a single model which references another model. Here are the migrations: class CreateCompanies < ActiveRecord::Migration def change create_table :companies do |t| t.string :name t.timestamps end end end class CreateDealers < ActiveRecord::Migration def change create_table :dealers do |t| t.references :company, index: true t.integer :rank, default: 1 t.timestamps end end end And then I ran this command from the terminal: neo4apis activerecord models Dealer --import-all-associations The result was this: NOTE: Specification.all called from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/bin/neo4apis:8:in `<top (required)>' Importing tables: dealers /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/lib/neo4apis/base.rb:35:in `add_node': No UUID specified for label `Company` (ArgumentError) from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/activerecord.rb:49:in `add_model_node' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/activerecord.rb:37:in `add_model_relationship' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/activerecord.rb:22:in `block (2 levels) in model_importer' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/activerecord.rb:17:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/activerecord.rb:17:in `block in model_importer' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/lib/neo4apis/base.rb:62:in `instance_exec' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/lib/neo4apis/base.rb:62:in `import' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:77:in `block (3 levels) in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:51:in `block (2 levels) in find_each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:51:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:51:in `block in find_each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:124:in `find_in_batches' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/activerecord-4.2.0/lib/active_record/relation/batches.rb:50:in `find_each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:76:in `block (2 levels) in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:67:in `each' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:67:in `block in import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/lib/neo4apis/base.rb:54:in `batch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:66:in `import_models_or_tables' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-activerecord-0.6.1/lib/neo4apis/cli/activerecord.rb:36:in `models' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in `run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in `invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in `dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:115:in `invoke' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:235:in `block in subcommand' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/command.rb:27:in `run' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/invocation.rb:126:in `invoke_command' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor.rb:359:in `dispatch' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/thor-0.19.1/lib/thor/base.rb:440:in `start' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/gems/neo4apis-0.5.1/bin/neo4apis:14:in `<top (required)>' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in `load' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/neo4apis:23:in `<main>' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in `eval' from /Users/harry/.rvm/gems/ruby-2.2.0@dealerdbsql/bin/ruby_executable_hooks:15:in `<main>' First off, thanks for your help! This is really useful. There errors are looking more tractable! I'm looking deeper, but I have a couple of questions first: Can you share the model (or a stripped down version) of the one from your app when you just tried to import one? For the second thing that you tried, do you also have ActiveRecord models for the dealers/companies tables? If not, then neo4apis-activerecord is probably trying to generate models, but it's not generating one for Company since you didn't specify it. I actually need to sit down and fix that. Right now on the command like if you specify models or tables, it actually (mostly) does the same thing and lets you specify either models or tables and it uses the ActiveSupport#Inflectors module to get what it needs. I should be more strict about that because if you haven't defined a model is your ruby application, it shouldn't try to generate one. Is parts one of your tables? I'm trying to figure out why it seems to think that @klass is equal to :part on this line: https://github.com/rails/rails/blob/master/activerecord/lib/active_record/associations/preloader/association.rb#L107 I don't think I'll be able to fix it tonight and I'll be busy this weekend, but I'll take another look again soon. Thanks! Yes, parts is a table in the original import. Except for one join table, every table has an associated AR model defined in the usual Rails way. Also, I appreciate your help with this, but please don't drop everything to fix my problem. I have other ways of getting the data into Neo4j and was really just interested in seeing what the results were when I tried your gem. That said, I'm happy to continue debugging if you'd like to explore some more. Ah! Sorry, I didn't see when you tagged me cause it was the day I was out and about. The faraday_middleware thing was cleared up by modifying your gemfile, right? Ok, back to this, and sorry for the lack of response ;) I've made a number of improvements, most of which probably won't help. But I've also added a --debug flag, so could you try again with that flag set and post the output if it doesn't work? Thanks! Using version 0.6.0. Running neo4apis activerecord models Dealer --import-all-associations Error: Could not find command "activerecord".
gharchive/issue
2015-01-28T21:45:55
2025-04-01T06:45:06.957563
{ "authors": [ "cheerfulstoic", "harrylove", "subvertallchris" ], "repo": "neo4jrb/neo4apis-activerecord", "url": "https://github.com/neo4jrb/neo4apis-activerecord/issues/9", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
53649898
adds unique: true, creates_unique_rel to ActiveNode and ActiveRel Resolves #473 Usage: class Student include Neo4j::ActiveNode has_many :out, :lessons, type: 'ENROLLED_IN', unique: true end # or for ActiveRel class Student include Neo4j::ActiveNode has_many :out, :lessons, rel_class: 'ENROLLED_IN' end class EnrolledIn from_node Student to_node Lesson creates_unique_rel end Both student.lessons << lesson and EnrolledIn.create(from_node: student, to_node: lesson) will now use CREATE UNIQUE instead of CREATE in the Cypher statement. Coverage increased (+0.01%) when pulling 6415ad4ad75430ae4eafa22298c9605d9e006a25 on create_unique into 616ded4453579856191229fabb368f2df1541448 on master. My favorite kind of comments: none! IT SHALL BE MERGED!
gharchive/pull-request
2015-01-07T16:20:49
2025-04-01T06:45:06.961943
{ "authors": [ "coveralls", "subvertallchris" ], "repo": "neo4jrb/neo4j", "url": "https://github.com/neo4jrb/neo4j/pull/660", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
96233399
Support outputting pretty cypher queries With pretty_logged_cypher_queries config option @subvertallchris Could you try this out when you get a chance? You should just need to use the pretty_cypher branches for both neo4j and neo4j-core as well as setting config.neo4j.pretty_logged_cypher_queries = true You might also need to set config.neo4j.log_cypher_queries = true
gharchive/pull-request
2015-07-21T05:44:17
2025-04-01T06:45:06.963723
{ "authors": [ "cheerfulstoic" ], "repo": "neo4jrb/neo4j", "url": "https://github.com/neo4jrb/neo4j/pull/878", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
101375208
More build errors? These seem to be directly related to Issue #14. I'm running Xubuntu in a VM, with the package libboost-all-dev installed, which appeared to be good enough for CMake to detect it being installed. Simply did cmake . and make, and I get this. What could be wrong here? Thanks. What version of boost are you running? There was a bug in boost for a while due to which this swap thing didn't compile. I think it wasn't fixed until 1.58 or so. As a workaround, try one of the following: add #include <boost/fusion/include/swap.hpp> to the top of src/parser_assembly/flowcontrol.cpp remove the lines ``` void swap(nihstro::Condition& a, nihstro::Condition& b) { boost::fusion::swap(a, b); } Your workaround did the trick! According to UbuntuUpdates.org, it seems I've got 1.55.0.2 (while running Xubuntu 15.04 to be exact). Which of the two exactly made it work? :p I did both and it did the trick. Seems when I read your reply at 3 in the morning I neglected to read the words "one of the following"... hehe, alright then. I'll add a wiki page or something to mention this workaround, I guess. Glad you got it working, though ;)
gharchive/issue
2015-08-17T09:11:52
2025-04-01T06:45:06.968400
{ "authors": [ "TheGag96", "neobrain" ], "repo": "neobrain/nihstro", "url": "https://github.com/neobrain/nihstro/issues/29", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
269439128
chore(package): update mocha to version 4.0.1 Closes #53 Codecov Report Merging #57 into master will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #57 +/- ## ===================================== Coverage 100% 100% ===================================== Files 3 3 Lines 81 81 Branches 14 14 ===================================== Hits 81 81 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 5758cf6...9b62128. Read the comment docs.
gharchive/pull-request
2017-10-29T21:45:05
2025-04-01T06:45:06.975958
{ "authors": [ "codecov-io", "neogeek" ], "repo": "neogeek/doxdox", "url": "https://github.com/neogeek/doxdox/pull/57", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
170492058
New Rust error message format Hello. Rust nightly now shows errors in a new format. It would be nice to support it in neomake (if the new format is fixed). Thanks. The new format is now in Rust stable. So, I believe it we should switch to it now. I currently use this errorformat, but it does not always work: let g:neomake_rust_clippy_maker = { \ 'exe': 'cargo', \ 'args': ['clippy'], \ 'errorformat': \ '%-Gerror: Could not compile%.%#,' \. '%-Gerror: aborting due to%.%#,' \. '%Eerror[%.%#]: %m,%C\\s%#--> %f:%l:%c,' \. '%Eerror: %m,%C\\s%#--> %f:%l:%c,' \. '%Wwarning: %m,%C\\s%#--> %f:%l:%c,' \. '%Ihelp: %m,' \. '%I\\s%#= help: %m,' \. '%I\\s%#= note: %m,' \, \ 'append_file': 0, \ } Check https://github.com/neomake/neomake/pull/613 I guess?! Closing this in favor of #613. @antoyo Please provide feedback there, and then we can get it finalized/merged.
gharchive/issue
2016-08-10T18:57:15
2025-04-01T06:45:06.983072
{ "authors": [ "antoyo", "blueyed" ], "repo": "neomake/neomake", "url": "https://github.com/neomake/neomake/issues/549", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1402312612
ArgumentOutOfRangeException when adding collection in Unity 2022.2 I get this error when adding a collection in Unity 2022.2 Seems to be Unity requiring a VisuelElement by index internally. The point of conflict comes from inside of CollectionEditor.AlternateComponentsGUI when the editor child is removed.
gharchive/issue
2022-10-09T15:41:09
2025-04-01T06:45:07.000815
{ "authors": [ "alexanderameye", "nomnomab" ], "repo": "neon-age/Smart-Hierarchy", "url": "https://github.com/neon-age/Smart-Hierarchy/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2123027313
tests: ignore pg_xact files in comparison Problem This test has been flaky for years. It's a very important correctness test, so we need to get it off the list of flaky tests to avoid ignoring other more severe failures. It's not obvious to me exactly why it's failing, but this is a binary-level difference in postgres structures that I'm not qualified to diagnose: the test is hoping that CHECKPOINT will guarantee that everything in pg_xact hits disk, but this is either not the case, or there is some subtle underlying issue that can result in different file content from the same stream of writes. Closes https://github.com/neondatabase/neon/issues/559 Summary of changes Exclude pg_xact files from comparison in check_restored_datadir_content Checklist before requesting a review [ ] I have performed a self-review of my code. [ ] If it is a core feature, I have added thorough tests. [ ] Do we need to implement analytics? if so did you add the relevant metrics to the dashboard? [ ] If this PR requires public announcement, mark it with /release-notes label and add several sentences in this section. Checklist before merging [ ] Do not forget to reformat commit message to not include the above checklist I'm pretty reluctant about this. If the test cannot validate pg_xact anymore, it reduces the coverage of the test significantly. I created https://github.com/neondatabase/neon/pull/6666, as yet another attempt fix the discrepancy. I was not able to reproduce the flakiness locally, so it's just based on a hunch that something happens in the database after the last checkpoint that isn't reflected in the basebackup yet. I think even if it doesn't fix the flakiness, it's a more robust approach anyway. If it doesn't fix the flakiness, let's revisit this. https://github.com/neondatabase/neon/pull/6666 is a plausible fix, so I'll close this and we can monitor the test to see if it stabilizes.
gharchive/pull-request
2024-02-07T13:21:44
2025-04-01T06:45:07.008522
{ "authors": [ "hlinnaka", "jcsp" ], "repo": "neondatabase/neon", "url": "https://github.com/neondatabase/neon/pull/6662", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
720582964
Pyls not reporting errors (or almost any diagnostics) nvim --version: NVIM v0.5.0-751-gd198aa511 Build type: RelWithDebInfo LuaJIT 2.1.0-beta3 Compilation: /usr/bin/gcc-5 -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -O2 -g -Og -g -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -DINCLUDE_GENERATED_DECLARATIONS -D_GNU_SOURCE -DNVIM_MSGPACK_HAS_FLOAT32 -DNVIM_UNIBI_HAS_VAR_FROM -DMIN_LOG_LEVEL=3 -I/home/travis/build/neovim/bot-ci/build/neovim/build/config -I/home/travis/build/neovim/bot-ci/build/neovim/src -I/home/travis/build/neovim/bot-ci/build/neovim/.deps/usr/include -I/usr/include -I/home/travis/build/neovim/bot-ci/build/neovim/build/src/nvim/auto -I/home/travis/build/neovim/bot-ci/build/neovim/build/include Compiled by travis@travis-job-7fcf59ed-01e6-4dcf-a363-4c9cb0163aca Features: +acl +iconv +tui See ":help feature-compile" system vimrc file: "$VIM/sysinit.vim" fall-back for $VIM: "/share/nvim" Run :checkhealth for more info :checkhealth result health#nvim_lsp#check ======================================================================== ## Checking language server protocol configuration - INFO: pyls: configuration checked. - INFO: rust_analyzer: configuration checked. health#nvim#check ======================================================================== ## Configuration - OK: no issues found ## Performance - OK: Build type: RelWithDebInfo ## Remote Plugins - OK: Up to date ## terminal - INFO: key_backspace (kbs) terminfo entry: key_backspace=^H - INFO: key_dc (kdch1) terminfo entry: key_dc=\E[3~ - INFO: $COLORTERM='truecolor' nvim-lsp version(commit hash): 60133c47e0fd82556d7ca092546ebfa8d047466e What language server: pyls 0.34.1-2 Operating system/version: Arch Linux Terminal name/version: alacirtty 0.5.0 $TERM: xterm-256color How to reproduce the problem from neovim startup init.vim --------- call plug#begin($HOME.'/.config/nvim/autoload/plugged') Plug 'neovim/nvim-lspconfig' Plug 'nvim-lua/diagnostic-nvim' call plug#end() lua <<EOF require'nvim_lsp'.pyls.setup{on_attach=require'diagnostic'.on_attach} require'nvim_lsp'.rust_analyzer.setup{on_attach=require'diagnostic'.on_attach} EOF let g:diagnostic_enable_virtual_text = 1 let g:space_before_virtual_text = 20 Create a random python file and write some gibberish: Actual behaviour As you can see, there are no syntax errors, mismatched brackets, "no member" errors etc. In fact I have been unable to force it to show anything other than whitespace warnings (like "trailing whitespace", "expected indentation block" or "expected 2 empty lines and found 1" ). Expected behaviour There should be an error in pretty much every single line above. Compare it to what happens with rust_analyzer: Now, if I add the lsp-status plugin and run :echo luaeval("require('lsp-status').diagnostics()") I also only get the 1 warning in that ridiculous python file, which (combined with the fact that rust works just fine) leads me to belive the problem is in the language server itself. Any advice on how to fix this? I suspect it's because of how you installed the server. Note that pyls leans on pylint, pyflakes, pycodestyle and pydocstyle for diagnostics. But these aren't installed by default. Hey, thank you so much for the quick reply. You are correct, after installing flake8 the code is linted the way it should. I had pylint installed already and thought that it would be sufficient by itself. Thanks again!
gharchive/issue
2020-10-13T17:31:42
2025-04-01T06:45:07.119227
{ "authors": [ "lithammer", "piotrmachura16" ], "repo": "neovim/nvim-lspconfig", "url": "https://github.com/neovim/nvim-lspconfig/issues/382", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2168143604
Table rows without Roman numerals / columns without letters E.g. Ch. 13 p. 1099 [1] (Receiving # from add_page_numbers.py but not making it into YAML) Ch. 3 p. 209 Ch. 3 p. 210 Ch. 4 p. 251 Ch. 4 p. 258 (note lexical row headers): Ch. 4 p. 261 [24], [25], [26] Ch. 5 p. 338 A tricky one: postTags describing relation between column entries ch. 3 p. 176 ch. 5 p. 455 (also [16] on p. 456): feature-value matrix where the sentences are row headers. Currently the column-headers are omitted (maybe that's OK).
gharchive/issue
2024-03-05T02:23:12
2025-04-01T06:45:07.183474
{ "authors": [ "nschneid" ], "repo": "nert-nlp/cgel", "url": "https://github.com/nert-nlp/cgel/issues/106", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
548437103
Rc/v0.27.0 0.27.0 (2020-01-11) Note: Version bump only for package ckb-sdk-js Codecov Report Merging #418 into develop will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## develop #418 +/- ## ======================================== Coverage 90.13% 90.13% ======================================== Files 26 26 Lines 1145 1145 Branches 268 268 ======================================== Hits 1032 1032 Misses 112 112 Partials 1 1
gharchive/pull-request
2020-01-11T14:36:18
2025-04-01T06:45:07.189939
{ "authors": [ "Keith-CY", "codecov-io" ], "repo": "nervosnetwork/ckb-sdk-js", "url": "https://github.com/nervosnetwork/ckb-sdk-js/pull/418", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
440524546
BinaryHash/CodeHash px prefix Does it has 0x prefix ? Originally posted by @classicalliu in https://github.com/nervosnetwork/ckb-sdk-swift/pull/53 RPC 接口是全部要求 0x 开头的,十六进制的字符串如果没有这个 prefix 会报错 Things to check: [ ] codeHash [ ] H256.zeroHash [ ] Script.alwaysSuccessHash RPC 接口是全部要求 0x 开头的,十六进制的字符串如果没有这个 prefix 会报错 Thanks for the info. Then we should always prefix them.
gharchive/issue
2019-05-06T03:02:19
2025-04-01T06:45:07.193041
{ "authors": [ "ashchan", "classicalliu" ], "repo": "nervosnetwork/ckb-sdk-swift", "url": "https://github.com/nervosnetwork/ckb-sdk-swift/issues/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2345859828
Error analysis pipeline Fixes #27 Description In this PR: A little refactor of some of Aidan's code preweight_tenure_error_analysis.py Run error analysis on all LSOAs for all features, before and after weighting Makes plots from the results Fixes # (issue) Instructions for Reviewer Don't worry about asf_heat_pump_suitability/getters/s3_getters.py - it is all copied from previous projects! Aidan's already reviewed code has been split into asf_heat_pump_suitability/pipeline/error_analysis/error_analysis_utils.py and asf_heat_pump_suitability/pipeline/error_analysis/sample/preweight_tenure_error_analysis.py. the only thing that needs review in these is the get_error_reduction function in error_analysis_utils Please check for bugs etc in asf_heat_pump_suitability/pipeline/reweight_epc/evaluate_reweighting.py Try running: python asf_heat_pump_suitability/pipeline/reweight_epc/evaluate_reweighting.py --sample Checklist: [ ] I have refactored my code out from notebooks/ [ ] I have checked the code runs [ ] I have tested the code [ ] I have run pre-commit and addressed any issues not automatically fixed [ ] I have merged any new changes from dev [ ] I have documented the code [ ] Major functions have docstrings [ ] Appropriate information has been added to READMEs [ ] I have explained this PR above [ ] I have requested a code review Thanks so much @crispy-wonton - your comments were extremely useful ! I've addressed your comments in this commit https://github.com/nestauk/asf_heat_pump_suitability/pull/28/commits/e75407168e71d3633c3ac0048929fed003571be0 I haven't rerun the run_script.py with the dummy rows included. I will try tomorrow and then rerun the evaluation. @crispy-wonton update: I have reran run_script.py with the outer join. There are 16788545 rows in this output. When the data is loaded we get this though: Which isn't what I expected (that row where UPRN= 10024356313.0 but UPRN_right = null?). I'm not quite sure if this matters, but it may represent a problem? latest evaluation is also run with changes to the script as in this PR + inclusion of dummy rows. New results are better for property type than before. The reweighting produces better Tenure proportions MAE 92.53% of the time The reweighting produces the same Tenure proportions MAE 0.1% of the time The reweighting produces better Property Type proportions MAE 93.5% of the time The reweighting produces the same Property Type proportions MAE 0.17% of the time The reweighting produces better Build Year proportions MAE 86.19% of the time The reweighting produces the same Build Year proportions MAE 1.55% of the time @crispy-wonton I just saw your comment - although I don't think you were concerned. So am I ok to merge this? @crispy-wonton I just saw your comment - although I don't think you were concerned. So am I ok to merge this? Yes I don't think it's a problem because I think it's an expected behaviour of the pipeline. I believe UPRN comes from the original EPC dataset and UPRN_right comes from the reweighted EPC rows. So any rows that were excluded from the reweighting because of our reweighting criteria would not be present in the reweighted EPC dataset, therefore they will appear as null UPRNs in the outer join. I think this is good to merge :)
gharchive/pull-request
2024-06-11T09:16:00
2025-04-01T06:45:07.228352
{ "authors": [ "crispy-wonton", "lizgzil" ], "repo": "nestauk/asf_heat_pump_suitability", "url": "https://github.com/nestauk/asf_heat_pump_suitability/pull/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1098717103
Merge queued tests and fix AddAppFromType Breaking change Proposed change Type of change [ ] Dependency upgrade [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (which adds functionality to an existing integration) [ ] Breaking change (fix/feature causing existing functionality to break) [ ] Code quality improvements to existing code or addition of tests Additional information This PR fixes or closes issue: fixes # This PR is related to issue: Link to documentation pull request: Checklist [ ] The code change is tested and works locally. [ ] Local tests pass. Your PR cannot be merged unless tests pass [ ] There is no commented out code in this PR. [ ] I have followed the [development checklist][dev-checklist] [ ] The code compiles without warnings (code quality chek) [ ] Tests have been added to verify that the new code works. If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated for www.home-assistant.io Codecov Report Merging #592 (8f28f58) into app_model (f697c36) will increase coverage by 0.90%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## app_model #592 +/- ## ============================================= + Coverage 85.06% 85.96% +0.90% ============================================= Files 106 111 +5 Lines 2805 3021 +216 Branches 305 318 +13 ============================================= + Hits 2386 2597 +211 + Misses 288 285 -3 - Partials 131 139 +8 Flag Coverage Δ unittests 85.96% <100.00%> (+0.90%) :arrow_up: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...el/Common/Extensions/ServiceCollectionExtension.cs 100.00% <100.00%> (ø) ...ime/NetDaemon.Runtime/Internal/NetDaemonRuntime.cs 85.00% <100.00%> (ø) ...mon.HassClient/Internal/Net/WebSocketClientImpl.cs 68.42% <0.00%> (-7.90%) :arrow_down: ...sClient/Internal/Net/WebSocketTransportPipeline.cs 90.43% <0.00%> (-0.87%) :arrow_down: ...Runtime/Common/Extensions/HostBuilderExtensions.cs 100.00% <0.00%> (ø) ...ntime/NetDaemon.Runtime/Internal/RuntimeService.cs 100.00% <0.00%> (ø) ...time/Common/Extensions/ServiceBuilderExtensions.cs 100.00% <0.00%> (ø) ...time/NetDaemon.Runtime/Internal/AppStateManager.cs 94.80% <0.00%> (ø) ...mon.HassClient/Internal/HomeAssistantConnection.cs 84.76% <0.00%> (+0.95%) :arrow_up: ... and 3 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update f697c36...8f28f58. Read the comment docs.
gharchive/pull-request
2022-01-11T06:15:18
2025-04-01T06:45:07.353719
{ "authors": [ "codecov-commenter", "helto4real" ], "repo": "net-daemon/netdaemon", "url": "https://github.com/net-daemon/netdaemon/pull/592", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1260201477
No URL provided I’m trying to download a YouTube playlist, opened in Shortcuts using a link in the pasteboard to follow the steps. Everything seems to operate as expected, however, after finishing up in the shell it returns back to Shortcuts and it shows ‘No URL provided’ Am I supposed to create a folder in Files app where the Shortcut writes to? I can’t seem to find that in the source code I’m reading in the Shortcuts app. You don't need to create any folder in Files. Your playlist videos are supposed to be saved on a special directory $SHORTCUTS When the process sends you back to Shortcuts it runs the shortcut with passing the path of your downloaded files. That's how the shortcut it's supposed to find your files. Seems like you are getting my error message for no input for some reason. Can you share which playlist you tried to download? was it audio or video? It was a YouTube playlist, audio only. I reinstalled the shortcut and managed to get the zip file in preview after returning to shortcuts. Great Shortcut you wrote and nice to have an introduction to a-Shell Mini, hadn’t heard about it before.
gharchive/issue
2022-06-03T18:05:20
2025-04-01T06:45:07.357357
{ "authors": [ "net00-1", "reinierrr" ], "repo": "net00-1/SW-DLT", "url": "https://github.com/net00-1/SW-DLT/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2714584340
TASK Build VRFs Attach List From Template fails Ansible Version ansible [core 2.17.0] Ansible Collection Versions Collection Version ---------------------------------------- ------- ansible.netcommon 5.3.0 ansible.posix 1.5.4 ansible.utils 2.11.0 cisco.dcnm 3.5.0 cisco.nac_dc_vxlan 0.3.0 cisco.nxos 5.3.0 community.general 8.5.0 Cisco NDFC Version 12.2.2 Cisco NX-OS Version 10.3(6) Which role is this issue related to? cisco.nac_dc_vxlan.validate Which section of the data model is this issue related to? vxlan.overlay_services Expected Behavior The extended service model data for the VRF attach groups should contain an 'mgmt_ip_address'. Actual Behavior ansible-playbook -i inventory.yaml deploy.yaml PLAY [marehler_vnd3] ******************************************************************************************************************************** <...> TASK [cisco.nac_dc_vxlan.common : Build VRFs Attach List From Template] ***************************************************************************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: ansible.errors.AnsibleUndefinedVariable: 'dict object' has no attribute 'mgmt_ip_address' fatal: [marehler_vnd3 -> localhost]: FAILED! => {"changed": false, "msg": "AnsibleUndefinedVariable: 'dict object' has no attribute 'mgmt_ip_address'"} Ansible Playbook --- # This is the main entry point playbook for calling the various roles in this collection. - hosts: marehler_vnd3 any_errors_fatal: true gather_facts: no roles: # ----------------------- # DataCenter Roles # Role: cisco.netascode_dc_vxlan.dtc manages NDFC workflows # - role: cisco.nac_dc_vxlan.dtc.create - role: cisco.nac_dc_vxlan.dtc.deploy - role: cisco.nac_dc_vxlan.dtc.remove Root Cause I think the issue is caused in the Validate role in Prepare Service Model. In prep_104_fabric_overlay_services.py, for the VRF attach groups, the switch hostname is replaced with the management IP address. However, ndfc_attach_vrfs.j2 used in task Build VRFs Attach List From Template expects an 'mgmt_ip_address' variable. for switch in model_data['vxlan']['overlay_services']['vrf_attach_groups_dict'][grp['name']]: if any(sw['name'] == switch['hostname'] for sw in switches): found_switch = next((item for item in switches if item["name"] == switch['hostname'])) if found_switch.get('management').get('management_ipv4_address'): switch['hostname'] = found_switch['management']['management_ipv4_address'] elif found_switch.get('management').get('management_ipv6_address'): switch['hostname'] = found_switch['management']['management_ipv6_address'] For the networks attach groups, such a 'mgmt_ip_address' is added instead of changing the hostname. for switch in model_data['vxlan']['overlay_services']['network_attach_groups_dict'][grp['name']]: if any(sw['name'] == switch['hostname'] for sw in switches): found_switch = next((item for item in switches if item["name"] == switch['hostname'])) if found_switch.get('management').get('management_ipv4_address'): switch['mgmt_ip_address'] = found_switch['management']['management_ipv4_address'] elif found_switch.get('management').get('management_ipv6_address'): switch['mgmt_ip_address'] = found_switch['management']['management_ipv6_address'] Relevant Debug Output No response Hi @marehler it looks like you have got yourself on a mixture of code versions here, from your versions above it looks like you are running on the 0.3.0 release but the changes to add mgmt_ip_address were made after that release so im guessing you have pulled certain files from the develop branch and hence now have the mismatch? if you either roll everything back to 0.3.0 or run everything from the current develop branch you shouldn't see this issue. Versioning issue, resolved as described above
gharchive/issue
2024-12-03T10:17:21
2025-04-01T06:45:07.364616
{ "authors": [ "marehler", "mthurstocisco" ], "repo": "netascode/ansible-dc-vxlan", "url": "https://github.com/netascode/ansible-dc-vxlan/issues/237", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1692782018
Blog on interrupts, softirqs and softnets in Linux Please review and Approve @andrewm4894 / @cakrit : Not sure how the context-switching blog came up here. I created a new branch after pushing the other one - weird. Closing the PR and recreating another one.. @andrewm4894 : I fixed the files being committed.
gharchive/pull-request
2023-05-02T17:04:32
2025-04-01T06:45:07.429166
{ "authors": [ "sashwathn" ], "repo": "netdata/blog", "url": "https://github.com/netdata/blog/pull/256", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
332225541
5.5尺寸 输入框bug 在5.5尺寸的设备上,在输入框输入大量表情, 在textview的高度大道最大值时,继续输入, 换行出现问题.仍能够输入文字呵呵表情,但不会自动换行,仍然显示之前的文字. 需要手动滑到最底部,接着输入,换行正常,显示正常; 目前只在5.5寸上发现这个问题. 4.7寸正常 感谢反馈,我们这边修复一下
gharchive/issue
2018-06-14T02:05:57
2025-04-01T06:45:07.505980
{ "authors": [ "emilyzyw", "xfy6238" ], "repo": "netease-im/NIM_iOS_UIKit", "url": "https://github.com/netease-im/NIM_iOS_UIKit/issues/139", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
361918363
Stream back deploy & function logs Add ability for user to see deploy logs in CLI. Also the ability for user to see function logs in the CLI (potentially separate issue) We could stream back deploy logs on netlify watch and netlify deploy Notes: If this requires adding leaking firebase client into the CLI, we should put this feature off until a bespoke logs API is usable via https://open-api.netlify.com/#/default Fallback potentially looks like: https://github.com/netlify/cli/issues/70 ? Future/potential alt API: Potentially a master netlify logs command This functionality was initially implemented in https://github.com/netlify/cli/pull/762 but later reverted because of NodeJS compatibility issues. I'm think maybe we can precompile incompatible code and include that. Opening again to investigate further. @RaeesBhatti That would be an amazing feature (that competition already has implemented), would be nice to give it a second shot I think 😃 I would love to see this! This has been added in netlify-cli@17.5.0 🎉 netlify logs:deploy to have an interactive selector for currently active builds netlify logs:functions to have an interactive selector for all functions on a site netlify logs:functions my-function-name to instantly stream the logs for a given function, in this case my-function-name
gharchive/issue
2018-09-19T20:39:16
2025-04-01T06:45:07.521334
{ "authors": [ "DavidWells", "Jolg42", "RaeesBhatti", "sarahetter", "whitep4nth3r" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/issues/119", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1606729089
Netlify-cli update to 13.0.0 fails Describe the bug When running yarn global add netlify-cli I get the following error: yarn global v1.22.18 [1/4] Resolving packages... [2/4] Fetching packages... [3/4] Linking dependencies... warning "netlify-cli > @netlify/build > ts-node@10.9.1" has unmet peer dependency "@types/node@*". [4/4] Building fresh packages... [2/2] ⠄ netlify-cli error /usr/local/share/.config/yarn/global/node_modules/netlify-cli: Command failed. Exit code: 1 Command: node ./scripts/postinstall.mjs Arguments: Directory: /usr/local/share/.config/yarn/global/node_modules/netlify-cli Output: file:///usr/local/share/.config/yarn/global/node_modules/@netlify/build/node_modules/@netlify/config/lib/files.js:2 import { getProperty, setProperty, deleteProperty } from 'dot-prop'; ^^^^^^^^^^^^^^ SyntaxError: Named export 'deleteProperty' not found. The requested module 'dot-prop' is a CommonJS module, which may not support all module.exports as named exports. CommonJS modules can always be imported via the default export, for example using: import pkg from 'dot-prop'; const { getProperty, setProperty, deleteProperty } = pkg; at ModuleJob._instantiate (node:internal/modules/esm/module_job:123:21) at async ModuleJob.run (node:internal/modules/esm/module_job:189:5) at async Promise.all (index 0) at async ESMLoader.import (node:internal/modules/esm/loader:526:24) at async postInstall (file:///usr/local/share/.config/yarn/global/node_modules/netlify-cli/scripts/postinstall.mjs:29:35) at async file:///usr/local/share/.config/yarn/global/node_modules/netlify-cli/scripts/postinstall.mjs:60:1 Steps to reproduce Run 12.1.1 Get warning that a new version is available Run yarn global add netlify-cli Configuration No response Environment System: OS: Linux 5.15 Ubuntu 22.04.1 LTS 22.04.1 LTS (Jammy Jellyfish) CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz Memory: 8.97 GB / 30.79 GB Container: Yes Shell: 5.1.16 - /bin/bash Node version 16.18.1 I was able to reproduce this, but I think this is an issue with yarn. Doing yarn global remove netlify-cli and afterwards yarn global add netlify-cli resolved the issue. I checked all our package.json files and all the dependencies on dot-prop seem correct. Uninstall/reinstall worked for me.
gharchive/issue
2023-03-02T12:33:24
2025-04-01T06:45:07.525607
{ "authors": [ "LeonMelis", "danez" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/issues/5523", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
933830407
refactor: move function builders to functions lib - Summary Moves the function builders (netlify-lambda and zip-it-and-ship-it) to a new location, where all the JS runtime files will live. Part of #2806. - Test plan N/A @eduardoboucas, that failing test is flaky. I'll look into it - if the PR passes on other OS I think it's ok @eduardoboucas, that failing test is flaky. I'll look into it - if the PR passes on other OS I think it's ok I'm not in a rush to merge this, so I don't mind waiting a bit more and restarting the CI if we want to be cautious. I'll leave it up to you. Thanks for looking into it!
gharchive/pull-request
2021-06-30T15:26:38
2025-04-01T06:45:07.528161
{ "authors": [ "eduardoboucas", "erezrokah" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/pull/2811", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
575017676
Upgrade @netlify/config Due to https://github.com/netlify/build/pull/868, the latest major release of @netlify/config had to merge the method for resolving the configuration and the method for retrieving the configuration file path. This breaking change requires the CLI to be updated accordingly. This function is causing memory leaks in Netlify CLI because we try to parse config every time the config file changes and it adds new event listeners. I would like to get more clarity on why we're adding event handlers. Or why we aren't using configorama.sync in @netlify/config (which avoids those event handlers). cc @ehmicky @DavidWells The latest version of @neltify/config (which is upgraded by this PR) does not use configorama anymore, so that should solve this problem. Cool, thanks @ehmicky Please note @netlify/config also needs to be upgraded in netlify-redirect-parser: https://github.com/netlify/netlify-redirect-parser/blob/cc359c546926a456f8cc03957182ec7532ef2914/netlify-config-parser.js#L12 to prevent those leaks. I can work on that. @ehmicky any update on this? I was waiting for PR approvals to merge it.
gharchive/pull-request
2020-03-03T23:06:07
2025-04-01T06:45:07.532750
{ "authors": [ "RaeesBhatti", "ehmicky", "erquhart" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/pull/732", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
225850454
Specific media folders for collections It would be awesome to specify its own media_folder for each collection. Use case A lot of collections consist of chunks of information of different nature with an image attached to each one. Examples: carousel, partners' logos, testimonials, site specific collections. Request It is plausible to keep images for each collection in a separate folder. This will require either to specify a media_folder parameter for each collection separately, or to specify a folder to the field with a widget image. This may be implemented as a part of #350. Keeping open for now. It seems that #350 will take a lot of effort to come into reality. May I bump my request to allow its own media folders for collections? It should be relatively simple. Maybe, I'll succeed to produce a viable pull request. @aperep #350 came out just after your comment, ironically. We'd definitely like to move to asset collections soon after 1.0, which will be out in early December. Closing in favor of #1046 since discussion is moving on that one.
gharchive/issue
2017-05-02T23:32:15
2025-04-01T06:45:07.537053
{ "authors": [ "aperep", "erquhart" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/399", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
903697334
Sign up and Log in not working for www.netlify.com Describe the bug Sign up and Log in Page redirect to star background and nothing more, can not log on or sign up. To Reproduce Go to 'https://www.netlify.com/' Click on 'Sign Up' See how no forms or anything appears Expected behavior You should be able to sign up for an account Screenshots Applicable Versions: OS: Windows 10 Browser version Version 90.0.4430.212 (Official Build) (64-bit) Additional context This is trying to access as a user Hi @xarielx, this is a repo to report issues with Netlify CMS. For Netlify related issues I recommend reaching out in https://answers.netlify.com/. Regardless, I'm able to both login and get to the sign up page. I recommend: Trying in incognito mode Trying in a different browser. Checking you don't have any browser extensions that can block the site My apologies, closing it
gharchive/issue
2021-05-27T12:52:44
2025-04-01T06:45:07.543322
{ "authors": [ "erezrokah", "xarielx" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/5440", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1894685523
NETOBSERV-1185 Console plugin async overview metrics This PR improve Overview panel loading & code: [X] Renamed records & metrics routes to /loki/flow/records /loki/flow/metrics [X] Show graphs as soon as metrics are available [X] Keep previously loaded graph dimensions in local storage [X] Reduce number of queries according to selected panels / features [X] Add number of queries in summary [X] Fix testing You can either check the number of queries in the side panel or in your browser debug -> network section to test this :nerd_face: */!\ QE automated tests might break since the API route has been renamed and loading behavior changed /!* After testing, I can confirm my comment here that showing only [dns/rtt/drops] panel won't save us from doing a standard flows query, so I think that should be fixed. Also a small thing, looks like the Globe icon alignment isn't perfect: Apart from that, this looks good Also a small thing, looks like the Globe icon alignment isn't perfect: should be better with that: https://github.com/netobserv/network-observability-console-plugin/pull/389/commits/2be8de9ef1a5024eb8611e93d7c1f487716c4835 @nathan-weinberg would it be fine to merge this PR before QE validation ? It may break your automated testings :thinking: I have 3 PRs needing these changes. If you prefer to take your time I will rebase them on this PR. Thanks ! /cc @Amoghrd @jpinsonneau If it's a breaking change, let's try and update the tests first so our regressions do not turn up false positives - i.e. prefer the rebase method :D Do we have a Jira for this change? Do we have a Jira for this change? I've created https://issues.redhat.com/browse/NETOBSERV-1348 for you guys /ok-to-test With changes in https://github.com/openshift/openshift-tests-private/pull/11959 the test cases remain stable: https://mastern-jenkins-csb-openshift-qe.apps.ocp-c1.prod.psi.redhat.com/job/ocp-veno/job/no-regression/280/ @jpinsonneau any manual verifications need to be done before merging? @jpinsonneau any manual verifications need to be done before merging? Not particularly. Just that the graphs still shows :smile_cat: @jpinsonneau any manual verifications need to be done before merging? Not particularly. Just that the graphs still shows 😸 Sounds good - I am spinning up another cluster today to investigate some frontend failures @Amoghrd wanted me to double-check so I can do that as well! /ok-to-test Full Success Run with our automated tests: https://mastern-jenkins-csb-openshift-qe.apps.ocp-c1.prod.psi.redhat.com/job/ocp-veno/job/no-regression/281/ /label qe-approved Thanks for the team effort ! :partying_face: Merging now
gharchive/pull-request
2023-09-13T14:40:38
2025-04-01T06:45:07.559614
{ "authors": [ "jotak", "jpinsonneau", "nathan-weinberg" ], "repo": "netobserv/network-observability-console-plugin", "url": "https://github.com/netobserv/network-observability-console-plugin/pull/389", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1658466667
Closing Project dialog is shown for 30+seconds with Latte plugin enabled Describe the bug When switching projects or closing phpStorm the Closing Project dialog is shown for 30+ seconds and hangs the process Environment (please complete the following information): PhpStorm version 2023.1 Plugin version 1.1.7 To Reproduce Steps to reproduce the behavior (or attach video): Open a project Switch to another project using cmd+o and select project - or just quit phpStorm See that the closing dialog shows for a very very long time Screenshots Additional context Disabling the Latte support plugin fixes the problem Sama problem: https://github.com/nette-intellij/intellij-neon/issues/74 Same problem here... i hope its not because a "paid" plugin has been introduced. Haven't encountered this problem so far, so it's not the plugin by itself. Have the same problem. Well the closing is in fact not the problem, but, when you for example open some project and you close it after few seconds, you have to wait about 30 seconds, because this plugin is indexing... Neon plugin is doing the same thing Fixed in 1.1.8 it will be available in 2 business days Still no realease? 🙏 https://plugins.jetbrains.com/plugin/7457-latte/versions New release is out and working. Thanks 😉 Yes, I probably misclicked somehow when uploading the new version. Or some error on the JetBrains side. Anyway, it's out now :) Yes, I probably misclicked somehow when uploading the new version. Or some error on the JetBrains side. Anyway, it's out now :) It didn't get applied automatically, you have to manually go to plugin and press update :-( I hope its not because they are pushing their payed plugin now :-(( I have no influence on these updates. This should be handled by PhpStorm itself. I just upload the new version via the web interface. Probably PhpStorm downloads information about new versions only once in a while. So maybe tomorrow or in a few hours it would offer you an update automatically.
gharchive/issue
2023-04-07T07:16:35
2025-04-01T06:45:07.575081
{ "authors": [ "Gappa", "badpenguin", "honzavaclavik", "mesour", "rostislavcz" ], "repo": "nette-intellij/intellij-latte", "url": "https://github.com/nette-intellij/intellij-latte/issues/202", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
601958444
CompositeByteBuf.addComponent doesn't guard agaist overflow Expected behavior CompositeByteBuf.addComponent should throw an exception when adding a component overflows the maximum size of the buffer (Integer.MAX_VALUE). Actual behavior Both writableBytes and/or readableBytes (depending on setting for increaseWriterIndex) as well as capacity overflow into negative numbers. Steps to reproduce Run the reproducer in scala. Use a higher setting, e.g. -Xmx8G to experience the issue before running out of heap. Output will be something like: 400000000 400000000 800000000 800000000 1200000000 1200000000 1600000000 1600000000 2000000000 2000000000 -1894967296 -1894967296 -1494967296 -1494967296 -1094967296 -1094967296 -694967296 -694967296 -294967296 -294967296 105032704 105032704 ... Minimal yet complete reproducer code (or URL to code) import io.netty.buffer.{ ByteBufAllocator, Unpooled } object BufTest { def main(args: Array[String]): Unit = { val alloc = ByteBufAllocator.DEFAULT val buf = alloc.compositeBuffer() // We'll keep adding 400MB components while (true) { val arr = new Array[Byte](400 * 1000 * 1000) buf.addComponent(Unpooled.wrappedBuffer(arr)) println(s"${buf.writableBytes()} ${buf.capacity()}") } } } Netty version 4.1.44 final JVM version (e.g. java -version) openjdk version "11.0.5" 2019-10-15 LTS OpenJDK Runtime Environment Zulu11.35+15-CA (build 11.0.5+10-LTS) OpenJDK 64-Bit Server VM Zulu11.35+15-CA (build 11.0.5+10-LTS, mixed mode) OS version (e.g. uname -a) Darwin Attilas-MacBook-Pro-4.local 18.7.0 Darwin Kernel Version 18.7.0: Mon Feb 10 21:08:45 PST 2020; root:xnu-4903.278.28~1/RELEASE_X86_64 x86_64 @szegedi sounds like a bug... would you be interested to provide a PR ? I can try. Seems to me like calling AbstractByteBuf.checkIndex from CompositeByteBuf.addComponent0 is the right solution. @szegedi https://github.com/netty/netty/pull/10197 PTAL @normanmaurer thanks for the super-quick turnaround on this one! @szegedi no problem ... happy to help. Would love to hear more about how faunadb uses netty :) In one word: heavily 😄. All of network communication is async I/O based on Netty socket channel handler pipelines. We use the ByteBuf API pervasively too even outside of immediate domain of networking (disk I/O, internal transformations). It's a nice (consistent, comprehensive, easy to understand) API to manipulate byte-based data, and if we're smart about it then we also minimize copying when it's time to send the bytes over the network.
gharchive/issue
2020-04-17T13:00:01
2025-04-01T06:45:07.588997
{ "authors": [ "normanmaurer", "szegedi" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/10194", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
950393377
ERR - LEAK: ByteBuf.release() was not called before it's garbage-collected. Expected behavior Below error should not come: 'ERROR ResourceLeakDetector:319 - LEAK: ByteBuf.release() was not called before it's garbage-collected.' Actual behavior 2021-07-21 15:45:00 ERROR ResourceLeakDetector:319 - LEAK: ByteBuf.release() was not called before it's garbage-collected. See https://netty.io/wiki/reference-counted-objects.html for more information. Recent access records: #1: io.netty.buffer.AdvancedLeakAwareByteBuf.getBytes(AdvancedLeakAwareByteBuf.java:250) org.apache.sshd.netty.NettySupport$1.getRawBytes(NettySupport.java:44) org.apache.sshd.common.util.buffer.ByteArrayBuffer.putBuffer(ByteArrayBuffer.java:150) org.apache.sshd.common.util.buffer.Buffer.putBuffer(Buffer.java:491) org.apache.sshd.common.session.helpers.AbstractSession.messageReceived(AbstractSession.java:510) org.apache.sshd.common.session.helpers.AbstractSessionIoHandler.messageReceived(AbstractSessionIoHandler.java:63) org.apache.sshd.netty.NettyIoSession.channelRead(NettyIoSession.java:181) org.apache.sshd.netty.NettyIoSession$Adapter.channelRead(NettyIoSession.java:216) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:748) #2: Hint: 'NettyIoSession$Adapter#0' will handle the message from this point. io.netty.channel.DefaultChannelPipeline.touch(DefaultChannelPipeline.java:116) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:748) #3: Hint: 'DefaultChannelPipeline$HeadContext#0' will handle the message from this point. io.netty.channel.DefaultChannelPipeline.touch(DefaultChannelPipeline.java:116) io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:748) #4: io.netty.buffer.AdvancedLeakAwareByteBuf.writeBytes(AdvancedLeakAwareByteBuf.java:634) io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:350) io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:151) io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:748) Created at: io.netty.buffer.PooledByteBufAllocator.newDirectBuffer(PooledByteBufAllocator.java:402) io.netty.buffer.AbstractByteBufAllocator.directBuffer(AbstractByteBufAllocator.java:187) io.netty.buffer.AbstractByteBufAllocator.directBuffer(AbstractByteBufAllocator.java:178) io.netty.buffer.AbstractByteBufAllocator.ioBuffer(AbstractByteBufAllocator.java:139) io.netty.channel.DefaultMaxMessagesRecvByteBufAllocator$MaxMessageHandle.allocate(DefaultMaxMessagesRecvByteBufAllocator.java:114) io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:150) io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:748) Steps to reproduce NA - The err occurs on a random basis. Minimal yet complete reproducer code (or URL to code) NA Netty version Initially '4.1.29.Final' version was being used in our app when we noticed the err. FYI- We are using the below mentioned jars in our app: netty-buffer netty-codec netty-common netty-handler netty-resolver netty-transport Then we upgraded above jars to '4.1.65.Final', but still the same error is appearing in our app logs. FYI - We are using below apache commons dependencies :- commons-codec - 1.10 commons-io - 2.4 commons-lang - 2.6 commons-logging - 1.0.3 and below 'org.apache.sshd' dependencies - version 2.1.0 for all: sshd-common shd-sftp sshd-scp sshd-core sshd-mina sshd-netty sshd-putty sshd-cli JVM version (e.g. java -version) Java HotSpot(TM) 64-Bit Server VM (25.281-b25) for linux-amd64 JRE (1.8.0_281-b25) OS version (e.g. uname -a) Operating System - Linux OS Version - 11 SP4 Is this info sufficient? We are checking the exact os version with our Unix team, once we get the details, we will update. Queries :- Is this issue fixed in any of the netty versions? If yes - could you please let us know - which version? If not - could you please let us know in which version this fix is expected? Could you please check apache commons dependencies and org.apache.sshd dependencies which we are using? Do we need to also update these libs version to support netty higher version? #1: io.netty.buffer.AdvancedLeakAwareByteBuf.getBytes(AdvancedLeakAwareByteBuf.java:250) org.apache.sshd.netty.NettySupport$1.getRawBytes(NettySupport.java:44) Release the Bytebuf by yourself If you've got the option else report the issue to ASF. It's not a Netty bug. Hi hyperxpro, Thanks for the prompt revert. Could you please guide me on how and where to release the Bytebuf? as there is no application class showing up in the logs, so we are not sure which part of the code is to be fixed / where to release the Bytebuf? and what is ASF? Just call Bytebuf#release method. If you don't have access to that Bytebuf then you should report this issue to ASF (Apache Software Foundation) on their issue tracker because org.apache.sshd.netty.NettySupport$1.getRawBytes(NettySupport.java:44) this class is responsible for releasing Bytebuf.
gharchive/issue
2021-07-22T07:51:03
2025-04-01T06:45:07.610976
{ "authors": [ "hyperxpro", "npdbleak" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/11502", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
740941485
Fix UnsafeByteBufUtil#setBytes() cause JVM crash (#10791) Motivation: Passing a null value of byte[] to the Unsafe.copyMemory(xxx) would cause the JVM crash Modification: Add null checking before calling PlatformDependent.copyMemory(src, xxx) Result: Fixes #10791 . @Ech0Fan can you add a unit test as well and can you also sign our icla: https://netty.io/s/icla @Ech0Fan can you add a unit test as well and can you also sign our icla: https://netty.io/s/icla Yes I wanted to add a unit test, but seems I didn't see any unit test for UnsafeByteBufUtil, so... should I add a unit test named UnsafeByteBufUtilTest even if it is A non-public class Thanks @Ech0Fan. Couple of comments: Probably should move the null check outside of the nonzero length check block I think we also need to do bounds check on src relative to srcIndex and length Thanks for comments. I think that would be nice. @Ech0Fan yes please just add a unit test for the class... its package private so it will work :) @Ech0Fan yes please just add a unit test for the class... its package private so it will work :) I've added the unit test, verify it please @netty-bot test this please @Ech0Fan thanks a lot!
gharchive/pull-request
2020-11-11T17:28:28
2025-04-01T06:45:07.617309
{ "authors": [ "Ech0Fan", "normanmaurer" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/10795", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1075731716
Update tcnative version in bom Motivaiton: The bom is out of sync with the project version of tcnative. This results in unresolved dependencies when using the bom. main (ae50917ad1)
gharchive/pull-request
2021-12-09T15:55:53
2025-04-01T06:45:07.618775
{ "authors": [ "Scottmitch" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/11902", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
376240566
pull Motivation: Explain here the context, and why you're making that change. What is the problem you're trying to solve. Modification: Describe the modifications you've done. Result: Fixes #. If there is no issue then describe the changes introduced by this PR. @johnou yep I do... that said this looks not like a "valid" pr to me
gharchive/pull-request
2018-11-01T03:32:35
2025-04-01T06:45:07.620539
{ "authors": [ "darklish", "normanmaurer" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/8451", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
727696934
PR#312 IOSXE/show lisp eid-table vrf User ipv4 map-cache Description Created parser for show lisp eid-table vrf User ipv4 map-cache Screenshots: Checklist: [x] I have updated the changelog. [x] All new and existing tests passed. [x] All new code passed compilation. Cisco PR
gharchive/pull-request
2020-10-22T20:24:15
2025-04-01T06:45:07.622938
{ "authors": [ "qduk" ], "repo": "networktocode/genieparser", "url": "https://github.com/networktocode/genieparser/pull/141", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
648448958
XGD_Runtime Error and Loading Video problem Hello, thank you to everyone who contributed to this project. When I try it on Jetson nano, the video feels like loading continuously. It plays after 5 seconds, loads again, plays again for 5 seconds, then pauses. I refresh the page but the same problem continue. It also gives errors related to XGD_Runtime on console. I created the docker image with the command below. docker build -f jetson-nano.Dockerfile -t "neuralet/smart-social-distancing:latest-jetson-nano" . Also I tried it on my PC, the result is same. What am i missing ? Hi, Thanks for your feedback :) It's weird. Can you try running the docker without sudo? yes, but i can't run docker without sudo command. yes, but i can't run docker without sudo command. Try: sudo usermod -aG docker $USER Then logout/login or reboot. @memina Is this solved? I am having the same issue without any display Hey, we just got around checking these old issues, sorry about the delay. The issue could be related to processing speed or the old frontend. Considering a few improvements have been done on the former and that we deprecated the old frontend for a new one on react, I presume this has been fixed. (Although the video streaming is still a bit rough around the edges) Feel free to try to reproduce it if you are still interested and report back If performance is still an issue. Elsewise I'll close this issue in a week :) Hey, we just got around checking these old issues, sorry about the delay. The issue could be related to processing speed or the old frontend. Considering a few improvements have been done on the former and that we deprecated the old frontend for a new one on react, I presume this has been fixed. (Although the video streaming is still a bit rough around the edges) Feel free to try to reproduce it if you are still interested and report back If performance is still an issue. Elsewise I'll close this issue in a week :)
gharchive/issue
2020-06-30T19:49:08
2025-04-01T06:45:07.646672
{ "authors": [ "JsonSadler", "memina", "nabil-bukhari", "renzodgc" ], "repo": "neuralet/smart-social-distancing", "url": "https://github.com/neuralet/smart-social-distancing/issues/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
539666468
Move user's configuration to a separate Makefile In order to avoid Git conflicts between existing complicated targets of current Makefile and the custom user's code while Template upgrade. Plus, change the documentation to teach users not to edit the "our" Makefile but only "theirs". -- "their" Makefile: `Makefile` -- include neuro.mk DATA_DIR?=data CONFIG_DIR?=config CODE_DIR?={{cookiecutter.code_directory}} NOTEBOOKS_DIR?=notebooks RESULTS_DIR?=results ... -- "our" Makefile: `neuro.mk` -- ##### COMMANDS ##### APT?=apt-get -qq PIP?=pip install --progress-bar=off -U --no-cache-dir NEURO?=neuro ifeq (${TRAIN_STREAM_LOGS}, yes) TRAIN_WAIT_START_OPTION=--wait-start --detach else TRAIN_WAIT_START_OPTION= endif ... not urgent for Jan2020 iteration. outdated
gharchive/issue
2019-12-18T13:01:08
2025-04-01T06:45:07.655122
{ "authors": [ "YevheniiSemendiak", "artemlops", "ayushkovskiy" ], "repo": "neuro-inc/cookiecutter-neuro-project", "url": "https://github.com/neuro-inc/cookiecutter-neuro-project/issues/262", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
493820094
Wrong calculation of MTR In the example below, the calculated MTR value is -53.41%, whereas it should be (633-303)/633=52%: When running the code in debug mode and calling this voxel, I get the correct value: (data_mt0[124,134,16]-data_mt1[124,134,16])/data_mt0[124,134,16] 0.521327 This issue might be related to a reported issue related to division of numpy array of integer. A solution would be to convert the input to float before doing the division. If confirmed, unit tests should be added to make sure this function behaves properly regardless the MRI input data type. ah! the use of np.true_divide seems to solve that bug: data_mtr = 100 * (data_mt0 - data_mt1) / data_mt0 data_mtr[124,134,16] -51.39968404423381 data_mtr = 100 * np.true_divide((data_mt0 - data_mt1), data_mt0) data_mtr[124,134,16] 52.13270142180095
gharchive/issue
2019-09-16T02:23:08
2025-04-01T06:45:07.677220
{ "authors": [ "jcohenadad" ], "repo": "neuropoly/spinalcordtoolbox", "url": "https://github.com/neuropoly/spinalcordtoolbox/issues/2441", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
710541372
Documentation showing error. https://spinalcordtoolbox.com/en/latest/user_section/command-line.html#sct-register-to-template Shows an error text and not the actual documentation. oopsi 😬 , thanks for catching that (and for reporting it!) @lrouhier ! My apologies. This has to do with one of my changes during the argparse conversion, see: https://github.com/neuropoly/spinalcordtoolbox/pull/2842/commits/4e97be6af8af5498da04fb50c51b497c7748eae9. The SCT course Slideshare links were replaced with iCloud links, but the iCloud links contain the % character. Apparently argparse throws a bit of a fit with % in help description strings, see also [link 1], [link 2]. I'll start on this ASAP. wow! that's an evil bug 🐞 another incentive to move away from icloud for SCT course (i'm already salivating at our future datalad-like SCT course 😛)
gharchive/issue
2020-09-28T20:04:28
2025-04-01T06:45:07.680973
{ "authors": [ "jcohenadad", "joshuacwnewton", "lrouhier" ], "repo": "neuropoly/spinalcordtoolbox", "url": "https://github.com/neuropoly/spinalcordtoolbox/issues/2927", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
307527389
why xrdp does not use freerdp,bugs in NeutrinoRDP are no longer fixed. such as: https://www.talosintelligence.com/vulnerability_reports/TALOS-2017-0337 The reason why xrdp switched from FreeRDP to NeutrinoRDP (fork) is stability. When we decided to fork FreeRDP, it may be 2012-2013, FreeRDP wasn't stable and makes drastic changes every day. We were tired to follow such changes. However, recent days, FreeRDP (2.0.0) seems to be stable. I personally think we should support both FreeRDP and NeutrinoRDP and provide users a choice which to use. Regarding NeutrinoRDP, "no longer fixed" is not correct. It is our fork. Can be fixed. Hi metalefty :) Some answer's can you copy in a "FaQ" and be able to close, therewith have you a Faq where the Peoples can read and you can linked every by closing :) best regards Blacky
gharchive/issue
2018-03-22T07:01:20
2025-04-01T06:45:07.691256
{ "authors": [ "blackcrack", "metalefty", "zhonghesc" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/1075", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
849872094
Configure xrdp on Ubuntu 20.04 Hi By setting port=tcp://:3389, I see the following netstat output $ sudo netstat -tulpn | grep xrdp tcp 0 0 0.0.0.0:3389 0.0.0.0:* LISTEN 9320/xrdp tcp6 0 0 ::1:3350 :::* LISTEN 6051/xrdp-sesman However, from my windows machine, I am not able to connect and returns error in the initiating remote connection.... What else should I check? Maybe still the daemon doesn't receive the request. @mahmoodn Here are some pointers for you:- Leave sudo tail -f /var/log/xrdp.log running and then try to connect. If you get output that will show whether you're getting through or not. Are you running a firewall? What does sudo ufw status tell you? Please see below: $ sudo ufw status [sudo] password for mahmood: Status: inactive $ ps aux | grep xrdp root 6051 0.0 0.0 24436 508 ? S Apr 04 0:00 /usr/sbin/xrdp-sesman xrdp 9320 0.0 0.0 24564 612 ? S Apr 04 0:00 /usr/sbin/xrdp mahmood 63171 0.0 0.0 20544 2804 pts/2 S+ 14:45 0:00 grep --color=auto xrdp But as you can see, nothing is recorded when I try to connect to the server. It looks like there's a name resolution or connectivity error here. If you had managed to reach the Ubuntu machine you'd have more in the log than that. Can you get a command line up on the Windows machine and ping the hostname you're trying to connect to? Without a firewall, ping should work. Check the IP address displayed by ping is the same on your Ubuntu machine with ip addr on the Ubuntu side. Yes I can ping that. However, the IP address assigned on the Ubuntu interface is an internal IP (172.20.54.10). There is an entry in the department firewall/router that maps an external IP to 172.20.54.10. We use the external IP from abroad and I am able to ping and ssh to the machine. The most likely thing in that case is that the departmental firewall which is NATting to your 172.20.0.0/16 address is not allowing TCP port 3389, which would be standard procedure. You should check with them. If that's the case and they won't open TCP/3389 (they shouldn't frankly), you can still make use of the ssh port if you've got PuTTY installed on your Windows machine, and you're able to make an ssh connection to your Ubuntu machine. The idea is to set up a tunnel over ssh so that you can connect to a local port which then gets forwarded over ssh to the remote Ubuntu machine. It's best not to use TCP/3389 for the local port as Windows itself may be listening on it. Use TCP/3390 instead. Procedure is:- Make sure you can use PuTTY to get to your Ubuntu machine from your Windows machine. Save the session settings you need to do this. While you're logged out, go to the Connection/SSH/Tunnels settings, and add a forward from local port 3390 to 3389 on your target machine. It'll look something like this, where <target> is the hostname or IP address of your external IP:-L3390 <target>:3389 Log in to the machine again over ssh using PuTTY. On the Windows machine, run MSTSC.EXE and connect to localhost:3390 If all is well, PuTTY will tunnel the TCP connection for you, and you'll be able to log in graphically using xrdp. @mahmoodn - do you need any more help with this? I didn't get a clear answer from my department if the port is opened or not! So, I wasn't able to go further. For now, I will close the topic. Thanks for your help.
gharchive/issue
2021-04-04T13:11:17
2025-04-01T06:45:07.699812
{ "authors": [ "mahmoodn", "matt335672" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/1854", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1342853667
When windows server 2016 enable NLA, Cann't connect with tls error via Proxy Mode xrdp is running on centos as a Proxy for user to connect windows server 2016, if the windows server enbale NLA, user cann't connect to the server. As logs shown above,the windows server is 192.168.213.238 [20220818-17:18:46] [INFO ] FreeRDP version major 1 minor 0 revision 1 [20220818-17:18:46] [INFO ] loaded module 'libxrdpneutrinordp.so' ok, interface size 411072, version 4 [20220818-17:18:46] [WARN ] The colormap is all NULL [20220818-17:18:46] [ERROR] libxrdp_query_channel: Channel index out of range. max channel index 4, received channel index 4 [20220818-17:18:47] [INFO ] lxrdp_connect: freerdp_connect returned 0 [20220818-17:18:47] [INFO ] Failure to connect [20220818-17:18:47] [ERROR] freerdp_connect Failed to destination :192.168.213.238:3389 [20220818-17:18:47] [ERROR] xrdp_wm_log_msg: Error connecting to: 192.168.213.238 [20220818-17:18:47] [INFO ] Error connecting to: 192.168.213.238 [20220818-17:18:59] [ERROR] xrdp_sec_recv: xrdp_mcs_recv failed [20220818-17:18:59] [ERROR] xrdp_rdp_recv: xrdp_sec_recv failed [20220818-17:18:59] [ERROR] libxrdp_process_data: xrdp_rdp_recv failed [20220818-17:18:59] [ERROR] xrdp_process_data_in: xrdp_process_loop failed [20220818-17:18:59] [ERROR] SSL_write: Failure in SSL library (protocol error?) [20220818-17:18:59] [ERROR] xrdp_iso_send: trans_write_copy_s failed [20220818-17:18:59] [ERROR] SSL_shutdown: Failure in SSL library (protocol error?) [20220818-17:18:59] [ERROR] Sending [ITU T.125] DisconnectProviderUltimatum failed ===================================================================== I captured the tls packet into the tls.zip : tls.zip Requires #1338 to solve this. when user account contains domain names, use username@domain can work properly while using domain\username which will returns Account Error
gharchive/issue
2022-08-18T09:50:15
2025-04-01T06:45:07.703392
{ "authors": [ "chiefliu", "metalefty" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/2345", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
246206634
Blank screen after login After logging in via xrdp i get a blank screen and nothing happens. Journalctl -xe gives the error "Listening socket is in wrong state, terminating listener". You should at least mention about xrdp version, client what you use, where your xrdp obtained from. Im using xrdp 0.9.1-7build1 on ubuntu 17.04 and downloaded xrdp with apt. xrdp.log [20170728-13:33:36] [INFO ] A connection received from: ::ffff:90.187.5.173 port 2091 [20170728-13:33:36] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:36] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20170728-13:33:37] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20170728-13:33:37] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20170728-13:33:37] [DEBUG] Security layer: requested 11, selected 0 [20170728-13:33:37] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:37] [ERROR] Listening socket is in wrong state, terminating listener [20170728-13:33:37] [INFO ] A connection received from: ::ffff:90.187.5.173 port 2092 [20170728-13:33:37] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:37] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20170728-13:33:37] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20170728-13:33:37] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20170728-13:33:37] [DEBUG] Security layer: requested 0, selected 0 [20170728-13:33:38] [DEBUG] xrdp_00000647_wm_login_mode_event_00000001 [20170728-13:33:38] [INFO ] Loading keymap file /etc/xrdp/km-00000407.ini [20170728-13:33:38] [WARN ] local keymap file for 0x00000407 found and doesn't match built in keymap, using local keymap file [20170728-13:33:38] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:42] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:44] [DEBUG] xrdp_wm_log_msg: connecting to sesman ip 127.0.0.1 port 3350 [20170728-13:33:45] [INFO ] xrdp_wm_log_msg: sesman connect ok [20170728-13:33:45] [DEBUG] xrdp_wm_log_msg: sending login info to session manager, please wait... [20170728-13:33:45] [DEBUG] return value from xrdp_mm_connect 0 [20170728-13:33:45] [INFO ] xrdp_wm_log_msg: login successful for display 10 [20170728-13:33:45] [DEBUG] xrdp_wm_log_msg: started connecting [20170728-13:33:45] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:49] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:49] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:52] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:53] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:56] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:56] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:59] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:00] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:03] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:03] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:06] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:07] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_wm_log_msg: connection problem, giving up [20170728-13:34:10] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_wm_log_msg: some problem [20170728-13:34:10] [DEBUG] xrdp_mm_module_cleanup [20170728-13:34:10] [DEBUG] Closed socket 16 (AF_INET6 ::1 port 59238) [20170728-13:34:10] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:34:10] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_mm_module_cleanup [20170728-13:34:10] [ERROR] Listening socket is in wrong state, terminating listener [20170728-13:34:14] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:17] [DEBUG] Closed socket 17 (AF_UNIX) xrdp-sesman.log [20170728-13:33:44] [INFO ] A connection received from ::1 port 59262 [20170728-13:33:45] [INFO ] ++ created session (access granted): username server, ip 0.0.0.0:2092 - socket: 12 [20170728-13:33:45] [INFO ] starting Xorg session... [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 5910) [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 6010) [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 6210) [20170728-13:33:45] [DEBUG] Closed socket 8 (AF_INET6 ::1 port 3350) [20170728-13:33:45] [DEBUG] Closed socket 7 (AF_INET6 ::1 port 3350) [20170728-13:33:45] [INFO ] Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp [20170728-13:33:55] [ERROR] X server for display 10 startup timeout [20170728-13:33:55] [INFO ] starting xrdp-sessvc - xpid=1611 - wmpid=1610 [20170728-13:33:55] [ERROR] X server for display 10 startup timeout [20170728-13:33:55] [ERROR] another Xserver might already be active on display 10 - see log [20170728-13:33:55] [DEBUG] aborting connection... [20170728-13:33:55] [INFO ] ++ terminated session: username server, display :10.0, session_pid 1609, ip 0.0.0.0:2092 - socket: 12 Im using xrdp 0.9.1-7build1 on ubuntu 17.04 and downloaded xrdp with apt. xrdp.log [20170728-13:33:36] [INFO ] A connection received from: ::ffff:90.187.5.173 port 2091 [20170728-13:33:36] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:36] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20170728-13:33:37] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20170728-13:33:37] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20170728-13:33:37] [DEBUG] Security layer: requested 11, selected 0 [20170728-13:33:37] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:37] [ERROR] Listening socket is in wrong state, terminating listener [20170728-13:33:37] [INFO ] A connection received from: ::ffff:90.187.5.173 port 2092 [20170728-13:33:37] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:33:37] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20170728-13:33:37] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20170728-13:33:37] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20170728-13:33:37] [DEBUG] Security layer: requested 0, selected 0 [20170728-13:33:38] [DEBUG] xrdp_00000647_wm_login_mode_event_00000001 [20170728-13:33:38] [INFO ] Loading keymap file /etc/xrdp/km-00000407.ini [20170728-13:33:38] [WARN ] local keymap file for 0x00000407 found and doesn't match built in keymap, using local keymap file [20170728-13:33:38] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:42] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:44] [DEBUG] xrdp_wm_log_msg: connecting to sesman ip 127.0.0.1 port 3350 [20170728-13:33:45] [INFO ] xrdp_wm_log_msg: sesman connect ok [20170728-13:33:45] [DEBUG] xrdp_wm_log_msg: sending login info to session manager, please wait... [20170728-13:33:45] [DEBUG] return value from xrdp_mm_connect 0 [20170728-13:33:45] [INFO ] xrdp_wm_log_msg: login successful for display 10 [20170728-13:33:45] [DEBUG] xrdp_wm_log_msg: started connecting [20170728-13:33:45] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:49] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:49] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:52] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:53] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:56] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:56] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:33:59] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:00] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:03] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:03] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:06] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:07] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_wm_log_msg: connection problem, giving up [20170728-13:34:10] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_wm_log_msg: some problem [20170728-13:34:10] [DEBUG] xrdp_mm_module_cleanup [20170728-13:34:10] [DEBUG] Closed socket 16 (AF_INET6 ::1 port 59238) [20170728-13:34:10] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:192.168.178.38 port 3389) [20170728-13:34:10] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:10] [DEBUG] xrdp_mm_module_cleanup [20170728-13:34:10] [ERROR] Listening socket is in wrong state, terminating listener [20170728-13:34:14] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:17] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:21] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:24] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:28] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:31] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:35] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:38] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:34:42] [DEBUG] Closed socket 17 (AF_UNIX) [20170728-13:36:14] [DEBUG] Closed socket 17 (AF_UNIX) xrdp-sesman.log [20170728-13:33:44] [INFO ] A connection received from ::1 port 59262 [20170728-13:33:45] [INFO ] ++ created session (access granted): username server, ip 0.0.0.0:2092 - socket: 12 [20170728-13:33:45] [INFO ] starting Xorg session... [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 5910) [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 6010) [20170728-13:33:45] [DEBUG] Closed socket 9 (AF_INET6 :: port 6210) [20170728-13:33:45] [DEBUG] Closed socket 8 (AF_INET6 ::1 port 3350) [20170728-13:33:45] [DEBUG] Closed socket 7 (AF_INET6 ::1 port 3350) [20170728-13:33:45] [INFO ] Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp [20170728-13:33:55] [ERROR] X server for display 10 startup timeout [20170728-13:33:55] [INFO ] starting xrdp-sessvc - xpid=1611 - wmpid=1610 [20170728-13:33:55] [ERROR] X server for display 10 startup timeout [20170728-13:33:55] [ERROR] another Xserver might already be active on display 10 - see log [20170728-13:33:55] [DEBUG] aborting connection... [20170728-13:33:55] [INFO ] ++ terminated session: username server, display :10.0, session_pid 1609, ip 0.0.0.0:2092 - socket: 12 Did you read /usr/share/doc/xrdp/README.Debian? Follow instructions written in the file. In this case, README.Debian doesn't solve the issue. Please forget it. I have a similar problem after updating to Debian 9. Please help me. Xorg is properly started but terminated immediately. startwm.sh needs to be adjusted. No feedback from the reporter, closing. Maybe startwm.sh needs to be adjusted. When you're back, please leave a comment. I'll reopen this. Problem solved with latest beta / release candidate.
gharchive/issue
2017-07-28T00:32:53
2025-04-01T06:45:07.729702
{ "authors": [ "finnbrue", "metalefty", "mightyraven", "tps800" ], "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/839", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
841796309
On click UI navigation on top get undefined value latest version When i click Top navigation, onStepChange return with undefined value. And top navigation lose focus @Hawerrr sorry for the delay - did you resolve this? if not - can you provide more details...
gharchive/issue
2021-03-26T10:17:33
2025-04-01T06:45:07.776442
{ "authors": [ "Hawerrr", "newbreedofgeek" ], "repo": "newbreedofgeek/react-stepzilla", "url": "https://github.com/newbreedofgeek/react-stepzilla/issues/167", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
230279460
Cannot write more than one Profile to MySQL database. ###_MASTER TEMPLATE: Use for bugs, tasks, and future requests. Also, label accordingly, using the labels on the right. ###_Issue Type - is it a bug, a task, or a future request?: BUG. ###_Severity (1-4; 1 - not severe, 4 - very severe): 2 ###_Priority (1-4; 1 - lowest priority, 4 - highest priority): 2 ###_Synopsis - describe bug/task/future request, in more detail: Cannot write more than one Profile to MySQL database. Consequently, you cannot list more than the first profile made under View Profiles. When you make a profile under Add Profiles, it'll then appear under the View profiles page. However, it only shows the first profile that you made. Any subsequent profiles show that they're accepted, but don't show up. ###_Expected Behavior - what the behavior is supposed to be, for bug/task/FR: It would either write more than one profile to the SQL DB, and then make them viewable, or the website would recognize that you already have one profile in the DB, and will prevent you from making another one. This is a dev scope that has to be decided by the Dev Team. ###_Actual Behavior - what the behavior is instead, for bugs only: You only see the very first profile made. Screenshot attached. This is because it only writes that first profile to the SQL database. ###_Steps to Repro the Issue, for bugs only: Step 1: Go to Profiles/Add a Profile on the Nav bar. Step 2: Fill out all of the fields, including uploading an image. Make it obvious test data by putting in some variant of 'test' in every field. Step 3: For the email field, please use either your own email, or example.com. We don't want to put in a stranger's email address. Step 4: Click the Add button. ###_Other Notes, for bug/task/future request: I am closing this issue; I have verified that I can add more than one profile now to the Profile Page. It also writes those profiles to the DB.
gharchive/issue
2017-05-22T04:03:15
2025-04-01T06:45:07.796004
{ "authors": [ "se-linde" ], "repo": "newmanix/itc260-sp15-gig-central", "url": "https://github.com/newmanix/itc260-sp15-gig-central/issues/112", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
224913038
Feature request: Activate When Plugged into External Display Hi, this is absolutely a "nice to have but not necessary" thing. It would be nice if there was an option to enable automatically when an external display is plugged in and disabled when it's unplugged. This would actually save me from quite a lot of reboots -- my Mac regularly crashes when connected to an external display and going to sleep -- disabling sleep only when connected to the display would resolve this perfectly. Created an MR for this.
gharchive/issue
2017-04-27T21:05:48
2025-04-01T06:45:07.797595
{ "authors": [ "caseyWebb", "lazyatom", "sturza" ], "repo": "newmarcel/KeepingYouAwake", "url": "https://github.com/newmarcel/KeepingYouAwake/issues/84", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
61673370
Add ability to check for container with given name This adds a new argument, -container-name, which will make check_docker check that a container with the specified name exists and is running. This is useful in cases where you use the same container more than once, but wish to ensure all instances are up. For example, two wordpress blogs using the container wordpress but named wordpress_A and wordpress_B. I just realized, this is one possible solution to the problem described in #6. Ping @didip @relistan @intjonathan. Feedback on this would be appreciated. This is great! Thanks for the contribution!
gharchive/pull-request
2015-03-14T18:33:44
2025-04-01T06:45:07.800088
{ "authors": [ "didip", "zoni" ], "repo": "newrelic/check_docker", "url": "https://github.com/newrelic/check_docker/pull/10", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
854734411
OverviewTemplate.js should support embedding functionality Relates to #1227 and PR #1201 Enhancement We'd like to be able embed index pages such as https://developer.newrelic.com/build-apps https://developer.newrelic.com/ab-test currently when you apply /embed/ to these pages you receive a 404 page Acceptance Criteria The OverviewTemplate.js should support embedding functionality All Overview index pages can now have /embed to their path and the content will be embedable. This may be as easy as just adding OverviewTemplate to this query @jpvajda / @alexronquillo Out of curiosity, why would we want to embed these pages? It looks like their purpose is to aid users in navigating around the developer site. I'm not sure I see the value of embedding these. Ah I see, we're using the overviewTemplate for the A/B test page. For the most part, we use that as a landing page template - it looks like we're using it here as a work-around to get a single page layout. @zstix bingo.. perhaps we need a new template then and not take this approach? @zstix I wonder if the correct way to approach this is build a single page template, vs trying to support /embed on the overviewTemplate, though a quick work around could be to enable that on the overviewTemplate, it seems really straight forward. The solution is to convert this page to a guide template instead of using an overview template, since this page is functioning as guide and less so an overview I'll check in with @alexronquillo if this will work instead of trying to add embed to the overview template. This is fine. Thanks. Honestly, I probably copied the OverviewTemplate from another index page. I'm still learning what templates are even used for. We're actually not embedding this page anyway, I just noticed that it couldn't be embedded even if we wanted to, which seemed like a bug.
gharchive/issue
2021-04-09T17:47:16
2025-04-01T06:45:07.805839
{ "authors": [ "alexronquillo", "jpvajda", "zstix" ], "repo": "newrelic/developer-website", "url": "https://github.com/newrelic/developer-website/issues/1240", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
687218405
[nri-bundle] Bundle logging psp Is this a new chart No What this PR does / why we need it: Updating the nri-bundle to use newrelic-logging version 1.2.0 We need to wait for PR 112 to be merged first (https://github.com/newrelic/helm-charts/pull/112) Which issue this PR fixes Special notes for your reviewer: Is the requirements.lock file automatically generated? This PR is dependent upon PR 112 https://github.com/newrelic/helm-charts/pull/112 Checklist [Place an '[x]' (no spaces) in all applicable fields. Please remove unrelated fields.] [x] Chart Version bumped [x] Variables are documented in the README.md [x] Title of the PR starts with chart name (e.g. [mychartname]) I guess the test fails because the other PR is not merged in yet
gharchive/pull-request
2020-08-27T12:52:31
2025-04-01T06:45:07.853557
{ "authors": [ "polfliet" ], "repo": "newrelic/helm-charts", "url": "https://github.com/newrelic/helm-charts/pull/114", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
729732491
Split register batches based on size This PR will split the register batches based on the json size Pull Request Test Coverage Report for Build 329531269 27 of 31 (87.1%) changed or added relevant lines in 3 files are covered. 1 unchanged line in 1 file lost coverage. Overall coverage increased (+0.03%) to 57.904% Changes Missing Coverage Covered Lines Changed/Added Lines % pkg/entity/id.go 0 4 0.0% Files with Coverage Reduction New Missed Lines % internal/agent/event_sender_vortex.go 1 73.21% Totals Change from base Build 329237691: 0.03% Covered Lines: 10981 Relevant Lines: 18964 💛 - Coveralls
gharchive/pull-request
2020-10-26T16:48:39
2025-04-01T06:45:07.861380
{ "authors": [ "coveralls", "cristianciutea" ], "repo": "newrelic/infrastructure-agent", "url": "https://github.com/newrelic/infrastructure-agent/pull/190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1050545770
Switch to IMDSv2 for metadata calls. I can see that inside lib/new_relic/agent/utilization/aws.rb, you are using IMDSv1 calls, to fetch instance details, however, it is a security recommendation that we use IMDSv2 calls for this purpose. AWS security hub recommends having IMDSv2 calls only. So, this gem using v1 calls is a blocker for us to move forward on this. Priority: Blocker hello @vinoth2710 , Thank you for letting us know that this issue is impacting you. This is a Duplicate of #552, so I will be closing this issue so it can be tracked in one ticket. Duplicate of #552
gharchive/issue
2021-11-11T03:56:32
2025-04-01T06:45:07.875171
{ "authors": [ "tannalynn", "vinoth2710" ], "repo": "newrelic/newrelic-ruby-agent", "url": "https://github.com/newrelic/newrelic-ruby-agent/issues/848", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1027778877
Update nr1-graphiql-notebook Acceptance Criteria [ ] Confirm with the programmability 🧙 that there are no concerns with removing this app. [ ] Follow the end of life process to remove this application from the catalog [ ] Archive the repo Usage I've let some people know our intentions to EOL this app just waiting back on any concerns, I've also changed the AC to be focused on the end of life steps we need to take.
gharchive/issue
2021-10-15T20:18:41
2025-04-01T06:45:07.877548
{ "authors": [ "aswanson-nr", "jpvajda" ], "repo": "newrelic/nr1-catalog", "url": "https://github.com/newrelic/nr1-catalog/issues/313", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2348855261
Plugin not found v5.2.0 Receiving the following error after upgrading from v5.1.0 to v5.2.0: Serverless plugin "serverless-newrelic-lambda-layers" not found. Make sure it's installed and listed in the "plugins" section of your serverless config file. Run "serverless plugin install -n serverless-newrelic-lambda-layers" to install it. Any ideas why this would happen? No other changes made to the serverless.yml. @lukehedger might be related to #449. Can you try again we released 5.2.1 Thanks @bizob2828
gharchive/issue
2024-06-12T13:59:33
2025-04-01T06:45:07.879605
{ "authors": [ "bizob2828", "lukehedger" ], "repo": "newrelic/serverless-newrelic-lambda-layers", "url": "https://github.com/newrelic/serverless-newrelic-lambda-layers/issues/448", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
747433
PID check failed. RNG must be re-initialized after fork(). Hint: Try Random.atfork() I just tried beeswithmachineguns for the first time and spun up 4 instances with the default settings and attacked oe of my urls with default settings and got this stacktrace: Read 4 bees from the roster. Connecting to the hive. Assembling bees. Each of 4 bees will fire 250 rounds, 25 at a time. Stinging URL so it will be cached for the attack. Organizing the swarm. Bee 0 is joining the swarm. Bee 1 is joining the swarm. Bee 2 is joining the swarm. Bee 3 is joining the swarm. No handlers could be found for logger "paramiko.transport" Traceback (most recent call last): File "./bees", line 5, in <module> main.main() File "/home/zz/src/beeswithmachineguns/beeswithmachineguns/main.py", line 121, in main parse_options() File "/home/zz/src/beeswithmachineguns/beeswithmachineguns/main.py", line 113, in parse_options bees.attack(options.url, options.number, options.concurrent) File "/home/zz/src/beeswithmachineguns/beeswithmachineguns/bees.py", line 319, in attack results = pool.map(_attack, params) File "/usr/lib/python2.7/multiprocessing/pool.py", line 199, in map return self.map_async(func, iterable, chunksize).get() File "/usr/lib/python2.7/multiprocessing/pool.py", line 491, in get raise self._value AssertionError: PID check failed. RNG must be re-initialized after fork(). Hint: Try Random.atfork() Python version is 2.7.1 linux distribution is arch Hi there I ran bees with 10 instances and everything worked fine, however when I ran it with 45 instances I now get the same error listed above. I am not very knowledgable in Python, but I am pretty sure I am using Paramiko 1.7.7.1 as I ran sudo pip install beeswithmachineguns --upgrade and it outputted the following: sudo pip install beeswithmachineguns --upgrade Downloading/unpacking beeswithmachineguns Running setup.py egg_info for package beeswithmachineguns Downloading/unpacking boto==2.0 (from beeswithmachineguns) Running setup.py egg_info for package boto Downloading/unpacking paramiko==1.7.7.1 (from beeswithmachineguns) Running setup.py egg_info for package paramiko Downloading/unpacking pycrypto>=2.1 (from paramiko==1.7.7.1->beeswithmachineguns) Running setup.py egg_info for package pycrypto Installing collected packages: beeswithmachineguns, boto, paramiko, pycrypto .... Here is the error trace I get: bees attack -n 20000 -c 250 -u http://my-server.eu-west-1.elb.amazonaws.com/ Read 45 bees from the roster. Connecting to the hive. Assembling bees. Each of 45 bees will fire 444 rounds, 5 at a time. Stinging URL so it will be cached for the attack. Organizing the swarm. Bee 0 is joining the swarm. Bee 1 is joining the swarm. ... Bee 43 is joining the swarm. Bee 44 is joining the swarm. No handlers could be found for logger "paramiko.transport" Traceback (most recent call last): File "/usr/local/bin/bees", line 5, in <module> main.main() File "/Library/Python/2.7/site-packages/beeswithmachineguns/main.py", line 127, in main parse_options() File "/Library/Python/2.7/site-packages/beeswithmachineguns/main.py", line 119, in parse_options bees.attack(options.url, options.number, options.concurrent) File "/Library/Python/2.7/site-packages/beeswithmachineguns/bees.py", line 325, in attack results = pool.map(_attack, params) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/multiprocessing/pool.py", line 199, in map return self.map_async(func, iterable, chunksize).get() File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/multiprocessing/pool.py", line 491, in get raise self._value AssertionError: PID check failed. RNG must be re-initialized after fork(). Hint: Try Random.atfork() If anyone could help I would be most appreciative. I love bees with machine guns so far, shame about this problem. For everyone else who hits this problem, it seems the problem exists with paramiko 1.7.7.1 as well. @dimaqq however provided a valid work around for the issue by modifying the paramiko module file paramiko/transport.py Here is a gist of paramiko/transport.py file if you want it
gharchive/issue
2011-04-09T09:55:47
2025-04-01T06:45:07.885009
{ "authors": [ "mattheworiordan", "zoranzaric" ], "repo": "newsapps/beeswithmachineguns", "url": "https://github.com/newsapps/beeswithmachineguns/issues/17", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2763693164
Docs: use proper markup for key presses Back in the days, we put the key strokes in double quotes. However, this convention was changed to AsciiDoc's kbd: syntax later. Apparently, we missed some key strokes during the migration. So, this commit changes the remaining ones to the proper syntax. Lowercase keys in the old syntax are now converted to uppercase. The old uppercase keys still remain uppercase, but get an explicit Shift prefix. coverage: 61.019%. remained the same when pulling 615c30cbdf57e689d80c89c22c10c824172dbcc8 on fix-key-stroke-markup into 8b779b7f573e63965fb9a27f55e3c052dca288df on master.
gharchive/pull-request
2024-12-30T21:02:41
2025-04-01T06:45:07.887511
{ "authors": [ "coveralls", "der-lyse" ], "repo": "newsboat/newsboat", "url": "https://github.com/newsboat/newsboat/pull/2968", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
354935474
Binary filters not filtering correctly Has a Value active: "true" filter active: As the above pictures show, binary column filters like Is Source and Is Binary filter correctly on "Has a Value", but fail to filter when a specific "true" or "false" value is selected in the filter drop-down box. The behavior is observed on all Binary choice columns. #301 merged, closing this.
gharchive/issue
2018-08-28T23:03:26
2025-04-01T06:45:07.907178
{ "authors": [ "majurg" ], "repo": "nexB/aboutcode-manager", "url": "https://github.com/nexB/aboutcode-manager/issues/295", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1412786813
AltID Motivation Many identity provider has both username-like and user_id-like identifier. We need to record both in case username is changed by the user after vaildator.Validate(). To make this new system do a minimal impact on current behaviour, here's a modest change to start. In the future, some of the user_id-based identifier should be put in Identity field, like twitter, github, discord and minds. Checklist [x] Models [x] Proof [x] ProofChain [x] validators [x] validator.Base LGTM
gharchive/pull-request
2022-10-18T08:24:55
2025-04-01T06:45:08.332681
{ "authors": [ "fengshanshan", "nykma" ], "repo": "nextdotid/proof_server", "url": "https://github.com/nextdotid/proof_server/pull/37", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1938592095
Allow to disable welcome page When passing the same :paths or :paths-fn arg as for build! to serve! Clerk will no longer the built-in welcome page and instead serve the generated or user-specified index page on root. This is a first step towards enabling easy sharing fully interactive Clerk documents including pagination &c. Also tested static builds locally, didn't see any regressions there.
gharchive/pull-request
2023-10-11T19:39:42
2025-04-01T06:45:08.359255
{ "authors": [ "mk", "zampino" ], "repo": "nextjournal/clerk", "url": "https://github.com/nextjournal/clerk/pull/561", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
2318238082
Add Clade-I build Adds a Clade-I build. The rooting is currently using least-squares (clock vs div). Starting this work in preparation for new DRC sequences (not yet submitted). Preview build at nextstrain.org/staging/mpox/clade-I - divergence view looks good, temporal less so. Todo [x] check rooting [x] GitHub action for automation [x] Check validity of re-using clock rate from hmpxv1 (invalid - as I presumed - now using an inferred rate from clade-I samples when excluding clade-Ib) See this slack thread (private) for more review comments.
gharchive/pull-request
2024-05-27T05:03:31
2025-04-01T06:45:08.368591
{ "authors": [ "jameshadfield" ], "repo": "nextstrain/mpox", "url": "https://github.com/nextstrain/mpox/pull/254", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1455880703
its literally just a cow saying configure me! its not even a kahoot flooder there is none Same problem for me
gharchive/issue
2022-11-18T22:03:22
2025-04-01T06:45:08.372064
{ "authors": [ "drakeg77", "iamcubeman12" ], "repo": "nexuspcs/kspammer", "url": "https://github.com/nexuspcs/kspammer/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1361945227
Add Cancer Research UK Manchester Institute to the organisation list Add Cancer Research UK Manchester Institute to the organisation list. We also need a white version of the logo (to use for the dark theme) Nudge for @SPPearce in case you missed this ☝🏻 Yeah, I saw it, I just don't have a white version of the logo to hand. Ah ok - we have CRUK Beatson already, pretty sure I faffed around making their white logo years ago. I'll see if I can make a white version for you quickly in Illustrator 👍🏻 Added 👍🏻 Thanks Phil! I have no experience of illustrator to do it myself. Looks good though. On Thu, 29 Sep 2022, 18:20 Phil Ewels, @.***> wrote: Merged #1242 https://github.com/nf-core/nf-co.re/pull/1242 into master. — Reply to this email directly, view it on GitHub https://github.com/nf-core/nf-co.re/pull/1242#event-7488818718, or unsubscribe https://github.com/notifications/unsubscribe-auth/AF55TWI5BT5PULXAQ3UHHBTWAXFW5ANCNFSM6AAAAAAQE7MUJQ . You are receiving this because you were mentioned.Message ID: @.***>
gharchive/pull-request
2022-09-05T13:13:14
2025-04-01T06:45:08.388111
{ "authors": [ "SPPearce", "ewels", "mashehu" ], "repo": "nf-core/nf-co.re", "url": "https://github.com/nf-core/nf-co.re/pull/1242", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1641762132
Added Atoplex bed file As discussed in https://github.com/nf-core/viralrecon/issues/353 @ljmesi Hi! Could you review this so we can merge? :) not sure how this one wasn't merged after approval, I'll take responsibility for pushing the button
gharchive/pull-request
2023-03-27T09:41:34
2025-04-01T06:45:08.389656
{ "authors": [ "ctuni", "klkeys" ], "repo": "nf-core/test-datasets", "url": "https://github.com/nf-core/test-datasets/pull/813", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
203906444
json error? Hello, Homebridge/Pi newb here... i've been trying to get homebridge up and running with my Raspberry Pi 2B and my Nest thermostat. I've encountered a few issues but this one has me stumped, any suggestions would be greatly appreciated! Thanks in advance :) pi@raspberrypi:~ $ homebridge *** WARNING *** The program 'nodejs' uses the Apple Bonjour compatibility layer of Avahi. *** WARNING *** Please fix your application to use the native API of Avahi! *** WARNING *** For more information see http://0pointer.de/avahi-compat?s=libdns_sd&e=nodejs *** WARNING *** The program 'nodejs' called 'DNSServiceRegister()' which is not supported (or only supported partially) in the Apple Bonjour compatibility layer of Avahi. *** WARNING *** Please fix your application to use the native API of Avahi! *** WARNING *** For more information see http://0pointer.de/avahi-compat?s=libdns_sd&e=nodejs&f=DNSServiceRegister [1/29/2017, 11:38:49 PM] Loaded plugin: homebridge-nest [1/29/2017, 11:38:49 PM] Registering platform 'homebridge-nest.Nest' [1/29/2017, 11:38:49 PM] --- [1/29/2017, 11:38:49 PM] There was a problem reading your config.json file. [1/29/2017, 11:38:49 PM] Please try pasting your config.json file here to validate it: http://jsonlint.com [1/29/2017, 11:38:49 PM] /usr/lib/node_modules/homebridge/lib/server.js:197 throw err; ^ SyntaxError: Unexpected token   in JSON at position 453 at Object.parse (native) at Server._loadConfig (/usr/lib/node_modules/homebridge/lib/server.js:191:19) at new Server (/usr/lib/node_modules/homebridge/lib/server.js:52:38) at module.exports (/usr/lib/node_modules/homebridge/lib/cli.js:26:16) at Object. (/usr/lib/node_modules/homebridge/bin/homebridge:17:22) at Module._compile (module.js:570:32) at Object.Module._extensions..js (module.js:579:10) at Module.load (module.js:487:32) at tryModuleLoad (module.js:446:12) at Function.Module._load (module.js:438:3) at Module.runMain (module.js:604:10) at run (bootstrap_node.js:394:7) at startup (bootstrap_node.js:149:9) at bootstrap_node.js:509:3 My config file... { "bridge": { "name": "Homebridge", "username": "CC:22:3D:E3:CE:30", "port": 51826, "pin": "031-45-154" }, "description": "This is an example configuration file with one fake accessory and one fake platform. You can use this as a template for creating your own configuration file containing devices you actually own.", "accessories": [{ }], "platforms": [{ "platform": "Nest", "token" : "c.N9YZ4RqBUr2Aga", "clientId": "7c1e9dbe-7105-47d7-a6ae-2c1f********", "clientSecret": "guz5vIoE5IwF67SdJ********", "code": "66X", "username" : "EMAIL, "password" : "PASSWORD"   }] } The config for the accessories is wrong. Just write "accessories": [],. The {} are only used, if you really want to define an accessory. I had a similar issue. Mine was: [1/29/2017, 8:01:22 PM] Loaded plugin: homebridge-hue [1/29/2017, 8:01:22 PM] Registering platform 'homebridge-hue.Hue' [1/29/2017, 8:01:22 PM] --- [1/29/2017, 8:01:22 PM] There was a problem reading your config.json file. [1/29/2017, 8:01:22 PM] Please try pasting your config.json file here to validate it: http://jsonlint.com [1/29/2017, 8:01:22 PM] /usr/local/lib/node_modules/homebridge/lib/server.js:197 throw err; ^ SyntaxError: Unexpected token : in JSON at position 258 at Object.parse (native) at Server._loadConfig (/usr/local/lib/node_modules/homebridge/lib/server.js:191:19) at new Server (/usr/local/lib/node_modules/homebridge/lib/server.js:52:38) at module.exports (/usr/local/lib/node_modules/homebridge/lib/cli.js:26:16) at Object.<anonymous> (/usr/local/lib/node_modules/homebridge/bin/homebridge:17:22) at Module._compile (module.js:570:32) at Object.Module._extensions..js (module.js:579:10) at Module.load (module.js:487:32) at tryModuleLoad (module.js:446:12) at Function.Module._load (module.js:438:3) The config.json is { "bridge": { "name": "Homebridge", "username": "CENSORED", "port": CENSORED, "pin": "CENSORED" }, "description": "This is the configuration file for my HomeBridge.", "platforms": [ "platform": "Hue", "name": "Hue", "users": { "CENSORED": "CENSORED" } ] } If you have an accessory or platform, then each of them must be a JSON object enclosed by { and }. If not, is simply an empty array []. Please use the linter which is mentioned in the error message. It'll tell you, what's wrong. SgtHobNob try this config file with you info. { "bridge": { "name": "Homebridge", "username": "CC:22:3D:E3:CE:30", "port": 51826, "pin": "031-45-154" }, "description": "This is an example configuration file with one fake accessory and one fake platform. You can use this as a template for creating your own configuration file containing devices you actually own.", "accessories": [{ }], "platforms": [{ "platform": "Nest", "token": "c.N9YZ4RqBUr2Aga", "clientId": "7c1e9dbe-7105-47d7-a6ae-2c1f********", "clientSecret": "guz5vIoE5IwF67SdJ********", "code": "66X", "username": "EMAIL", "password": "PASSWORD" }] } josephswebpagegames try this config file with your info { "bridge": { "name": "Homebridge", "username": "CENSORED", "port": "CENSORED", "pin": "CENSORED" }, "description": "This is the configuration file for my HomeBridge.", "platforms": [{ "platform": "Hue", "name": "Hue" }] } Can you all enclose json code in three back ticks ` in extra lines please?
gharchive/issue
2017-01-29T23:53:26
2025-04-01T06:45:08.424685
{ "authors": [ "Neelrp12", "SgtHobNob", "josephswebpagegames", "maplesteve", "snowdd1" ], "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/1094", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
164559401
iOS 10 does NOT support homebridge iOS 10 beta does NOT currently support homebridge. The system is currently running fine, and is showing no errors in its console. On my iOS device, it is unable to add the device stating Couldn't Add Accessory Home couldn't connect to this accessory. Good username. Dupe? https://github.com/nfarina/homebridge/issues/662 Yup its a dupe, issue #662 address this issue. Please close the ticket. It's your ticket, I think you can close it. I am just an observer here! Nope don't think I can close, Any source Mods, please close this ticket. iOS 10 does support homebridge I have it running perfectly fine after a little bit of tweaking. Though I will look more into issue #662
gharchive/issue
2016-07-08T15:45:58
2025-04-01T06:45:08.427926
{ "authors": [ "AppleTechy", "IComplainInComments", "irrg" ], "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/699", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
168008288
add advance in journal entry related field Mr. Roby ask me to put an advance because he want to have relate advance in journal entry. The reason that i don't inherit in Advance module because if someone has to inherit that related field in JE, my inherit for advance will be replaced. Finally, i discuss with Mr. Roby and he decided to put in standard. ok krub
gharchive/pull-request
2016-07-28T03:48:22
2025-04-01T06:45:08.429039
{ "authors": [ "bank-netforce", "songpon" ], "repo": "nfco/netforce", "url": "https://github.com/nfco/netforce/pull/363", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2649832634
shuffle slides day 2.. re-shuffled slides a bit...: 53-exercise-starthere-part4.md: sharing the ARC 54-datahub-datasharing.md: some additional notes on ARC sharing, datahub terminolgy, etc. 55-exercise_fairproject_arc.md: create their own ARCs 60-data-publications.md: generics about publishing 65-exercise-starthere-part5.md: how to do it with dataplant @Mestizia
gharchive/pull-request
2024-11-11T16:41:57
2025-04-01T06:45:08.432002
{ "authors": [ "Brilator" ], "repo": "nfdi4plants/training-material", "url": "https://github.com/nfdi4plants/training-material/pull/6", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
663727690
Adapt playlist extraction to YouTubes architectual changes This PR addresses multiple errors with playlist video extraction. Fixes: #684 Fixes: #670 and possibly other issues. This PR is necessary, because YouTube recently changed something at their backand which changed the format of the playlist html responses, so the old way of extracting the youtube urls won't work anymore. YouTube is now relying on putting all relevant information inside a js variable and populating the site via html. The corresponding js var is window["ytInitialData"] and contains json data about anything which should be displayed on the page. The videos are now listed inside the json variable and can easily be extracted via json "tree-walking" as shown in the code. (I can add comments if neccessary, just add a change request thingie). Another thing that changed was the way continuations/pagination works now. Before the update, getting the urls from a playlist with more than 100 songs was fairly simple, because the page contained the url which contained the next page of videos. YouTube replaced this with a ajax api endpoint call, which needs a continuation token. This token can be extracted from the json as well as shown in the code. This ajax endpoint also needs some headers now (don't ask me why, I simpy found them by looking inside the developer tools and stripped the unnecessary). These are also included inside the code. This endpoint contains similar data to the data initially provided inside the html page of the playlist. This update by YouTube has a positive impact on performance of long playlists, because downloading a json is using way less time and bandwidth than downloading the entire website and extracting the urls via regex. @tooxo Thank you for the PR -- could you add to the description to provide context e.g. https://medium.com/@hugooodias/the-anatomy-of-a-perfect-pull-request-567382bb6067 :) Added some detail, please excuse my poor grammar I'm on mobile @RONNCC @tooxo thanks for change! -- would love assistance with setting up auto CI testing too 😉 thanks for the continued support! @RONNCC important addition to this PR: #691 (Also opened a ci related pr #690)
gharchive/pull-request
2020-07-22T12:34:54
2025-04-01T06:45:08.437534
{ "authors": [ "RONNCC", "dahmadjid", "tooxo" ], "repo": "nficano/pytube", "url": "https://github.com/nficano/pytube/pull/687", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
2533851602
backpressure? Does this library handle backpressure? Node's documentation talks about this.push being able to return false meaning stop pushing more items. The doesn't handle it. It seems to https://github.com/nfroidure/streamqueue/blob/main/src/index.ts#L72 but looks like there is no test for it. There are some this.push() where result isn't looked at when switching streams. So I agree it might be good enough. Anyway here is another implementation using a transformer which feels simpler to me. I just pipe the streams into this. import { Readable, Transform, TransformOptions } from 'stream'; import { EventEmitter } from 'events'; export class StreamQueue extends Transform { private _done: boolean; private ended: boolean; private currentStream: Readable | undefined private queue: Readable[] constructor(opts: TransformOptions) { super(opts); this.queue = []; this.currentStream = undefined; this._done = false; this.ended = false; } done(){ this._done = true; return this; } _destroy(err, callback) { for (let v of [this.currentStream, ...this.queue]) { v!.destroy(err) } } _transform(chunk: any, encoding: string, callback: any) { this.push(chunk); callback(); } // Add a readable stream to the queue add(...args: Readable[]) { if (this.ended) throw Error('ended') this.queue = [...this.queue, ...args] if (!this.currentStream) { this._processNextStream(); } } _processNextStream() { this.currentStream = this.queue.shift(); if (!this.currentStream){ if (this._done) { this.push(null); this.ended = true; } return; } this.currentStream.on('end', () => this._processNextStream()) this.currentStream.pipe(this) } } /* test: const buffers = ["a","b"].map((x) => Buffer.from(x)) const streams = buffers.map((x) => Readable.from(x)) const sq = new StreamQueue({objectMode: false}) sq.add(...streams) const a = await sq.toArray() console.log(a) // prints "a" and "b" as Buffers so seems to work. */ I could have missed some edge cases. So this issue turns into is the given implementation the most simple to get the job done ?
gharchive/issue
2024-09-18T14:06:48
2025-04-01T06:45:08.444330
{ "authors": [ "MarcWeber", "nfroidure" ], "repo": "nfroidure/streamqueue", "url": "https://github.com/nfroidure/streamqueue/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
229622002
[invalid] nodemon / npm run serve crashed on rebuild I'm submitting a ... (check one with "x") [x] bug report => check the README and search github for a similar issue or PR before submitting [ ] support request => check the README and search github for a similar issue or PR before submitting [ ] feature request Current behavior npm run serve crashes on npm run build:universal-prod or build:universal-dev. Expected/desired behavior I think it should not crash by rebuilding server.js Minimal reproduction of the problem with instructions My steps (win10): git clone https://github.com/ng-seed/universal.git cd universal npm install npm run build:universal-prod npm run serve (new console) npm run build:universal-prod Output: D:\path\universal>npm run serve > @ng-seed/universal@0.4.0-beta.2 serve D:\path\universal > npm run serve:universal > @ng-seed/universal@0.4.0-beta.2 serve:universal D:\path\universal > nodemon ./.server/server.js [nodemon] 1.11.0 [nodemon] to restart at any time, enter `rs` [nodemon] watching: *.* [nodemon] starting `node ./.server/server.js` Express server listening on http://localhost:8000 [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] restarting due to changes... [nodemon] starting `node ./.server/server.js` module.js:472 throw err; ^ Error: Cannot find module 'D:\path\universal\.server\serv er.js' at Function.Module._resolveFilename (module.js:470:15) at Function.Module._load (module.js:418:25) at Module.runMain (module.js:605:10) at run (bootstrap_node.js:423:7) at startup (bootstrap_node.js:147:9) at bootstrap_node.js:538:3 [nodemon] app crashed - waiting for file changes before starting... What is the motivation / use case for changing the behavior? Please tell us about your environment: Win 2012R2, NodeJS 7, current repo Angular version: 2.0.X 4.1 as in this repo. Browser: Console. Language: [all | TypeScript X.X | ES6/7 | ES5] Node (for AoT issues): node --version = @ortreum title says everything, it's all about nodemon. You can modify the package.json to force use your project using another tool than nodemon. Hello, @fulls1z3 fulls1z3 then please use a working version of nodemon in your package.json because I expect a usable version if I clone this project. I'm not that deep into this subject and can't just "use another tool". Michael
gharchive/issue
2017-05-18T10:20:04
2025-04-01T06:45:08.488541
{ "authors": [ "fulls1z3", "ortreum" ], "repo": "ng-seed/universal", "url": "https://github.com/ng-seed/universal/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
333712168
Unable to use ng-zorro-antd I'm submitting a ... (check one with "x") [ ] Regression (a behavior that used to work and stopped working in a new release) [x] Bug report <!-- Please check the repository for a similar issue or PR before submitting --> [x] Support request => <!-- Please check the repository for a similar issue or PR before submitting --> [ ] Feature request [ ] Documentation issue or request Current behavior Can't bind to 'nzCount' since it isn't a known property of 'nz-badge'. I've added the NgZorroAntdModule and everything but it's not recognizing it ... I've created to projects: Clear angular project generated with angular-cli and added NgZorroAntdModule according to the documentation https://github.com/vlados/ant-test second project created as a copy of the current status of ng-seed/universal using the documentation https://github.com/vlados/universal-test-ant You can see the project running with stackblitz https://stackblitz.com/github/vlados/ant-test https://stackblitz.com/github/vlados/universal-test-ant The problem is not releated to the project. When you have lazyModule you need to import the NgZorroAntdModule in the each lazy loaded module
gharchive/issue
2018-06-19T15:01:05
2025-04-01T06:45:08.493307
{ "authors": [ "vlados" ], "repo": "ng-seed/universal", "url": "https://github.com/ng-seed/universal/issues/442", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
241617132
Dependency upgrade This PR includes upgrade all dependencies to latest and the removal of reflect-metadata in favor of core-js This PR includes upgrade all dependencies to latest and the removal of reflect-metadata in favor of core-js I assume this was closed in favor of #410 That's correct. The only part that has really changed is the webpack-test.config.js. The CommonsChunkPlugin doesn't play nice with karma and I don't have a clean fix atm. Seems to be a common issue. For the moment I clear the inported plug-ins. Will find a better solution this weekend. By the way, we are using styleUrls and templateUrls, the angular2-template-loader plugin will change it to require at build time if your interested in a PR for that? :+1: Nice work. For my own project our app is pretty big so I've been avoiding bundling templates in our Angular.js app to help initial load times. If you have a smaller app (or have set up lazy loading) importing styleUrls makes sense to me. Idk what the maintainers would suggest for this starter.
gharchive/pull-request
2017-07-10T06:59:39
2025-04-01T06:45:08.504020
{ "authors": [ "aciccarello", "jbadeau" ], "repo": "ngParty/Angular1-scaffold", "url": "https://github.com/ngParty/Angular1-scaffold/pull/409", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
144618623
GEOWAVE-686: added an ingest format for avro encoded using the geowav… …e simplefeaturecollection schema this is branched off GEOWAVE-684 so either this can be pulled into GEOWAVE-684 first and then GEOWAVE-684 can be pulled into 0.9.1, or I can resubmit this PR into 0.9.1 if GEOWAVE-684 is first pulled into 0.9.1
gharchive/pull-request
2016-03-30T14:55:22
2025-04-01T06:45:08.505450
{ "authors": [ "rfecher" ], "repo": "ngageoint/geowave", "url": "https://github.com/ngageoint/geowave/pull/695", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2015723893
fix: use SecretBytes type for cert values to prevent accidental printing Proposed changes We save the values of the provided certs that we retrieve from Kubernetes secrets in the Certificates attribute on the Certificates struct. This is sensitive information that we want to make sure stays out of the logs and any stack traces. A common approach to this is to create a type definition for sensitive values that implements Stringer and JSON interfaces and cast the sensitive data to that value. Help I'm very new to Go. I may have made some non-idiomatic choices or failed to use features which could have improved this. Checklist Before creating a PR, run through this checklist and mark each as complete. [x] I have read the CONTRIBUTING document [x] If applicable, I have added tests that prove my fix is effective or that my feature works [x] If applicable, I have checked that any relevant tests pass after adding my changes [x] I have updated any relevant documentation (README.md and CHANGELOG.md) @4141done do you need anything from me to merge this? @4141done do you need anything from me to merge this? If you have time would you be able to run a test pass on it? That's what I keep not having time to go through. Otherwise I think we are good. Confirmed integration test works.
gharchive/pull-request
2023-11-29T03:22:48
2025-04-01T06:45:08.567005
{ "authors": [ "4141done", "ciroque" ], "repo": "nginxinc/nginx-loadbalancer-kubernetes", "url": "https://github.com/nginxinc/nginx-loadbalancer-kubernetes/pull/147", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
525021438
angular 9.0.0-rc.2 Cannot read property 'onDestroy' of undefined I'm using nrwl NX and jest. When I run jest I'm getting TypeError: Cannot read property 'onDestroy' of undefined node_modules/@ngneat/until-destroy/bundles/ng:/@ngneat/until-destroy/lib/until-destroy.ts:32:48 at Object.__decorate (../../../../node_modules/tslib/tslib.js:93:99) Has anything changed in the latests RC's that could cause this to happen? Update I forked and updated until-destroy to the latest angular rc release, but the tests work, so it hasn't got anything to do with angular/core. Nx uses the library jest-preset-angular and nx is currently using version 7.1.1 jest-preset-angular has released a version 8.0.0, but nx still hasn't updated to that version, so that might fix it. The only thing i can think of is that jest-preset-angular doesn't compile to ivy code and thus ɵcmp will be null. But I'm not sure how to get around it, or if I just have to wait for nx to catch up to angular 9. :( Hi, this is our peer dependencies requirements: "peerDependencies": { "@angular/core": ">=9.0.0-next.10" } So I guess you'll have to wait or use the older version. I have some issue when try run my app in storybook. I can't use old version, because i use Ivy. This lib really useful, but we had to remove because of this bug. We have the same issue in storybook using Ivy For some reason, Storybook doesn't use Angular Ivy with my setup and I had the same error. I added postinstall script with ngcc option to package.json according to the official docs and now everything works well. After making these changes you may need to delete node_modules (rm -rf) and install them again. P.S. to check if Ivy enabled you can import it from angular/core import { Component, OnInit, ɵivyEnabled as ivyEnabled, OnDestroy } from '@angular/core'; ngOnInit(): void { console.log('ivyEnabled', ivyEnabled); }
gharchive/issue
2019-11-19T13:49:14
2025-04-01T06:45:08.579308
{ "authors": [ "0x6368656174", "NetanelBasal", "akirpichnikov", "demike", "leon" ], "repo": "ngneat/until-destroy", "url": "https://github.com/ngneat/until-destroy/issues/42", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
778256812
API permission problem Hi, An API user can list all zones even if only one domain is listed on the ACL On the web interface the same user can see only his domain. A simple test : curl -sSfL --stderr - --request GET --header 'Content-Type: application/json' --header 'X-API-Key: API_KEY' --data '' https://dnsadmin/api/v1/servers/localhost/zones If you change the server from localhost to pdnsadmin then it should give you the correct domains, like it is written here: https://github.com/ngoduykhanh/PowerDNS-Admin/blob/2917c47fd1c78194d3462687b3a31c17c42e308e/powerdnsadmin/routes/api.py#L960 If you change the server from localhost to pdnsadmin then it should give you the correct domains, like it is written here: https://github.com/ngoduykhanh/PowerDNS-Admin/blob/2917c47fd1c78194d3462687b3a31c17c42e308e/powerdnsadmin/routes/api.py#L960 Thank you but i think it's a security issue because any user can use localhost and list any domains So I've added if server_id == 'pdnsadmin' or server_id == 'localhost': Thank you but i think it's a security issue because any user can use localhost and list any domains So I've added if server_id == 'pdnsadmin' or server_id == 'localhost': Well, actually I don't think this is a security breach. Any unauthenticated user can dig/nslookup to a dns server and see what domains/records are configured there, let alone this is an authenticated api. I think this is a misleading api, not a security problem. Well, actually I don't think this is a security breach. Any unauthenticated user can dig/nslookup to a dns server and see what domains/records are configured there, let alone this is an authenticated api. I think this is a misleading api, not a security problem. I guess that you're expecting this behaviour because you're distributing API keys to limit managers view over the API. It's reasonable to think that you may not want the see all domains on this endpoint because it's a bit disturbing and inconsistent with the idea that a user could list any domain and get an error trying to access it. In my setup I have set PDNS as a frontend to use API keys to limit user actions to a list of allowed domain. The changes will be made through Terraform to allow them to create A and PTR records when they pop up virtual instances on the delegated domains, so it uses the server/localhost endpoint. In my opinion it should be consistent accross the two endpoints. If there's no regression I think It's the way to go if it only requires a tiny change in the code. I guess that you're expecting this behaviour because you're distributing API keys to limit managers view over the API. It's reasonable to think that you may not want the see all domains on this endpoint because it's a bit disturbing and inconsistent with the idea that a user could list any domain and get an error trying to access it. In my setup I have set PDNS as a frontend to use API keys to limit user actions to a list of allowed domain. The changes will be made through Terraform to allow them to create A and PTR records when they pop up virtual instances on the delegated domains, so it uses the server/localhost endpoint. In my opinion it should be consistent accross the two endpoints. If there's no regression I think It's the way to go if it only requires a tiny change in the code. Yes, it's not a real security issue, but i would like the same view on web interface and on the API. It's more difficult to list and dig or nslookup each domain or they need to use an external service. Some account are used to create acme records for wildcard letsencrypt domain and some script check all the domain right and give error if the domain is not allowed. Generaly i'm creating an api access for each server which need to generate these entries. Before the last release, API and Web access gave the same view. Generaly on a API view we can see only resources than we're authorized to see or to modify. So it's logic to permit Adminstrator or Operator to see any domain but an user don't need to see it. if server_id == 'pdnsadmin': if g.apikey.role.name not in ['Administrator', 'Operator']: domain_obj_list = g.apikey.domains else: domain_obj_list = Domain.query.all() return json.dumps(domain_schema.dump(domain_obj_list)), 200 else: resp = helper.forward_request() return resp.content, resp.status_code, resp.headers.items() But if server is not 'pdnsadmin' user can list all domains We can see some restriction in the code: @api_bp.route('/pdnsadmin/zones', methods=['GET']) @api_basic_auth def api_login_list_zones(): if current_user.role.name not in ['Administrator', 'Operator']: domain_obj_list = get_user_domains() else: domain_obj_list = Domain.query.all() domain_obj_list = [] if domain_obj_list is None else domain_obj_list return json.dumps(domain_schema.dump(domain_obj_list)), 200 So i think their are missed checks on this release after the API changes. Yes, it's not a real security issue, but i would like the same view on web interface and on the API. It's more difficult to list and dig or nslookup each domain or they need to use an external service. Some account are used to create acme records for wildcard letsencrypt domain and some script check all the domain right and give error if the domain is not allowed. Generaly i'm creating an api access for each server which need to generate these entries. Before the last release, API and Web access gave the same view. Generaly on a API view we can see only resources than we're authorized to see or to modify. So it's logic to permit Adminstrator or Operator to see any domain but an user don't need to see it. if server_id == 'pdnsadmin': if g.apikey.role.name not in ['Administrator', 'Operator']: domain_obj_list = g.apikey.domains else: domain_obj_list = Domain.query.all() return json.dumps(domain_schema.dump(domain_obj_list)), 200 else: resp = helper.forward_request() return resp.content, resp.status_code, resp.headers.items() But if server is not 'pdnsadmin' user can list all domains We can see some restriction in the code: @api_bp.route('/pdnsadmin/zones', methods=['GET']) @api_basic_auth def api_login_list_zones(): if current_user.role.name not in ['Administrator', 'Operator']: domain_obj_list = get_user_domains() else: domain_obj_list = Domain.query.all() domain_obj_list = [] if domain_obj_list is None else domain_obj_list return json.dumps(domain_schema.dump(domain_obj_list)), 200 So i think their are missed checks on this release after the API changes. The API is in two parts: /pdnsadmin is for managing PDNSA /servers is just proxying requests to the PDNS Server. Furthermore, these are not the same endpoints: server/pdnsadmin/zones returns only zones allowed to the APIKey with their corresponding PDNS Admin id server/localhost/* forwards each subpath request to PDNS. We can filter the response of PDNS to allowed zones, that would we better thant changing the behavior. You can see it does not give the same amount of information: api/v1/servers/pdnsadmin/zones [ { "id": 1, "name": "test.com" }, { "id": 2, "name": "test.fr" }, { "id": 3, "name": "test.info" } ] api/v1/servers/localhost/zones [ { "account": "infra", "dnssec": false, "edited_serial": 2020120901, "id": "test.com.", "kind": "Master", "last_check": 0, "masters": [], "name": "dwadm.in.", "notified_serial": 2020120901, "serial": 2020120901, "url": "/api/v1/servers/localhost/zones/test.com." }, { "account": "infra", "dnssec": false, "edited_serial": 2020120702, "id": "deverywa.re.", "kind": "Master", "last_check": 0, "masters": [], "name": "test.fr.", "notified_serial": 2020120702, "serial": 2020120702, "url": "/api/v1/servers/localhost/zones/test.fr." }, { "account": "testing", "dnssec": false, "edited_serial": 2021010402, "id": "test.info.", "kind": "Master", "last_check": 0, "masters": [], "name": "test.info.", "notified_serial": 2021010402, "serial": 2021010402, "url": "/api/v1/servers/localhost/zones/test.info." } ] The API is in two parts: /pdnsadmin is for managing PDNSA /servers is just proxying requests to the PDNS Server. Furthermore, these are not the same endpoints: server/pdnsadmin/zones returns only zones allowed to the APIKey with their corresponding PDNS Admin id server/localhost/* forwards each subpath request to PDNS. We can filter the response of PDNS to allowed zones, that would we better thant changing the behavior. You can see it does not give the same amount of information: api/v1/servers/pdnsadmin/zones [ { "id": 1, "name": "test.com" }, { "id": 2, "name": "test.fr" }, { "id": 3, "name": "test.info" } ] api/v1/servers/localhost/zones [ { "account": "infra", "dnssec": false, "edited_serial": 2020120901, "id": "test.com.", "kind": "Master", "last_check": 0, "masters": [], "name": "dwadm.in.", "notified_serial": 2020120901, "serial": 2020120901, "url": "/api/v1/servers/localhost/zones/test.com." }, { "account": "infra", "dnssec": false, "edited_serial": 2020120702, "id": "deverywa.re.", "kind": "Master", "last_check": 0, "masters": [], "name": "test.fr.", "notified_serial": 2020120702, "serial": 2020120702, "url": "/api/v1/servers/localhost/zones/test.fr." }, { "account": "testing", "dnssec": false, "edited_serial": 2021010402, "id": "test.info.", "kind": "Master", "last_check": 0, "masters": [], "name": "test.info.", "notified_serial": 2021010402, "serial": 2021010402, "url": "/api/v1/servers/localhost/zones/test.info." } ] Provided patch can be used if @ngoduykhanh does not wish to include it. Provided patch can be used if @ngoduykhanh does not wish to include it. I agree with you guys. IMO, the result should be consistent with web UI and API. I merged the patch from @jbe-dw . I agree with you guys. IMO, the result should be consistent with web UI and API. I merged the patch from @jbe-dw .
gharchive/issue
2021-01-04T17:50:24
2025-04-01T06:45:08.605915
{ "authors": [ "Caez83", "RoeiGanor", "jbe-dw", "ngoduykhanh" ], "repo": "ngoduykhanh/PowerDNS-Admin", "url": "https://github.com/ngoduykhanh/PowerDNS-Admin/issues/856", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
242767384
Debugger connection check fails to work Hello! I'm running speedracer under Linux, and for some reason Chrome fails to open port 9222 for debugging if I don't use sudo. The problem is that speedracer actually outputs the issue: chrome-launcher chrome error: Error: connect ECONNREFUSED 127.0.0.1:9222 +4ms (at least with debug output enabled) but doesn't stop after that. Here's the sample output: DEBUG=* speedracer run race.js chrome-launcher locate chrome +0ms chrome-launcher trying: /Applications/Google Chrome Canary.app/Contents/MacOS/Google Chrome Canary +0ms chrome-launcher auto locate chrome +1ms chrome-launcher chrome found: /usr/bin/google-chrome +8ms chrome-launcher spawn chrome with flags: --headless,--remote-debugging-port=9222,--disable-gpu,--disable-translate,--disable-extensions,--disable-background-networking,--safebrowsing-disable-auto-update,--disable-sync,--metrics-recording-only,--disable-default-apps,--no-first-run,--user-data-dir=/tmp,--js-flags="--expose-gc" +1ms chrome-launcher check debugger ready: 1 +2ms chrome-launcher chrome error: Error: connect ECONNREFUSED 127.0.0.1:9222 +4ms chrome-launcher check debugger ready: 2 +501ms chrome-launcher chrome ready +1ms server start server +0ms server listening on port: 3000 +3ms server serving directory: /home/speedracer +0ms runner-server listening on port: 3001 +0ms driver load file race.js +38ms director file loaded: race.js +114ms director no races found: race.js +3s I assume stopping at failures like that and outputting them in non-debug mode could potentially save some time for other users. Thanks in advance! assume stopping at failures like that and outputting them in non-debug mode could potentially save some time for other users. Thanks in advance! You're totally right 😆, I'm addressing this issue and many others in 0.3.0 (very slowly but surely). Back to your problem... chrome-launcher spawn chrome with flags: --headless,--remote-debugging-port=9222,--disable-gpu,--disable-translate,--disable-extensions,--disable-background-networking,--safebrowsing-disable-auto-update,--disable-sync,--metrics-recording-only,--disable-default-apps,--no-first-run,--user-data-dir=/tmp,--js-flags="--expose-gc" +1ms This means that it did started Chrome. chrome-launcher check debugger ready: 1 +2ms chrome-launcher chrome error: Error: connect ECONNREFUSED 127.0.0.1:9222 +4ms chrome-launcher check debugger ready: 2 +501ms chrome-launcher chrome ready +1ms Then comes the a period when Speed Racer waits for the debugger to answer, it usually takes some time so it retries several times. Here it tries on time, ECONNREFUSED 127.0.0.1:9222 means the debugger can't be reached because it's not ready yet. Then it retries a second time 500ms later and logs chrome-launcher chrome ready +1ms. Which means Chrome is started and Speed Racer is talking to it. So your problem does not come from Chrome AFAIK. director no races found: race.js +3s It means that it can't find any race in /home/speedracer. Would you mind giving me the command line you typed and the tree structure of this directory? It might be related to #30 which has the same symptoms and matches with the date breaking changes have been introduced to rollup-stream. Closing this for now. Trying the 0.3.0 tag I end up with director no races found: src/ui/EmailField/race.js +3s instead Hey, sorry for the late answer here. Would you mind posting me the content of src/ui/EmailField/race.js? Thanks!
gharchive/issue
2017-07-13T16:59:20
2025-04-01T06:45:08.623539
{ "authors": [ "Wain-PC", "joaomilho", "ngryman" ], "repo": "ngryman/speedracer", "url": "https://github.com/ngryman/speedracer/issues/25", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
131892549
Release 1.2 I have the impression there are several small things that need to be done before release 1.2 and I'm afraid I may forget something so I would like to open this issue for reference. I also invite people to start looking for error, in the current master or in PR #128. [x] issue #61 [x] issue #127 [x] issue #118 [x] issue #124 [x] roll version [x] tutorials (in special issue #118) (make and add link) [ ] conda package (make and add instructions) [x] re-build doc [x] remove old select region in favor of SelectRegion_dev [x] pep8 [x] Update README to use the User Forum email for issues and discussions. (this list don't reflect execution order) +1 for the list! The one thing I think could wait is the select region stuff. This could be a sub version of 1.2 in my opinion. if you mean the SelectRegion_dev, this was kind of a question on you. But I would like to know why; what is missing in the dev version? Honestly I haven't looked at it in a while. Let me take a look and get back to you. @jjhelmus When we get this further along, could you help us build a conda package? I'd be happy to help get a conda package built for ARTView, just ping me when you make or are about to make the release and I can start testing out some recipes. Thanks! It should be pretty soon. Great, I'm traveling this week so might not be as prompt, but I'll keep an eye out. @gamaanderson What do you propose for the select region? I would like to keep the display functionality, but agree that having two separate methods will become problematic. what funcionalities of do you like to mantain, that you cannot do in one way or the other in with the new version? I think they are all there. @jjhelmus what would be the install instruction once the conda package is made? Run conda install -c jjhelmus artview. The jjhelmus part could be replaced with another conda "channel", that has just need where I've been putting the packages I've built. I believe the functionality is there, I meant that I'd like to maintain the setup from the Display as it is for now in terms of easily plotting histogram, bringing up table, applying a mask, etc. The newer select region is more modular probably and therefore in the long run more flexible. I just want to make sure we don't lose functionality for the user. This sort of goes back to a way back discussion I had with @scollis about open source radar or meteorology having a channel on conda. Would this be something to explore? @jjhelmus is the Guru here.. But yes.. this would be a great idea.. Nick wrote: This sort of goes back to a way back discussion I had with @scollis https://github.com/scollis about open source radar or meteorology having a channel on conda. Would this be something to explore? — Reply to this email directly or view it on GitHub https://github.com/nguy/artview/issues/129#issuecomment-187440150. @nguy since the code is already there, I have no problem to let the current SelectRegion accessible through the display. But it will become a part of display and lose its status as a components, therefore not accessible through the artview module or any mode. I know it's a balance, but I feel that this is a tool for quick data querying. Let's keep it for the time being, with the thought of any other tools developed should use the newer version. Do you think it wise to rename the code to indicate that it is display-based and not a component? For example, we could rename the class from SelectRegion to DisplaySelectRegion or something and have display prominent in all naming. We could then also rename SelectRegion_dev to SelectRegion. ok, I will rename it. Thanks @gamaanderson. Could you also replace the the contact in the README file to the user forum email? We can probably close some of the outstanding Issues that have been addressed as well. Let me know if you want me to do any of this. Okay, it appears that rebuilding docs and rolling version are the two parts that are left. I'll do that today and submit a PR. At this point we should probably then build the conda package and go forward with the release. If we manage to get this sorted, it would be fun to release on Monday since it's a leap year! Ended up pushing the new docs and rolling the rev directly into the master. @jjhelmus Can we get started with the conda release if you have time? Do you think we should do a "what new" in relation to the latest release? It is so much, that I have my doubts if it makes sense. I think that's a great idea. We could put this in the README maybe. Then I was going to post that What's New and announcement in the Users forum and the pyaos mailing list. The README is already too long! Ha, that is true! We should probably move some of this to the Documentation. @nguy I'll start working on the artview conda build this week. I'll keep you all updated if I run into any question. Thanks! @jjhelmus I just pushed a minor update. I don't think this affects the conda build, but just in case.... yes, it does afect. If the conda package is build from the old code then the tool would continue broken. I created test artview conda packages from the ae1331f56 commit which I labeled as "1.2.3". These are available on my Anaconda.org channel with the label "testing". If you use git tag to tag a release it will be easier to create the conda package for 1.2. Just let me know the tag/commit that should be considered 1.2 for the conda package and I can update the package quickly. Thank you a lot @jjhelmus @nguy PR #132 can be merged. I would also like to call attention to the ISRELEASED flag in setup.py line 57, in case you haven't already seen it. I want to make sure I follow what to do here. I can merge #132 (with the ISRELEASED flag turned on - thanks @gamaanderson I did not see this.) Apply the tag to it with something like conda release as description. Do I need to do a formal release via Github also? After all of this the "real" artview conda package will be released. Or do I need to just do step 2 and the rest can come afterward. Sorry, try to figure out the workflow here. You can do steps 3 and 4 anytime after after you change ISRELEASED to True and tag the release, there is no immediacy. Setting ISRELEASE to True indicates that commit corresponds to the "official" release version and tagging makes it easier to find rather than having to use the commit hash. It is good to make a second commit immediately after to flip ISRELEASED back to False and roll the version number. Examples from the last release of Py-ART: REL Release 1.6.0 MAINT: start v1.7+dev There are other workflows that projects use for this process that involves release branches, etc, but I've never found that necessary. Thanks @jjhelmus for walking us through this! Okay the tag is done. The method for Py-ART seems to be a good one to follow. @gamaanderson any thoughts? This would give us a benchmark for development. FYI, git push doesn't push tags by default, you need to use git push --tags. done, thanks FYI, this will close Issue #105 I believe. Also, @jjhelmus would you be willing to help with future conda releases? If so, we should ping you when appropriate? Happy to. I'll set up the system for building the conda package on conda-forge so that anyone can make a release by submitting a pull request. Great! You'll probably need to lead me through that as well eventually. We should sit and chat at SEA conference to discuss parallel developments. Sounds like a plan. I have always been in favor of following pyart example when applicable. Me too, let @jjhelmus do the hard work. ;p Okay back on topic, do we need anything else for the conda package then? A full suite of conda packages for artview 1.2 made and available on Anaconda.org. I need to fix a few items on conda-forge before moving the build there. Excellent, thank you again and let us know anything else we can do to help. I created a conda-forge feedstock for artview which allows for reproducible builds of the conda package using online CI services. Anyone can submit a pull request that repository when updates are needed. The command to install the packages produced by this setup is conda install -c https://conda.anaconda.org/conda-forge artview. I prefer this more open method for building but will continue to maintain the packages on my personal anaconda.org channel, jjhelmus. The official V1.2 release is in the wild! Realized I forgot to do the ISRELEASED. So pushed this as REL commit and updated the Release tag. Then pushed a v1.3dev commit afterwards. Can this be closed? Yep
gharchive/issue
2016-02-06T19:51:21
2025-04-01T06:45:08.664684
{ "authors": [ "gamaanderson", "jjhelmus", "nguy", "scollis" ], "repo": "nguy/artview", "url": "https://github.com/nguy/artview/issues/129", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2351017641
WiFi QR Code Hi, This is a wonderful system with great results. Thank you! The problem I am having is when using this class to generate a WiFi QR code. The result doesn't work because the password is missing. I used the below code: $ssid = 'My-Guest-Wifi'; $auth = 'WPA'; $pass = 'APassWord'; $qrcode = new QRCode('WIFI:S:'.$ssid.';T:'.$auth.';P:'.$pass.';;'); $qrcode->create('tmp/atest.png'); The result I got was this: However, when I do a straight cURL Post call to the API, the result includes the password and works perfectly. Here's the post fields included in the cURL call: {"data":"WIFI:S:My-Guest-Wifi;T:WPA;P:APassWord;","size":600,"download":true} So, am I doing something wrong in the code? There are no examples here on how to do WiFi QR codes, so I had to look at the how it was done on the qr-code-monkey website by looking at their JS file. But, since it works using cURL, I am guessing that there might be something happing with this class. Any help is appreciated. Steve Ok, So, I think I know what is happening. Instead of: $qrcode = new QRCode('WIFI:S:'.$ssid.';T:'.$auth.';P:'.$pass.';;'); I changed it to: $data = "WIFI:S:$ssid;T:$auth;P:$pass;"; $qrcode = new QRCode($data); And it worked! So, I think the key is to use double quotes instead of separating variables with single quotes. So, problem solved. Again, thanks for a great class! Steve
gharchive/issue
2024-06-13T12:27:26
2025-04-01T06:45:08.669849
{ "authors": [ "stemo40" ], "repo": "nguyenary/QRCodeMonkey", "url": "https://github.com/nguyenary/QRCodeMonkey/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1988825667
fix mappings of player and engine for reporting tournament results As you can see here, when I was Engine 0 and Player R it gave the win (and score) incorrectly to Engine 1 (i.e. position R). tournament = tilewe.engine.Tournament([ MichaelEngine("Michael 1"), tilewe.engine.RandomEngine("Random 2"), tilewe.engine.RandomEngine("Random 3"), tilewe.engine.RandomEngine("Random 4") ]) Michael is player red B: 45 ( I3 I4 Z4 O4 T4 I5 T5 V5 X5 Z5 ) Y: 41 ( I4 Z4 F5 I5 U5 V5 W5 X5 Y5 Z5 ) R: 80 ( I4 I5 ) G: 56 ( I4 O4 I5 V5 W5 X5 Z5 ) Finished: True Winner: R Game 1: winners: [1] scores: [41, 80, 45, 56] total wins: [0, 1, 0, 0] total scores: [41, 80, 45, 56] After the fix you can see even when Engine 0 is player G, it correctly gives the win and score to Engine 0: Michael is player green Game 1: winners: [0] scores: [89, 48, 45, 56] total wins: [1, 0, 0, 0] total scores: [89, 48, 45, 56] Ah yep, that makes sense
gharchive/pull-request
2023-11-11T08:03:17
2025-04-01T06:45:08.685252
{ "authors": [ "maconard", "nhamil" ], "repo": "nhamil/python-tilewe", "url": "https://github.com/nhamil/python-tilewe/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
481596207
JDBC sampling We just implemented DataFrame sampling per #71. But sometimes we would want to do sampling when we are reading data from DB, i.e. even before the DataFrame is created, as the whole data set may not fit in memory. The are a few possible ways to implement SqlLoader.sample(n): JDBC ResultSet has a roundabout API to check the size of the result: rs.last(); int size = rs.getRow(); rs.first(); // doesn't work with TYPE_FORWARD_ONLY ResultSets Aside from ResultSet type limitations, this looks like it may have its own memory inefficiencies (how does rewinding affect fetch size window?) . There's also a reservoir sampling algorithm that would hopefully allow to do that without JDBC tricks. As reservoir sampling loses the original ordering of the records, looks like we'll need to collect the record numbers and then sort the resulting sample by them. This changes the performance from O(N) to something worse, but since the sample size M is usually much smaller than N, it should still be O(N) at the extremes. Implemented reservior sampling with sampling API available for both TableLoader and SqlLoader.
gharchive/issue
2019-08-16T12:54:22
2025-04-01T06:45:08.688039
{ "authors": [ "andrus" ], "repo": "nhl/dflib", "url": "https://github.com/nhl/dflib/issues/73", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
433795264
fix: breaking change for timegridDisplayPrimayTime Please check if the PR fulfills these requirements [x] It's submitted to right branch according to our branching model [x] It's right issue type on title [ ] When resolving a specific issue, it's referenced in the PR's title (e.g. fix #xxx[,#xxx], where "xxx" is the issue number) [x] The commit message follows our guidelines [x] Tests for the changes have been added (for bug fixes/features) [x] Docs have been added/updated (for bug fixes/features) [x] It does not introduce a breaking change or has description for the breaking change Description Thank you for your contribution to TOAST UI product. 🎉 😘 ✨ ok to test
gharchive/pull-request
2019-04-16T13:57:12
2025-04-01T06:45:08.691299
{ "authors": [ "jungeun-cho" ], "repo": "nhn/tui.calendar", "url": "https://github.com/nhn/tui.calendar/pull/304", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
757851934
can i save the undo state in broswer's memory? Summary i use img-editor in my work. i want to save the use action in memory for fix user's mistakes eg. close broswer, refresh broswer. @qiphon Currently, the undo state is not saved separately in the browser. But I review it positively. canvas.tojsondata is a good api. but it has a little more things to do now
gharchive/issue
2020-12-06T05:29:55
2025-04-01T06:45:08.692877
{ "authors": [ "lja1018", "qiphon" ], "repo": "nhn/tui.image-editor", "url": "https://github.com/nhn/tui.image-editor/issues/494", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1115628569
Variables defined in .env.development not set in nhost_auth and nhost_storage containers I wrote an .env.development that looks like this: MY_SPECIAL_ENV_VAR=so contentful Then I ran nhost dev, and dumped the nhost_auth and nhost_storage containers' environments — MY_SPECIAL_ENV_VAR was not defined in either container. But it should be, right? (Oddly MY_SPECIAL_ENV_VAR was defined in the nhost_hasura container's environment 🤷 ) Environment variables will be available in: Hasura Serverless functions So I'd say it looks good. Ah, I see. Unless I'm misunderstanding, I believe this documentation will need updating. https://github.com/nhost/cli/#environment-variables https://github.com/nhost/cli/wiki/Environment#envdevelopment-file In both places, it suggests that the env variables will be available in all containers. Thanks for pointing out the bug in wiki. We will update it. I'm closing this issue now.
gharchive/issue
2022-01-27T00:15:42
2025-04-01T06:45:08.703756
{ "authors": [ "elitan", "mrinalwahal", "shyndman" ], "repo": "nhost/cli", "url": "https://github.com/nhost/cli/issues/141", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1344387914
feat: allow user to specify the path to look for/download hasura binary Should work like right now but allowing for a custom path to look for/download the hasura cli binary. I understand intention, but it won't fully work. Check these lines basically if default version from CLI (taken from docker tag of graphql-engine image) is newer than yours then the path you provide will be overriden with downloaded hasura binary is newer than yours then the path you provide will be overriden with downloaded hasura binary that's the intention. As per the description: "allowing for a custom path to look for/download the hasura cli binary"
gharchive/pull-request
2022-08-19T12:39:22
2025-04-01T06:45:08.705830
{ "authors": [ "dbarrosop", "elephant3" ], "repo": "nhost/cli", "url": "https://github.com/nhost/cli/pull/286", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
822892246
Update services and corresponding tests Task: AB#11053 There's still a bit of updated test coverage required here. I've done some manual tests but planning to work on filling in the unit test gaps today, all being well.
gharchive/pull-request
2021-03-05T09:03:02
2025-04-01T06:45:08.707010
{ "authors": [ "bjss-chris" ], "repo": "nhs-digital-gp-it-futures/BuyingCatalogueOrdering", "url": "https://github.com/nhs-digital-gp-it-futures/BuyingCatalogueOrdering/pull/310", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
433713568
Branch marked as Default is empty The master branch is marked as default (and therefore shows when you browse to the Repo), but this only contains a readme.md file. It might be helpful merging the latest / current branch into master. We don't use master branch.... guidance is that we use develop so best to make that one default. The repository has been updated to use 'develop' and 'master' only. Subsequent changes to the spec can be made in develop for review, prior to merging into 'master' for publication.
gharchive/issue
2019-04-16T10:50:55
2025-04-01T06:45:08.709157
{ "authors": [ "NaminderSoorma", "TimCoates", "bushra" ], "repo": "nhsconnect/NHS-Scheduling-API", "url": "https://github.com/nhsconnect/NHS-Scheduling-API/issues/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
404754033
Clean cookie code For a short term solution, the cookie consent code was taken directly from other nhs.uk solutions, with a simpler design applied to it. As the design is simpler and a user controls cookies with the browser (rather than in site settings), a lot of the code isn't required and can be simplified. The code is also currently compiled by webpack and isn't very readable in this repo. It needs to be changed to the pre-compiled code and assets. @bradmorton1 are you aware of these errors relating to the cookie banner javascript? Chrome console Uncaught SyntaxError: Unexpected token e in JSON at position 1 at JSON.parse (<anonymous>) at getCookie (cookieconsent.js:40) at checkCookie (cookieconsent.js:143) Build warnings [16:36:53] Finished 'minifyJS' after 1.56 s app/scripts/cookieConsent.js (269:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (283:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (299:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (315:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (331:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (345:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (361:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification app/scripts/cookieConsent.js (375:6) Use of eval is strongly discouraged, as it poses security risks and may cause issues with minification Yeah, the first error is present because it's dropping the cookie but doesn't need to contain most of the info in the cookie. The 2nd lot is because it's using compiled webpack code, and we probably don't want that. Can make improvements some time this week This was fixed in release 1.0.3
gharchive/issue
2019-01-30T12:45:13
2025-04-01T06:45:08.713343
{ "authors": [ "AdamChrimes", "bradmorton1" ], "repo": "nhsuk/nhsuk-service-manual", "url": "https://github.com/nhsuk/nhsuk-service-manual/issues/82", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }