id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
63900144
How to construct the query for this case? If I get 2 Realm classes: @interface Foo : RLMObject @end @interface Bar : RLMObject @property Foo *foo; @end How can I construct an query to find all Foo objects which all Bar objects point to. There's no shorthand way to do this, but the following code snippet should work: RLMRealm *realm; // Whatever your Realm is NSMutableArray *pointedToFoos = [NSMutableArray array]; for (Bar *bar in [Bar allObjectsInRealm:realm]) { if (bar.foo) { [pointedToFoos addObject:bar.foo]; } } What if I want to get an instance of as a result? Unfortunately, it's not currently possible to query directly on back links, so there's no way to get an RLMResults of that collection. Could you filter the Bar objects with an NSPredicate where self.foo != nil? Yes, you could. Given NSPredicate's syntax, it'd be nice to be able to do something like this: [[Bar objectsWhere:@"foo != nil"] valueForKey:@"foo"]; but we don't currently support that in Realm. But even if that worked, you'd get an NSArray of Foos, which wouldn't auto-update. The best thing you can do now if you need an auto-updating RLMResults is to follow @dmiedema's suggestion and query RLMResults *foos = [Bar objectsWhere:@"foo != nil"] and every time you need to access a Foo, use ((Bar *)foos[index]).foo.
gharchive/issue
2015-03-24T04:21:28
2025-04-01T06:45:36.542567
{ "authors": [ "dismory", "dmiedema", "jpsim", "segiddins" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/issues/1685", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
107374872
Errors in xcode after updating to iOS9 support I started to get these errors after updating latest xcode support for iOS I have added the screenshots http://imgur.com/a/Fw140 update Errors Pods/Realm/Realm/RLMAccessor.mm:248:46: Expected ')' Pods/Realm/Realm/RLMAccessor.mm:256:13: Use of undeclared identifier 'val' Pods/Realm/Realm/RLMAccessor.mm:257:37: Use of undeclared identifier 'val' Pods/Realm/Realm/RLMAccessor.mm:258:65: Use of undeclared identifier 'options'; did you mean 'optind'? Pods/Realm/Realm/RLMAccessor.mm:297:20: Redefinition of 'RLMSetValue' Pods/Realm/Realm/RLMAccessor.mm:693:13: No matching function for call to 'RLMSetValue' issue was fixed after updating the latest version sorry. close this thread Glad your resolved whatever this was, @nuke99!
gharchive/issue
2015-09-20T05:40:54
2025-04-01T06:45:36.546041
{ "authors": [ "jpsim", "nuke99" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/issues/2547", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
196973794
Feature request. File synchronization Could you at least give an idea how to implement this? Currently realm supports database synchronization only We're really focusing on synchronizing Realm databases rather than arbitrary files. There are certainly ways to build a more general purpose file synchronization algorithms on top of Realm's synchronization, but no out of the box solution. One obvious way is to store data in a Realm rather than another file format. This is an option if you're trying to synchronize say a document format that's used with your app (like .docx or .pages bundles). If you can't store the file data itself with Realm, you could also store the file metadata and actions. For example, every time a file operation is performed, record it in a side Realm, so you can track which files have been modified, in what way, in what order. Thanks for your response but there are some problems: 1)if I store files into Realm directly then the base may grow unpredictably. Additionally Realm has a limitation for file size size - they should be less than 15mb 2)if I store files outside Realm then I have troubles with file synchronization But anyways thanks for your response. We're working to address some file growth issues in realm/realm-core#2343, this is something we very much want to fix, so I encourage you to share more details there.
gharchive/issue
2016-12-21T16:22:36
2025-04-01T06:45:36.549202
{ "authors": [ "gerchicov-bp", "jpsim" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/issues/4475", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
513147870
pod update or use Swift Package Manager always failed Realm framework version: ? latest Realm Object Server version: ? Xcode version: ? Xcode11.1 iOS/OSX version: ? Mac OS 10.15 when I update RealmSwift, I use "pod update", but I get a error " Pre-downloading: Realm from https://github.com/realm/realm-cocoa.git, branch master, submodules true [!] Error installing Realm [!] Failed to download 'Realm': [!] /usr/bin/git clone https://github.com/realm/realm-cocoa.git /var/folders/7c/pc0hhjfj5s72q2ty1jrm6crc0000gn/T/d20191028-14811-dgthqb --template= Cloning into '/var/folders/7c/pc0hhjfj5s72q2ty1jrm6crc0000gn/T/d20191028-14811-dgthqb'... error: RPC failed; curl 18 transfer closed with outstanding read data remaining fatal: the remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed " Then, I use Swift Package Manager. I get a error " Cloning into '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore'... Submodule 'external/catch' (https://github.com/catchorg/Catch2) registered for path 'Realm/ObjectStore/external/catch' Cloning into '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore/external/catch'... error: RPC failed; curl 18 transfer closed with outstanding read data remaining fatal: the remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed fatal: clone of 'https://github.com/catchorg/Catch2' into submodule path '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore/external/catch' failed Failed to clone 'external/catch'. Retry scheduled Cloning into '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore/external/catch'... error: RPC failed; curl 18 transfer closed with outstanding read data remaining fatal: the remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed fatal: clone of 'https://github.com/catchorg/Catch2' into submodule path '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore/external/catch' failed Failed to clone 'external/catch' a second time, aborting Cloning into '/Users/sven/Library/Developer/Xcode/DerivedData/SPMTest-ditlgorglmvpzfebwatrofwqalay/SourcePackages/checkouts/realm-cocoa/Realm/ObjectStore/external/catch'... Failed to recurse into submodule path 'Realm/ObjectStore' " I can use other framework normally @yumengqing It looks like you are "just" having problems downloading? It's not clear how we can help with that? @bmunkholm spm is consistently failing on this package after running "Reset package cache". So although it might be network related, SEO is telling me there might be something ~unique~ specifically to this repo that is causing git to an edge case
gharchive/issue
2019-10-28T07:52:43
2025-04-01T06:45:36.558104
{ "authors": [ "CodeOcenS", "bmunkholm", "rromanchuk" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/issues/6317", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
849559071
Realm crash: terminating with uncaught exception of type realm::KeyNotFound: No such object Goals I want my app to work without crashes Expected Results I expected my app to not crash Actual Results Thread 10#0 0x00007fff6112f766 in kevent () #1 0x000000010423bcb9 in realm::util::network::Service::IoReactor::wait_and_activate(std::__1::chrono::time_point<std::__1::chrono::steady_clock, std::__1::chrono::duration<long long, std::__1::ratio<1l, 1000000000l> > >, std::__1::chrono::time_point<std::__1::chrono::steady_clock, std::__1::chrono::duration<long long, std::__1::ratio<1l, 1000000000l> > >) () #2 0x000000010423c342 in realm::util::network::Service::IoReactor::wait_and_advance(std::__1::chrono::time_point<std::__1::chrono::steady_clock, std::__1::chrono::duration<long long, std::__1::ratio<1l, 1000000000l> > >, std::__1::chrono::time_point<std::__1::chrono::steady_clock, std::__1::chrono::duration<long long, std::__1::ratio<1l, 1000000000l> > >, bool&, realm::util::network::Service::OperQueue<realm::util::network::Service::AsyncOper>&) () #3 0x000000010423cc95 in realm::util::network::Service::Impl::run() () #4 0x00000001041b331d in realm::sync::Client::run() () #5 0x000000010434ba4d in void* std::__1::__thread_proxy<std::__1::tuple<std::__1::unique_ptr<std::__1::__thread_struct, std::__1::default_delete<std::__1::__thread_struct> >, realm::_impl::SyncClient::SyncClient(std::__1::unique_ptr<realm::util::Logger, std::__1::default_delete<realm::util::Logger> >, realm::SyncClientConfig const&, std::__1::shared_ptr<realm::SyncManager const>)::'lambda0'()> >(void*) () #6 0x00007fff61167109 in _pthread_start () #7 0x00007fff61162b8b in thread_start () Thread 11#0 0x00007fff6112d882 in __psynch_cvwait () #1 0x00007fff61167425 in _pthread_cond_wait () #2 0x000000010423fc52 in realm::util::network::Service::Impl::resolver_thread() () #3 0x000000010423fb2d in void* std::__1::__thread_proxy<std::__1::tuple<std::__1::unique_ptr<std::__1::__thread_struct, std::__1::default_delete<std::__1::__thread_struct> >, realm::util::network::Service::Impl::add_resolve_oper(std::__1::unique_ptr<realm::util::network::Service::ResolveOperBase, realm::util::network::Service::LendersOperDeleter>)::'lambda'()> >(void*) () #4 0x00007fff61167109 in _pthread_start () #5 0x00007fff61162b8b in thread_start () Realm notification listener (12)#0 0x00007fff6112f766 in kevent () #1 0x0000000104246908 in realm::_impl::ExternalCommitHelper::listen() () #2 0x0000000104246a8e in void* std::__1::__thread_proxy<std::__1::tuple<std::__1::unique_ptr<std::__1::__thread_struct, std::__1::default_delete<std::__1::__thread_struct> >, realm::_impl::ExternalCommitHelper::ExternalCommitHelper(realm::_impl::RealmCoordinator&)::$_0> >(void*) () #3 0x00007fff61167109 in _pthread_start () #4 0x00007fff61162b8b in thread_start () CRASH HERE ===> Realm notification listener (13)#0 0x00007fff6113133a in __pthread_kill () #1 0x00007fff61166e60 in pthread_kill () #2 0x00007fff200fab94 in abort () #3 0x00007fff20253818 in abort_message () #4 0x00007fff20244e65 in demangling_terminate_handler() () #5 0x00007fff201780d9 in _objc_terminate() () #6 0x00007fff20252c47 in std::__terminate(void (*)()) () #7 0x00007fff202555b6 in __cxa_rethrow () #8 0x0000000104246b38 in void* std::__1::__thread_proxy<std::__1::tuple<std::__1::unique_ptr<std::__1::__thread_struct, std::__1::default_delete<std::__1::__thread_struct> >, realm::_impl::ExternalCommitHelper::ExternalCommitHelper(realm::_impl::RealmCoordinator&)::$_0> >(void*) () #9 0x00007fff61167109 in _pthread_start () #10 0x00007fff61162b8b in thread_start () Console: uncaught exception in notifier thread: N5realm11KeyNotFoundE: No such object libc++abi.dylib: terminating with uncaught exception of type realm::KeyNotFound: No such object terminating with uncaught exception of type realm::KeyNotFound: No such object Steps for others to Reproduce I have a pretty complex setup but I can provide all details if needed. The crash reproduces with 100% frequency. A brief explanation below. I have 2 Mongo Atlas databases. The first for public data to which all users have access. The second for user data. Public data is lazily loaded from the network using an insert trigger for one of the user collections. The app listens for a user object in the user database and for a container object in the public database using separate Realms. The container object is always present and has relationships to all other public data and just aggregates this data. The app observes that object. Realm function adds leaf objects and so the container data became available when relationships are resolved. So what I got from logs is that when I add an object to the user database I receive an update about that, then Realm trigger calls function to update 3 different tables with data when it received from the network and app crashes after that, probably on trying to fetch an update. It might be a race condition of some sort because the Realm function updates those objects concurrently. The app works fine after restart. I tried to terminate and start sync again without success. I tried to erase and fetch all data without success. I tried to deleted and install the app without success. Update 1: I tried to put 5 seconds delay between fetches but it didn't help so it looks like the problem is more general. Update 2: When I removed container object observation the app stopped to crash and works fine even after the restart when new data is populated. So the root cause seems to be the observation during relationship resolve from nil to some new object. App logs: 04.03 05:03:34.282 | R | Performing write 04.03 05:03:34.286 | R | User Realm update 04.03 05:03:34.287 | R | Deletions: [] | Insertions: [] | Modifications: [0] 04.03 05:03:34.287 | R | User objects count: 1 uncaught exception in notifier thread: N5realm11KeyNotFoundE: No such object 2021-04-03 05:03:35.514950+0300 Divtracker[74750:22811057] uncaught exception in notifier thread: N5realm11KeyNotFoundE: No such object libc++abi.dylib: terminating with uncaught exception of type realm::KeyNotFound: No such object terminating with uncaught exception of type realm::KeyNotFound: No such object Trigger logs: Logs: [ "04.03 02:03:34.626 | Symbol: EVV", "04.03 02:03:34.645 | Company for symbol 'EVV' not found. Fetching and fixing data...", "04.03 02:03:34.646 | Previous day price for symbol 'EVV' not found. Fetching and fixing data...", "04.03 02:03:34.731 | Previous day price for symbol 'EVV' not found. Fetching and fixing data...", "04.03 02:03:34.791 | Inserting company for symbol 'EVV'...", "04.03 02:03:34.791 | Company for symbol 'EVV' successfully inserted", "04.03 02:03:34.850 | Inserting previous day price for symbol 'EVV'...", "04.03 02:03:34.850 | Previous day price for symbol 'EVV' successfully inserted", "04.03 02:03:34.946 | Inserting dividends container for symbol 'EVV'...", "04.03 02:03:34.947 | Dividends for symbol 'EVV' successfully inserted" ] Sync logs: OK - Apr 03 5:03:35+03:00 - 61ms - SyncWrite Source: Write originated from MongoDB Logs: [ "Upload message contained 1 changeset(s)", "Integrating upload required conflict resolution to be performed on 0 of the changesets", "Latest server version is now 6" ] Partition: P Write Summary: { "CompanyRealmModel": { "inserted": [ "EVV" ] }, "DividendsContainerRealmModel": { "inserted": [ "EVV" ] }, "PreviousDayPriceRealmModel": { "inserted": [ "EVV" ] } } Code Sample I can provide Realm functions, schemes, objects, and code I use for objects observing later if needed. Version of Realm and Tooling ProductName: Mac OS X ProductVersion: 10.15.6 BuildVersion: 19G2021 /Applications/Xcode.app/Contents/Developer Xcode 12.4 Build version 12D4e /usr/local/bin/pod 1.10.1 Realm (10.7.2) RealmSwift (10.7.2) /bin/bash GNU bash, version 3.2.57(1)-release (x86_64-apple-darwin19) /usr/local/bin/carthage 0.37.0 (not in use here) /usr/bin/git git version 2.24.3 (Apple Git-128) The other thing I noted is that when I delete leaf objects they are no longer available when I use realm.objects(Leaf.self) but they still exist in the container object. Do I need to delete them manually? It doesn't look right though 😕 I reworked db not use any relationships and collecting everything manually as a workaround Hey @anton-plebanovich ! Thanks for providing all the information. Since you already noted you can reproduce this issue 100% I think it would be a good idea to just use that as a starting point. In case this information (code, repo, etc.) is not supposed to be posted publicly, can you send it to realm-help@mongodb.com so we can have a look at it? Thank you! Hey @DominicFrei , thanks for the response. I already reworked the database and the source code for the app but I'll try to set up a demo project you can use tomorrow. @anton-plebanovich Alright. :) Glad to hear your rework works for you. No rush with the repo then. Whenever you'll be able to put something together we'll sure appreciate it! 👍 @DominicFrei sadly, I can't reproduce it anymore I tried to recreate the same setup but everything just works right now. I think we can close it as not reproducible or I can try it during the week since it might be some weird race conditions related to the network or server load I can't capture right now. @anton-plebanovich Thank you for the feedback. I'll close it then and in case you stumble upon it again I'm happy to re-open as soon as we got more information. 👍 I'm hitting this same crash now, after moving my previously-fine app to the new @Persisted property wrapper. I don't have anything more to report currently, but just adding a data point. @bdkjones I have seen a similar issue where this exception occurs when a primary key value is nil but the property is declared as non nil. I had solved it by declaring my property as optional. e.g. @Persisted(primaryKey: true) var _id: ObjectId // becomes @Persisted(primaryKey: true) var _id: ObjectId? Maybe you are hitting something similar?
gharchive/issue
2021-04-03T02:24:34
2025-04-01T06:45:36.571500
{ "authors": [ "DominicFrei", "anton-plebanovich", "bdkjones", "leemaguire" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/issues/7198", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
156848815
Optimize accessor class creation In total this cuts the time to create the accessor classes in half. In absolute terms it's not an exciting speedup (it was only ~10 ms for our test schema before on an iPad Air running 9.1), but supporting schema changes while the app is running involves creating a lot more accessor classes, which lead to it starting to be an issue. This appears to break Realm Swift's reflection, at least in ObjectUtil.getOptionalProperties(_:). This is looking really good. I think the only things are getting the changelog entry to the right spot and realigning with the "managed/unmanaged" naming that was just merged. I've rebased this and brought it back up to date as I found myself wanting the refactored accessor creation code from this for the array of primitives stuff.
gharchive/pull-request
2016-05-25T20:58:19
2025-04-01T06:45:36.574250
{ "authors": [ "jpsim", "tgoyne" ], "repo": "realm/realm-cocoa", "url": "https://github.com/realm/realm-cocoa/pull/3651", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
58218548
follow xdg standard. fixes #122 Moves the rebar3 config (including templates and plugins) to ~/.config/rebar3 and the package cache to ~/.cache/rebar3 fixes #122 There's template docs that will need updates too with this: https://github.com/rebar/rebar3/blob/master/doc/templates.md#global-variables
gharchive/pull-request
2015-02-19T14:57:37
2025-04-01T06:45:36.675001
{ "authors": [ "ferd", "tsloughter" ], "repo": "rebar/rebar3", "url": "https://github.com/rebar/rebar3/pull/166", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
421776269
web application with many entity and large data with grid Hi all and pardon me because i'm a beginner in node js framework. imagine we have a system with 100 entity for data entry. Is there a sample or step by step video or article for easiest solutions and with lowest coding for making this project with these features : 1- create model(poco) layer and just write entity properties and its validation attribute for backend and frontend validations. 2- create a form automatic with a grid with these features : a) paging & sorting & filtering for each column header. b) data entry (insert and delete and update) for master entity. c) open each row by plus sign click for show details entities. d) add some custom buttons in rows for execute an action or api on server and database. 2- project can work with sqlite database or sqlserver. 3- project can give access or deny some web page and insert & delete & update &custom buttons. and I just write a query for create a ViewModel for grids. It sounds like you’re looking for a “data grid” - this library is for grid layout components
gharchive/issue
2019-03-16T06:30:50
2025-04-01T06:45:36.678059
{ "authors": [ "jxnblk", "mammadkoma" ], "repo": "rebassjs/grid", "url": "https://github.com/rebassjs/grid/issues/174", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2152458241
optimize animate_blob for filterquality currently, filterquality used for the spin blob of the loading image has its preset to high: https://github.com/rebels2638/ScoutingApp2024/blob/2e9d6a08899aca6e5c6563c42f30442e32787253/lib/blobs/animate_blob.dart#L36 i feel like either low or medium would work well here, see: https://api.flutter.dev/flutter/dart-ui/FilterQuality.html in reference to: https://api.flutter.dev/flutter/widgets/Image/filterQuality.html not useful
gharchive/issue
2024-02-24T19:40:28
2025-04-01T06:45:36.680615
{ "authors": [ "exoad" ], "repo": "rebels2638/ScoutingApp2024", "url": "https://github.com/rebels2638/ScoutingApp2024/issues/18", "license": "BSD-4-Clause", "license_type": "permissive", "license_source": "github-api" }
910816881
Has this extension been abandoned? I have a minor tweak I'd like to suggest, but it looks like this extension is no longer in development. Is it? Thanks I plan on coming back to it. What's your suggestion?
gharchive/issue
2021-06-03T20:08:13
2025-04-01T06:45:36.695081
{ "authors": [ "arandorion", "reblws" ], "repo": "reblws/tab-search", "url": "https://github.com/reblws/tab-search/issues/88", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
230223158
esc2text output error Hi, I'm trying to capture the output send by an Epson tm-t20ii printer and converting that to text. I am able to capture data with nc (or as i prefer a perl script). When i try to print that file using "cat tmp.bin > /dev/usb/lp0" i get a correct looking receipt. But when i try to run it through esc2text i get the following output : Click to expand pi@raspberrypi:~/pos/tmp $ cat output.txt WARNING: Unknown command (J WARNING: Unknown command (J WARNING: Unknown command (J WARNING: Unknown command P WARNING: Unknown command 8 WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command 0 WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command 0 WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command 0 WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command 1 WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command B WARNING: Unknown command WARNING: Unknown command B WARNING: Unknown command WARNING: Unknown command B WARNING: Unknown command WARNING: Unknown command B WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command B WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command P WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command ? WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ` WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command 8 WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command ? WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command WARNING: Unknown command @ WARNING: Unknown command @ WARNING: Unknown command WARNING: Unknown command ? L?w0p018? @?<@? @???????B@???@??@???L?0|?????? @???@?8p@??????D???B??????|??????p ?,A????PcP??@@0???0B?? ??@@@<0@? ??@@@?0 ?? B????@@@?? ? B@@@?@ ? $@@@P@ @ $?? 0B@@p@ @ ???B@@?_???~??|?????? ????? @@@?@?@???????<?@???? D"A?? @???? @B??q???????? @BD???? ?@@BD??????@?D???0 @A(????00 ~(????0???@???0@@??? @ p? @??A?? @ ? 0?@???????>???? ???? ?????????@?@?????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????>???????? @C?0@ @?? @B????@ @@? @B??B@ @@? @D?B@ @@BD? ?@ @@BD? P@ @?~(? ?@ B(? @ @?B @ @B?@ @ @B?@ @ @B??? @ @B?A 8 0?@@?????σ?>??/???> ?????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????@???>??????????????8 ? A? A@@ ?@@ ??? @?@@ @@?? @? @?@@ @@@? @? @?@@ @@@ ? @?@@ @@@ ? @?@@ @@@ ??????A@ @@@ ???@ ~@ @@@" ? B@ @@@? !@?? A@? @@? !@??@ @?@@ @@?@@ ????@ @?@@? ? A???p@ @@@????>??????????p????@??|??>??????????}?????? ??A ?? ? ? @?@???? @@? ??@???? @@!? ??? @@"? ??@? ?@ ? @@$? P@E ? ?@ @@(? ?@E <?? @@6? @I ?? ?? @@!? ? @I ? @@ ?? ? H?? @@ @?A? P? ? @@? @?A? P? ?@? ? ??? P?? ?@??8??~????? @????????????@?????>???|??????????????? A??@? ? ?? @? @@? ??? @?! @@ ? @?! @@!B0 @? @ @@!B0@? @ @@"??? @@ @@?0?@@? @? @@?0 @? @ @@ ? @? @ @@ !? @ @? @ @@? !?@ @? ? ? ` !??p@@?? ?????~??~??????????????? L,90p018???????????????????????????????????????????????????????????????????????????????????????????????????8|??>@ ??? ? ? @@@ ? @@?????? ? @@ ?? ? @@?@ ? @@ @ ? @ ?? ? @ @ ? @ ?@?? @ ?@ ? @@????@ ??0A@?? ???9????>p??>??????| ??@0@@@?G? ? H@ @!P@? ??! 0 ! @ ? @ ? @@@? @ ?? ??9??|????????????????????????????????????????????????????????????????????????????????????????????????????|????>@ ??? @ ???? @0@? G?? ?H@? !P@? ?? ! ? 0 ! ? @ ? ? @ ?@?@ ? @0 ??? 0???8??>????????|? ??? ? @ ?| 0@? @ @ @@?? @0 @@ @@ @@@??? ???????8??|?|??????|? ?? ? ? ??> @@@A? G?@@?? ?H@@A@ !P@@A@? ! @A??! @A? A @?@? ? @!?@?? 0??p?|??? The file in question, comes from a working windows POS, setup using correct epson drivers. When i compare the file to the receipt-with-logo.bin demo file, they seem to be totally different. Making me think i need to convert this more before running it through esc2text ? Any idea's ? tmp.bin.zip This is quite an interesting test file, thanks for posting it. It definitely shows some buggy behaviour. I'll work a bit with this file and get back to you. It looks like it's mostly raster data, which we can't extract yet, but I can certainly make some improvements based on what you've posted here. So perhaps the printer driver is rastering it. Is this known behavior for Epson printers ? I will try using generic text driver in an hour and update. Op 22 mei 2017 om 09:06 heeft Michael Billington <notifications@github.commailto:notifications@github.com> het volgende geschreven: This is quite an interesting test file, thanks for posting it. It definitely shows some buggy behaviour. I'll work a bit with this file and get back to you. It looks like it's mostly raster data, which we can't extract yet, but I can certainly make some improvements based on what you've posted here. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://github.com/receipt-print-hq/escpos-tools/issues/20#issuecomment-303016499, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AEU69W-_RiM2zCZvvKb8TTX8nmnM6kx7ks5r8TQBgaJpZM4NhnBe. I seem to pretty much the same output get using generic text driver Working on this in a branch called 'bugfix/unrecognised-commands', should be more like this- [DEBUG] SelectPeripheralDeviceCmd [DEBUG] UnknownDataCmd [DEBUG] UnknownDataCmd [DEBUG] UnknownDataCmd [DEBUG] UnknownCommandOneArg [DEBUG] UnknownCommandOneArg [DEBUG] SelectPaperEndSensorsCmd [DEBUG] SelectDefaultLineSpacingCmd [DEBUG] SelectInternationalCharacterSetCmd [DEBUG] SelectCodeTableCmd [DEBUG] EnableSmoothingCmd [DEBUG] CommandTwoArgs [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] SetAbsolutePrintPosCmd [DEBUG] GraphicsLargeDataCmd [DEBUG] GraphicsDataCmd [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] SetAbsolutePrintPosCmd [DEBUG] GraphicsLargeDataCmd [DEBUG] GraphicsDataCmd [DEBUG] PrintAndFeedCmd (LineBreak) [DEBUG] FeedAndCutCmd The receipt is certainly wrapped up as raster images, so you can't retrieve the content with this parser yet. Unpacking the content in thoseGraphicsLargeDataCmd calls (responsible for all the gibberish before) will happen as part of issue #6. Do you know of any POS printer driver that doesn't do this ? I can always try use another driver, and see if the Epson wings it I directly generate the receipts using mike42/escpos-php library, then use raw printing to deliver the data. I think the issue you were having is solved in the latest master branch. Can you please try running your test file through the tools again to verify? You might just need to activate the imagick extension if you don't have it yet. Expect text file to be created with some whitespace only, and a list of commands on the terminal: php escimages tmp.bin > tmp.txt Expect two images of parts of the receipt to be written in pbm format, with a copy of each in png format as well- php esc2text.php tmp.bin -v Expect HTML file to be written- php esc2html.php tmp.bin > tmp.html This output should appear visually similar to the actual receipt: I've tried it now and have gotten the images. Now is there any way we can go about extracting the text from those images without running each file through tesseract ? Yes, to extract text from images, you will need an OCR tool, since there's no extra data we can recover from the ESC/POS. If you need perfect output, you will need to train tesseract for the font being used. It sounds like you are familiar with it, but for future readers, I'll point out that the PBM files can be scooped up in alphabetical order and OCR'ed directly, no need to waste CPU cycles with PNG compression or TIFF conversion. # tesseract sudo apt-get install tesseract-ocr imagemagick rm *.pbm php escimages.php input_files/tmp.bin convert -append *.pbm pbm:- | tesseract - - cat out.txt Also, am I able to include this sample file in the repository, under the MIT license, for regression testing purposes? Note: Tried pdfsandwich as a second option, was slow and the output was poor quality. # pdfsandwich sudo apt-get install pdfsandwich wkhtmltopdf php esc2html.php input_files/tmp.bin > tmp.html wkhtmltopdf --page-size A7 tmp.html tmp.pdf pdfsandwich tmp.pdf pdftotext -layout tmp_ocr.pdf cat tmp_ocr.txt Closing this off, I believe this bug has now been resolved. I'm sorry for the late reply, i was busy on another task. Yes you can include the sample file, no problem. This issue is resolved yes, thanks 👍 💯 test.zip Sorry to jump in here, but I am having the same problem with my test file. When I do php escpos2text.php test.bin, I get the following output: ... WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command GS WARNING: Unknown command DLE NUL WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command FS WARNING: Unknown command ESC W PuTTYWARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W WARNING: Unknown command ESC W ... When I use: php escimages test.bin > test.txt, I get: Could not open input file: escimages The receipt is generated with the Seiko CAPD347 Driver, so the ESC/POS format should be correct ... but it looks different to the example provided. I used: php esc2html.php ascii.bin > output.html and i get a blank page with any text. so what's the problem? this is my file: ascii.txt
gharchive/issue
2017-05-21T14:07:04
2025-04-01T06:45:36.752333
{ "authors": [ "AlphaWarrioR", "KlustoR", "fatiezzahra", "mike42", "nsatech" ], "repo": "receipt-print-hq/escpos-tools", "url": "https://github.com/receipt-print-hq/escpos-tools/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
957681750
SwiftUI image not found on older iOS versions. dyld: Library not loaded: /System/Library/Frameworks/SwiftUI.framework/SwiftUI Referenced from: /var/containers/Bundle/Application/MyApp.app/MyApp Reason: image not found I worked around this by adding: -weak_framework SwiftUI to the Other Linter Flags in the Parchment build as suggested by Apple in the iOS 13 release notes: https://developer.apple.com/documentation/ios-ipados-release-notes/ios-13-release-notes Hi! Could you check if this is still a problem with the v3.1.0 release? But we don't commit Pods.xcodeproj, and then how can I save -weak_framework SwiftUI code changes? @torburg Here is a nice post install code you put at the bottom of your Podfile, which will do this every time you run pod install. post_install do |installer| installer.pods_project.targets.each do |target| if target.name == 'Parchment' target.build_configurations.each do |config| config.build_settings['OTHER_LDFLAGS'] = '-weak_framework SwiftUI' end end end end This should be fixed in the master branch now thanks to @kambala-decapitator 🙌 Will release a new version soon
gharchive/issue
2021-08-02T02:30:25
2025-04-01T06:45:36.789442
{ "authors": [ "Starsky89", "rechsteiner", "tkirby", "torburg" ], "repo": "rechsteiner/Parchment", "url": "https://github.com/rechsteiner/Parchment/issues/583", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2466857880
Add Profile: Rishav Kumar Sinha Profile Submission Please paste a screenshot of your GitHub profile here. Profile Category Please select one or more categories for your profile: [x] Badge 🎖️ [x] Minimalistic ✨ [ ] Dynamic 🔄 [x] Icons 🎯 [x] Backgrounds 😎 [ ] GIFS 🖼️ [ ] Game Mode 🚀 [ ] Code 👨‍💻 Next Steps Once you've filled out the information, please comment below: @all-contributors please add @<username> for review @all-contributors please add @RishavKumarSinha for review
gharchive/issue
2024-08-14T21:16:58
2025-04-01T06:45:36.794199
{ "authors": [ "RishavKumarSinha" ], "repo": "recodehive/awesome-github-profiles", "url": "https://github.com/recodehive/awesome-github-profiles/issues/360", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1115069113
[Annotations] Hide score for annotated labels in text classification Annotated labels should hide info about prediction score (%). Some cases could be confusing to the annotators. what about this issue ? cc @dvsrepo @Amelie-V @leiyre @dcfidalgo I remember we had a discussion about this when the score changed to 100% when selecting the label. Now it always stays the same reflecting the prediction score, which to me seems ok. Yes, let's close this I think
gharchive/issue
2022-01-26T14:17:28
2025-04-01T06:45:36.796228
{ "authors": [ "dcfidalgo", "dvsrepo", "frascuchon" ], "repo": "recognai/rubrix", "url": "https://github.com/recognai/rubrix/issues/1039", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1335689886
add specific host and port to launch ui when use rb.log import rubrix as rb rb.init(api_url="http://172.26.33.174:5555",) When i create rb.init above appear erorr : [Errno 111] Connection refused how to rb,init with specific host and port I can't run rubix on google colab Hi @Dean98AI and thanks for reporting. Is there a rubrix instance running on that location? Before you connect the server using the rb.init, you should run the server instance using the python -m rubrix command. You can follow the setup documentation https://rubrix.readthedocs.io/en/stable/getting_started/setup%26installation.html
gharchive/issue
2022-08-11T09:18:53
2025-04-01T06:45:36.798909
{ "authors": [ "Dean98AI", "frascuchon" ], "repo": "recognai/rubrix", "url": "https://github.com/recognai/rubrix/issues/1676", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1156403483
bundle: Add infrastructure-features annotation to the csv Signed-off-by: Nitin Goyal nigoyal@redhat.com https://bugzilla.redhat.com/show_bug.cgi?id=2058185 /lgtm /cherry-pick release-4.10
gharchive/pull-request
2022-03-02T04:58:36
2025-04-01T06:45:36.840585
{ "authors": [ "SanjalKatiyar", "agarwal-mudit", "iamniting" ], "repo": "red-hat-storage/odf-operator", "url": "https://github.com/red-hat-storage/odf-operator/pull/191", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1177559123
Fix paths of static files. Without the patch, the program does not seem to serve static files correctly. Thanks for the contributing! I will check it a bit later - looks like it could be fragile in case of malformed request (ie: without /) I can confirm this issue
gharchive/pull-request
2022-03-23T03:48:29
2025-04-01T06:45:36.862315
{ "authors": [ "WillyPillow", "coderofsalvation", "reddec" ], "repo": "reddec/trusted-cgi", "url": "https://github.com/reddec/trusted-cgi/pull/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
156822677
Encode and decode query parameters 👓 @ajacksified @nramadas :fish: :fish:
gharchive/pull-request
2016-05-25T18:47:16
2025-04-01T06:45:36.864194
{ "authors": [ "ajacksified", "nramadas", "schwers" ], "repo": "reddit/node-platform", "url": "https://github.com/reddit/node-platform/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
170332813
Removing the old container and volume during deployment Removing the old docs service container and volume during deployment. This fixes #511. Using Ansible's command module, instead of shell, when possible. lgtm 👍
gharchive/pull-request
2016-08-10T05:45:41
2025-04-01T06:45:36.880952
{ "authors": [ "davidalber", "doug-wade" ], "repo": "redfin/react-server", "url": "https://github.com/redfin/react-server/pull/516", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1593762351
Appstudio update test-component-pac-swdf Pipelines as Code configuration proposal Pipelines as Code CI/test-component-pac-swdf-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 6 seconds init ✅ Succeeded 18 seconds clone-repository ✅ Succeeded 22 seconds build-container ✅ Succeeded 12 seconds sanity-inspect-image ✅ Succeeded 12 seconds deprecated-base-image-check ✅ Succeeded 9 seconds clair-scan ✅ Succeeded 33 seconds clamav-scan ✅ Succeeded 9 seconds sbom-json-check ✅ Succeeded 9 seconds sanity-label-check ✅ Succeeded 9 seconds sanity-optional-label-check ✅ Succeeded 6 seconds show-summary Pipelines as Code CI/test-component-pac-swdf-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 6 seconds init ✅ Succeeded 18 seconds clone-repository ✅ Succeeded 22 seconds build-container ✅ Succeeded 12 seconds sanity-inspect-image ✅ Succeeded 12 seconds deprecated-base-image-check ✅ Succeeded 33 seconds clamav-scan ✅ Succeeded 9 seconds clair-scan ✅ Succeeded 9 seconds sbom-json-check ✅ Succeeded 9 seconds sanity-label-check ✅ Succeeded 9 seconds sanity-optional-label-check ✅ Succeeded 6 seconds show-summary
gharchive/pull-request
2023-02-21T16:22:33
2025-04-01T06:45:36.900554
{ "authors": [ "redhat-appstudio-qe-bot1" ], "repo": "redhat-appstudio-qe/devfile-sample-hello-world", "url": "https://github.com/redhat-appstudio-qe/devfile-sample-hello-world/pull/3071", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1875877146
Appstudio update hacbs-test-project-1kqe Pipelines as Code configuration proposal To start the PipelineRun, add a new comment with content /ok-to-test For more detailed information about running a PipelineRun, please refer to Pipelines as Code documentation Running the PipelineRun To customize the proposed PipelineRuns after merge, please refer to Build Pipeline customization Pipelines as Code CI has failed.There was an error creating the PipelineRun: hacbs-test-project-1kqe-on-pull-request- creating pipelinerun hacbs-test-project-1kqe-on-pull-request- in namespace rhtap-demo-fmrs-tenant has failed. Tekton Controller has reported this error: the namespace of the provided object does not match the namespace sent on the request
gharchive/pull-request
2023-08-31T16:34:24
2025-04-01T06:45:36.903726
{ "authors": [ "rhtap-qe-bots" ], "repo": "redhat-appstudio-qe/hacbs-test-project", "url": "https://github.com/redhat-appstudio-qe/hacbs-test-project/pull/8348", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2182693371
RHTAPSRE-467: Support the configuration of provider URL When onboarding self-hosted SCM PAC can't determine the the base url, so it should be provided explicitly. Allow users to mention the provider (SCM) URL using an annotation on the component. /test application-service-e2e thanks for the review @johnmcollier. I fixed the formatting. @johnmcollier I didn't understand from the test logs if it failed because of my change or because of an infra issue/flaky test. Can you please help me debug it? /retest @gbenhaim The e2e test failure seems to have been a flake that's resolved itself. I see some flakiness in our controller tests as well, so I've retested those to confirm @gbenhaim why cannot we retrieve git provider URL from the Component source URL? @mmorhun I followed the configuration fields PAC provides. Since PAC doesn't retrieve the SCM host for self-hosted SCMs from the repository url, I don't think we should do it as well. Nothing promises that the hostname of the SCM API server can be computed from the repository url, for example in a case where cloning is done through a proxy.
gharchive/pull-request
2024-03-12T21:20:19
2025-04-01T06:45:36.907107
{ "authors": [ "gbenhaim", "johnmcollier", "mmorhun" ], "repo": "redhat-appstudio/application-service", "url": "https://github.com/redhat-appstudio/application-service/pull/450", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1175047658
Add script for testing building via AppStudio Script for secret creation and building via application and components by AppStudio. Replacement of m2-builds.sh after merge of https://github.com/redhat-appstudio/application-service/pull/74 I have an issue with my environment such that application creation fails... error: timed out waiting for the condition on applications/test-application maybe this script should abort if that happens and does not attempt to create the components?
gharchive/pull-request
2022-03-21T08:24:15
2025-04-01T06:45:36.909360
{ "authors": [ "Michkov", "scoheb" ], "repo": "redhat-appstudio/infra-deployments", "url": "https://github.com/redhat-appstudio/infra-deployments/pull/195", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1433099148
fix: add fix to make test error Signed-off-by: Sri Vignesh sselvan@redhat.com @srivickynesh the fix looks good, but can you modify the commit title to reference the HACBS-1256 ticket?
gharchive/pull-request
2022-11-02T12:53:43
2025-04-01T06:45:36.910515
{ "authors": [ "dirgim", "srivickynesh" ], "repo": "redhat-appstudio/integration-service", "url": "https://github.com/redhat-appstudio/integration-service/pull/64", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2046186621
feat: improve RHTAPBUGS Impact view and add filtering AD-147 We want to improve the view in QD of RHTAPBUGS Impact Add columns of fields Title from Jira Created Date Closed Date Add filtering of the labels as in Jira: we add a new column Labels a user can use to filter Bugs in the view the filter supports sub-string, which means if user start writing it will display bugs with label contains the entered sub-string entered . if a user enters open it will display all bugs with label got a sub-string of open /lgtm
gharchive/pull-request
2023-12-18T09:39:41
2025-04-01T06:45:36.913677
{ "authors": [ "flacatus", "kasemAlem" ], "repo": "redhat-appstudio/quality-dashboard", "url": "https://github.com/redhat-appstudio/quality-dashboard/pull/254", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1379655643
OLM-installed controller-manager get OOMKilled The patch operator controller-manager pod is get OOMKilled due to memory usage. On the node, we see: Sep 20 15:54:40 ctl-1 kernel: oom-kill:constraint=CONSTRAINT_MEMCG,nodemask=(null),cpuset=crio-370504c90f749d6c5b1a4dc36ee9404d2467c669e01a56f30ac77c1f238c313a.scope,mems_allowed=0-1,oom_memcg=/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod966588f9_b8c8_4eda_a907_dd30fe21972e.slice/crio-370504c90f749d6c5b1a4dc36ee9404d2467c669e01a56f30ac77c1f238c313a.scope,task_memcg=/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod966588f9_b8c8_4eda_a907_dd30fe21972e.slice/crio-370504c90f749d6c5b1a4dc36ee9404d2467c669e01a56f30ac77c1f238c313a.scope,task=manager,pid=2231452,uid=1000760000 Sep 20 15:54:40 ctl-1 kernel: Memory cgroup out of memory: Killed process 2231452 (manager) total-vm:8090736kB, anon-rss:500676kB, file-rss:25904kB, shmem-rss:0kB, UID:1000760000 pgtables:1976kB oom_score_adj:999 This is when trying to apply a single, minimal patch; you can see the complete test in https://github.com/larsks/patch-operator-test. This seems like a bug; nothing we're doing here should result in substantial memory consumption. If we wanted to try increasing the resource limits for the pod, is there a way to influence the resources section of the deployment? Or is our only alternative to deploy without using OLM and edit the manifests directly? Is there anything we can do to help figure out why we're hitting memory limits? please see this: https://github.com/redhat-cop/patch-operator#patch-controller-performance-considerations and this: https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md#resources Thanks. I had already seen the first document, but we only have a single patch defined on the entire cluster (and the cluster isn't even hosting user workloads yet, so the number of total secrets isn't huge [approx 2300 right now]). I'm surprised we're hitting limits in that situation. I'm following the second document to increase the resource allocation for the controller pod. Increasing the resource quota appears to have addressed the problem.
gharchive/issue
2022-09-20T16:01:36
2025-04-01T06:45:36.922159
{ "authors": [ "larsks", "raffaelespazzoli" ], "repo": "redhat-cop/patch-operator", "url": "https://github.com/redhat-cop/patch-operator/issues/48", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
648278460
Update naming to follow gatekeeper format OPA Gatekeeper suggests a structure: https://github.com/open-policy-agent/gatekeeper/tree/master/library which konstraint also uses: https://github.com/plexsystems/konstraint#how-template-and-constraint-naming-works I am thinking it is a good idea to follow that. cc @tylerauerbeck @pabrahamsson @springdo This also means splitting the current policies out more, which is probably a good thing and something we'd of needed to do eventually. CC @ckavili Applied via PR: https://github.com/redhat-cop/rego-policies/pull/69
gharchive/issue
2020-06-30T15:36:37
2025-04-01T06:45:36.925376
{ "authors": [ "garethahealy" ], "repo": "redhat-cop/rego-policies", "url": "https://github.com/redhat-cop/rego-policies/issues/67", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1126955596
Remove unused go imports What type of PR is this: /kind cleanup @feloy I am curious how you cleaned up unused packages? I forgot to do it from the previous PR (https://github.com/redhat-developer/odo/pull/5433) /approve /lgtm
gharchive/pull-request
2022-02-08T08:47:15
2025-04-01T06:45:36.930180
{ "authors": [ "feloy", "valaparthvi" ], "repo": "redhat-developer/odo", "url": "https://github.com/redhat-developer/odo/pull/5444", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1124127284
[BUG] Unable to mount NFS with Fedora 35 What happened: I attempted to deploy an application with a PVC that directly referenced an NFS PersistentVolume. The PVC and PersistentVolume were bound correctly, however when the kubelet tried to mount it gives the following error: Warning FailedMount 15s (x9 over 2m23s) kubelet MountVolume.SetUp failed for volume "plex-data" : mount failed: exit status 32 Mounting command: mount Mounting arguments: -t nfs <my-nas>:/volume2/plex /var/lib/kubelet/pods/<pod-uuid>/volumes/kubernetes.io~nfs/plex-data Output: mount: /var/lib/kubelet/pods/<pod-uuid>/volumes/kubernetes.io~nfs/plex-data: bad option; for several filesystems (e.g. nfs, cifs) you might need a /sbin/mount.<type> helper program. What you expected to happen: Mounting with NFS works on Fedora 35 and Microshift How to reproduce it (as minimally and precisely as possible): Create a PersistentVolume that mounts an NFS share: Create a PVC that binds directly to the PersistentVolume Deploy a Pod which uses the PVC as a volume Anything else we need to know?: Environment: Microshift version (use microshift version): 4.8.11 Hardware configuration: x86_64 OS (e.g: cat /etc/os-release): NAME="Fedora Linux" VERSION="35 (Server Edition)" ID=fedora VERSION_ID=35 VERSION_CODENAME="" PLATFORM_ID="platform:f35" PRETTY_NAME="Fedora Linux 35 (Server Edition)" ANSI_COLOR="0;38;2;60;110;180" LOGO=fedora-logo-icon CPE_NAME="cpe:/o:fedoraproject:fedora:35" HOME_URL="https://fedoraproject.org/" DOCUMENTATION_URL="https://docs.fedoraproject.org/en-US/fedora/f35/system-administrators-guide/" SUPPORT_URL="https://ask.fedoraproject.org/" BUG_REPORT_URL="https://bugzilla.redhat.com/" REDHAT_BUGZILLA_PRODUCT="Fedora" REDHAT_BUGZILLA_PRODUCT_VERSION=35 REDHAT_SUPPORT_PRODUCT="Fedora" REDHAT_SUPPORT_PRODUCT_VERSION=35 PRIVACY_POLICY_URL="https://fedoraproject.org/wiki/Legal:PrivacyPolicy" VARIANT="Server Edition" VARIANT_ID=server Kernel (e.g. uname -a): 5.15.17-200.fc35.x86_64 Others: Relevant Logs $ oc describe pod plex-6d88457c5-pf96x Name: plex-6d88457c5-pf96x Namespace: plex Priority: 0 Node: k3s001-kaplan.lan/192.168.86.20 Start Time: Fri, 04 Feb 2022 06:37:40 -0500 Labels: app.kubernetes.io/instance=plex app.kubernetes.io/name=plex pod-template-hash=6d88457c5 Annotations: <none> Status: Pending IP: IPs: <none> Controlled By: ReplicaSet/plex-6d88457c5 Containers: plex: Container ID: Image: ghcr.io/k8s-at-home/plex:v1.24.1.4931-1a38e63c6 Image ID: Port: 32400/TCP Host Port: 0/TCP State: Waiting Reason: ContainerCreating Ready: False Restart Count: 0 Limits: memory: 2Gi Requests: cpu: 200m memory: 256Mi Liveness: tcp-socket :32400 delay=0s timeout=1s period=10s #success=1 #failure=3 Readiness: tcp-socket :32400 delay=0s timeout=1s period=10s #success=1 #failure=3 Startup: tcp-socket :32400 delay=0s timeout=1s period=5s #success=1 #failure=30 Environment: ADVERTISE_IP: http://<IP_REDACTED>:32400/ TZ: America/New_York Mounts: /config from config (rw) /data from data (rw) /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-d8dc7 (ro) Conditions: Type Status Initialized True Ready False ContainersReady False PodScheduled True Volumes: config: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: plex-config ReadOnly: false data: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: plex-data ReadOnly: false kube-api-access-d8dc7: Type: Projected (a volume that contains injected data from multiple sources) TokenExpirationSeconds: 3607 ConfigMapName: kube-root-ca.crt ConfigMapOptional: <nil> DownwardAPI: true ConfigMapName: openshift-service-ca.crt ConfigMapOptional: <nil> QoS Class: Burstable Node-Selectors: <none> Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s node.kubernetes.io/unreachable:NoExecute op=Exists for 300s Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning FailedScheduling 10m default-scheduler running PreBind plugin "VolumeBinding": binding volumes: timed out waiting for the condition Normal Scheduled 2m23s default-scheduler Successfully assigned plex/plex-6d88457c5-pf96x to <HOST> Warning FailedMount 20s kubelet Unable to attach or mount volumes: unmounted volumes=[data], unattached volumes=[config data kube-api-access-d8dc7]: timed out waiting for the condition Warning FailedMount 15s (x9 over 2m23s) kubelet MountVolume.SetUp failed for volume "plex-data" : mount failed: exit status 32 Mounting command: mount Mounting arguments: -t nfs <NFS_HOST>:/volume2/plex /var/lib/kubelet/pods/9746b74d-8c45-42a2-9960-fabe21448e87/volumes/kubernetes.io~nfs/plex-data Output: mount: /var/lib/kubelet/pods/9746b74d-8c45-42a2-9960-fabe21448e87/volumes/kubernetes.io~nfs/plex-data: bad option; for several filesystems (e.g. nfs, cifs) you might need a /sbin/mount.<type> helper program. Note that mounting directly on the host works just fine. Hmmm, I have a couple of questions: Is that a containerised install? or AIO? In AIO it would make sense if the nfs packages aren't installed. is nfs-utils installed on the host?, I suspect we may need to add that dependency to the rpm spec at least. Thanks for the report @adambkaplan sorry to see this so late. So I was able to get this working by using upstream's NFS CSI Driver. The host has nfs-utils installed, but in my deployment Microshift is running as a container, and its base image (ubi-minimal) doesn't have nfs-utils. I wrote a blog post on how to do this: https://adambkaplan.com/post/2022-02-06-plex-on-microshift/ Deploying the NFS CSI driver on containerized MicroShift seem to work. Closing this bug.
gharchive/issue
2022-02-04T12:02:43
2025-04-01T06:45:36.954430
{ "authors": [ "adambkaplan", "mangelajo", "oglok" ], "repo": "redhat-et/microshift", "url": "https://github.com/redhat-et/microshift/issues/599", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2271052657
Added Kaoto-bot as reviewer for the Kaoto operator Thanks for submitting your Operator. Please check the below list before you create your Pull Request. New Submissions [ ] Are you familiar with our contribution guidelines? [ ] Are you familiar with our operator pipeline? [ ] Have you tested your Operator with all Custom Resource Definitions packaging? [ ] Have you tested your Operator in all supported installation modes? [ ] Have you considered whether you want to use semantic versioning order? [ ] Is your submission signed? [ ] Is operator icon set? Your submission should not [ ] Add more than one operator bundle per PR [ ] Modify any operator [ ] Rename an operator [ ] Modify any files outside the above mentioned folders [ ] Contain more than one commit. Please squash your commits. Operator Description must contain (in order) [ ] Description of the managed Application and where to find more information [ ] Features and capabilities of your Operator and how to use it [ ] Any manual steps about potential pre-requisites for using your Operator Operator Metadata should contain [ ] Human readable name and 1-liner description about your Operator [ ] Valid category name1 [ ] One of the pre-defined capability levels2 [ ] Links to the maintainer, source code and documentation [ ] Example templates for all Custom Resource Definitions intended to be used [ ] A quadratic logo Remember that you can preview your CSV here. -- 1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need 2 For more information see here /pipeline restart operator-hosted-pipeline There is a bug in the pipeline affecting PRs that do not affect new bundles. I'll manually merge this as the change is trivial and the fix will require some time. Thanks!
gharchive/pull-request
2024-04-30T10:08:03
2025-04-01T06:45:36.965265
{ "authors": [ "mporrato", "oscerd" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/4428", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
732160507
More indications in the screenshots It is a general comment but in the screenshots, it could be very helpful to highlight where attendees need to focus on, to click on or to focus on especially when the whole console is captured. Sorry for the late response, we've since added clear highlighting in the screenshots to make actions clear, thank you!
gharchive/issue
2020-10-29T09:45:03
2025-04-01T06:45:36.972784
{ "authors": [ "cedricclyburn", "mcouliba" ], "repo": "redhat-scholars/openshift-starter-guides", "url": "https://github.com/redhat-scholars/openshift-starter-guides/issues/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1122809688
Unsubscribe handling in async Redis responds to an unsubscribe with one or many replies, depending on the current subscribe state. When channel/pattern names are provided in the unsubscribe command each given name will trigger a reply even if duplicated or not subscribed to. To know when we can switch the internal subscribe state we need to make sure all pending replies have been processed before doing the switch. This is done by bookkeeping pending additional unsubscribe replies. Fixes #1039 The alternative solution would be to only send unsubscribe to channels/patterns that is subscribed to, i.e. change the command content, which is a bit more intrusive. We still need to bookkeep which channels/patterns that we have sent an unsubscribe to. Nice improvement, thanks!
gharchive/pull-request
2022-02-03T08:51:12
2025-04-01T06:45:36.974816
{ "authors": [ "bjosv", "michael-grunder" ], "repo": "redis/hiredis", "url": "https://github.com/redis/hiredis/pull/1047", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1171025144
Support COMMAND DOCS Description COMMAND DOCS Closes #1934 Checklist [ ] Does npm test pass with this change (including linting)? [ ] Is the new or changed code fully tested? [ ] Is a documentation update included (if this change modifies existing APIs, or introduces new ones)? Codecov Report Merging #2043 (04d1232) into master (be51abe) will decrease coverage by 0.36%. The diff coverage is 45.45%. :exclamation: Current head 04d1232 differs from pull request most recent head 832519d. Consider uploading reports for the commit 832519d to get more accurate results @@ Coverage Diff @@ ## master #2043 +/- ## ========================================== - Coverage 94.83% 94.46% -0.37% ========================================== Files 360 361 +1 Lines 3366 3377 +11 Branches 411 412 +1 ========================================== - Hits 3192 3190 -2 - Misses 89 100 +11 - Partials 85 87 +2 Impacted Files Coverage Δ packages/client/lib/commands/COMMAND_DOCS.ts 40.00% <40.00%> (ø) packages/client/lib/client/commands.ts 100.00% <100.00%> (ø) packages/client/lib/errors.ts 95.45% <0.00%> (-4.55%) :arrow_down: packages/client/lib/client/commands-queue.ts 76.47% <0.00%> (-3.93%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update be51abe...832519d. Read the comment docs.
gharchive/pull-request
2022-03-16T13:28:23
2025-04-01T06:45:36.986322
{ "authors": [ "Avital-Fine", "codecov-commenter" ], "repo": "redis/node-redis", "url": "https://github.com/redis/node-redis/pull/2043", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1327723921
Is map type unordered? Spec suggests encoding json as an example and json is unordered, however it doesn't explicitly states this. Ordered hash maps would complicate implementation for some languages, like go(map) or python(dict). copy from https://github.com/antirez/RESP3/issues/30 Let me point out Python dict is specified as ordered (since Python 3). @dumblob see this comment https://github.com/antirez/RESP3/issues/30#issuecomment-557050925
gharchive/issue
2022-08-03T20:01:12
2025-04-01T06:45:36.988906
{ "authors": [ "dumblob", "leibale" ], "repo": "redis/redis-specifications", "url": "https://github.com/redis/redis-specifications/issues/6", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
529211640
redisson 中的锁,怎么通过Redis Desktop Manager中看到 // 多久会自动释放, int leaseTime = 60; int waitTime = 5; RLock rLock = redisson.getLock(one.getId()); try { boolean flag = rLock.tryLock(waitTime, leaseTime, TimeUnit.SECONDS); if (flag) { log.info("取到锁"); executorThread.submit(new SinglePipelineContinueRunExecutor(customPipelineHistory)); rLock.unlock(); log.info("释放锁"); } } catch (Exception e) { log.error(e.getMessage()); } finally { rLock.unlock(); } redisson 中的锁,怎么通过Redis Desktop Manager中看到 Expected behavior Actual behavior Steps to reproduce or test case Redis version Redisson version 3.11.5 Redisson configuration 普通的RLock是以HASH的形式保存的 RLock rLock = redisson.getLock(one.getId()); one.getId() 就是key 啊
gharchive/issue
2019-11-27T09:10:35
2025-04-01T06:45:37.009895
{ "authors": [ "jackygurui", "kakukeme", "walkersing" ], "repo": "redisson/redisson", "url": "https://github.com/redisson/redisson/issues/2435", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
546760591
How to build from source and package as single jar without all optional dependency? Is your feature request related to a problem? Please describe. Redission collection is used in my application (single jar) and want to package needed classes without all optional dependencies to reduce app size. Describe the solution you'd like Split feature (optional dependencies are used) as sub module. Describe alternatives you've considered Remove all optional dependencies java classes from my jar (Follow dependency document in wiki). But I don't think it is right approach and whether have potential problems. there is redisson.jar already built from sources. Why can't you use it? Do you mean redission-all.jar? I don't use spring, hibernate and other additional features. And I want to package redission classes with all must have dependencies into single jar. When I trying build redission by maven shade plugin, I got a single jar file which includes all optional dependencies, for example javax.cache.*, org.jodd, net.bytebuddy, io.projectreactor,io.reactivex.rxjava2. I want to remove them to reduce file size. But I cann't do this through maven build. If I don't use liveobject service, how to remove them? I tried to remove theme from src tree, but I get compile error, some classes are used by others, such as RedissonObjectBuilder. Even I keep them, if we don't include org.jodd/net.bytebuddy, I cann't build successfully. Can I remove these dependencies without errors? Finally, I want to to use redission distributed collections by sync way/async way. Reactive way/ Rx way will be removed and two ways dependencies will removed at same time.
gharchive/issue
2020-01-08T10:04:47
2025-04-01T06:45:37.013440
{ "authors": [ "kitepad", "mrniko" ], "repo": "redisson/redisson", "url": "https://github.com/redisson/redisson/issues/2527", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1951749572
RediSearch: Issues in FT.AGGREGATE implementation Steps to reproduce or test case Try to call the aggregate API passing the withCursor option: AggregationOptions aggregateOptions = AggregationOptions.defaults().load("$").withCursor(10); ... AggregationResult aggregationResult = redissonSearch.aggregate("index", "*", aggregateOptions); Expected behavior The resulting Redis query should like like this: "FT.AGGREGATE" "index" "*" "LOAD" "1" "$" "WITHCURSOR" "COUNT" 10 Actual behavior The resulting Redis query looks like this: "FT.AGGREGATE" "index" "*" "LOAD" "1" "$" "SORTBY" "0" "WITHCURSOR" Redis version redis_version:6.2.12 ... module:name=search,ver=20609,api=1,filters=0,usedby=[],using=[ReJSON],options=[handle-io-errors] module:name=ReJSON,ver=20407,api=1,filters=0,usedby=[search|graph],using=[],options=[handle-io-errors] module:name=timeseries,ver=10810,api=1,filters=0,usedby=[],using=[],options=[handle-io-errors] module:name=graph,ver=21010,api=1,filters=0,usedby=[],using=[ReJSON],options=[] module:name=bf,ver=20405,api=1,filters=0,usedby=[],using=[],options=[] Redisson version 3.23.4 Redisson configuration default Additional Information There seem to be several issues in the aggregateAsync implementation: https://github.com/redisson/redisson/blob/b0f8bb756002c50672a51a6059837411b5b6a2b7/redisson/src/main/java/org/redisson/RedissonSearch.java#L507C39-L507C53 Issue 1: There is no COUNT clause after WITHCURSOR The issue is at line 573. We should check on getCursorCount instead of getCount. if (options.getCount() != null) { args.add("COUNT"); args.add(options.getCount()); } should be if (options.getCursorCount() != null) { args.add("COUNT"); args.add(options.getCursorCount()); } Issue 2: There is a SORTBY clause in the query, even though no sorting options were provided The issue is at line 544. The condition should be inverted. if (options.getSortedByFields().isEmpty()) { should be if (!options.getSortedByFields().isEmpty()) { Issue 3: MAXIDLE option is not honored The issue is at line 578. We should add MAXIDLE instead of COUNT. if (options.getCursorMaxIdle() != null) { args.add("COUNT"); args.add(options.getCount()); } should be if (options.getCursorMaxIdle() != null) { args.add("MAXIDLE"); args.add(options.getCursorMaxIdle()); } Fixed. Thanks for report
gharchive/issue
2023-10-19T09:45:01
2025-04-01T06:45:37.022062
{ "authors": [ "kgopin", "mrniko" ], "repo": "redisson/redisson", "url": "https://github.com/redisson/redisson/issues/5381", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2160654459
在使用redssion(3.16.1)框架过程中 遇见问题 org.redisson.client.RedisException: ERR bad lua script for redis cluster, all the keys that the script uses should be passed using the KEYS array, and KEYS should not be in expression. channel: [id: 0x1de471a8, L:/192.168.18.52:41918 - R:r-2ze4114ca3030d64pd.redis.rds.aliyuncs.com/39.97.25.119:6565] command: (EVAL), params: [local insertable = false; local v = redis.call('hget', KEYS[1], ARGV[5]); if v == false then inserta..., 8, DF:test:cancel:queue1127, redisson__timeout__set:{DF:test:cancel:queue1127}, redisson__idle__set:{DF:test:cancel:queue1127}, redisson_map_cache_created:{DF:test:cancel:queue1127}, redisson_map_cache_updated:{DF:test:cancel:queue1127}, redisson__map_cache__last_access__set:{DF:test:cancel:queue1127}, redisson_map_cache_removed:{DF:test:cancel:queue1127}, {DF:test:cancel:queue1127}:redisson_options, ...] https://github.com/redisson/redisson/issues/1538#issuecomment-826583188
gharchive/issue
2024-02-29T08:06:54
2025-04-01T06:45:37.026550
{ "authors": [ "Eason0803", "mrniko" ], "repo": "redisson/redisson", "url": "https://github.com/redisson/redisson/issues/5660", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
454075499
Fixed bug with set RedissonHttpSessionConfiguration.class keyPrefix When i integration spring session, in RedissonSessionRepository.class, the created event can't be trigger, if set keyPrefix on RedissonHttpSessionConfiguration.class. Cause event listener add on construct, at the same time, add create event listener apply keyPrefix. However, RedissonSessionRepository.class keyPrefix set on RedissonHttpSessionConfiguration.class. Code fragment RedissonSessionRepository.class createdTopic = redisson.getPatternTopic(getEventsChannelPrefix() + "*", StringCodec.INSTANCE); RedissonHttpSessionConfiguration.class RedissonSessionRepository repository = new RedissonSessionRepository(redissonClient, eventPublisher, keyPrefix); if (StringUtils.hasText(keyPrefix)) { repository.setKeyPrefix(keyPrefix); } Please retain changes related to bug fix only
gharchive/pull-request
2019-06-10T09:20:21
2025-04-01T06:45:37.028965
{ "authors": [ "hs20xqy", "mrniko" ], "repo": "redisson/redisson", "url": "https://github.com/redisson/redisson/pull/2155", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
772924314
Create values-schema mechanism to update values to conform to schema changes User story As a developer making breaking changes to the values-schema, I want to have a mechanism that will copy the corresponding values to their new locations. Problem We break the schema when: We move properties around (more frequent) We delete props (more frequent) We change property type (less frequent) otomi-core should have a mechanism that migrates values to conform to the new spec. Proposed solution A schema will get a version (semver, patches are additions, minors are breaking changes) and a changes property holding information about breaking changes: The new location (if changed) by mapping old to new The list of deleted props The new type (if changed) by giving a go template for transformation (see schema change below): Note: the schema version has NO relationship with otomi-core's version. Workflow: Every time a developer changes the schema which includes these changes as DOD, the otomi migrate-values command should be performed that will massage the values to migrate to the new structure. Schema change: Sample part: version: 0.23.7 changes: locations: charts.bla.someProp: someNewRootProp.someProp deletions: - charts.bla.someOtherProp mutations: # image tag went from semver to glob charts.bla.image.tag: 'printf "v%s"' Tasks: core: [ ] Enrich values-schema.yaml [x] changes property [ ] a lint function to ensure that making a breaking change will be checked before commit [x] otomi migrate-values-rev to show all history from previous versions from version control. [ ] Create otomi migrate-values script that migrates otomi-values forward based on changes defined, and sets latest version in values. [x] Connect otomi-tasks to bin/migrate-values.sh and bin/otomi migrates-values [x] Load otomi-values file as json with fileName as argument to start modification per file [ ] Create a mechanism that will be able to pin an otomi-values repository to a certain version so the modification can take place [x] getNewVersion() mock [ ] getNewVersion() [x] getOldVersion() mock [ ] getOldVersion() [ ] otomi-values modification [x] displacements () [ ] deletions () [ ] mutations () [ ] CLI functionality api: [ ] Add extra call to tools server just before deploy: migrate-values Definition Of Done [ ] Tasks are done [ ] (Unit) Tests are added to code [ ] Refactoring: [ ] Rewire exported functions [ ] Type safety for typescript [ ] (Architecture) Design Record(s) have been added as adr/*.md and appended to list in adr/_index.md [ ] Specs and demo files have been updated to reflect code changes [ ] Documentation has been updated (docs/lifecycle-management/versioning) [ ] Functionality, code, and/or documentation has been peer reviewed [ ] Relevant team members have been notified Copied from unassigned issues to link it with pull request for changes in otomi-core. A schema will get a version (semver, patches are additions, minors are breaking changes) Let's just refer to SemVer 2.0.0 then for clarity? https://semver.org thanks for all the hard work and congrats with the lessons learned, but this darling will be closed in favor of our simplifying efforts ;) Reopening as we need this to migrate values forward automatically. Parking this once more as this seems more complex seen the different files involved, and knowing we don't have much need for it yet. I have some new insights. I think we maybe should reconsider how we will approach this problem. I was reading Designing Data-Intensive Applications, and the problem domain is "schema evolution". Schema evolution has solutions in other serialisation/encoding formats. Thrift/Protobuf's solution is described like so: As you can see from the examples, an encoded record is just the concatenation of its encoded fields. Each field is identified by its tag number (the numbers 1 , 2 , 3 in the sample schemas) and annotated with a datatype (e.g., string or integer). If a field value is not set, it is simply omitted from the encoded record. From this you can see that field tags are critical to the meaning of the encoded data. You can change the name of a field in the schema, since the encoded data never refers to field names, but you cannot change a field’s tag, since that would make all existing encoded data invalid. You can add new fields to the schema, provided that you give each field a new tag number. If old code (which doesn’t know about the new tag numbers you added) tries to read data written by new code, including a new field with a tag number it doesn’t recognize, it can simply ignore that field. The datatype annotation allows the parser to determine how many bytes it needs to skip. This maintains forward compatibility: old code can read records that were written by new code. What about backward compatibility? As long as each field has a unique tag number, new code can always read old data, because the tag numbers still have the same meaning. The only detail is that if you add a new field, you cannot make it required. If you were to add a field and make it required, that check would fail if new code read data written by old code, because the old code will not have written the new field that you added. Therefore, to maintain backward compatibility, everyof the schema must be optional or have a default value. Removing a field is just like adding a field, with backward and forward compatibility concerns reversed. That means you can only remove a field that is optional (a required field can never be removed), and you can never use the same tag number again (because you may still have data written somewhere that includes the old tag number, and that field must be ignored by new code). Datatypes and schema evolution What about changing the datatype of a field? That may be possible check the documentation for details but there is a risk that values will lose precision or get truncated. For example, say you change a 32-bit integer into a 64-bit integer. New code can easily read data written by old code, because the parser can fill in any missing bits with zeros. However, if old code reads data written by new code, the old code is still using a 32-bit variable to hold the value. If the decoded 64-bit value won’t fit in 32 bits, it will be truncated. A curious detail of Protocol Buffers is that it does not have a list or array datatype, but instead has a repeated marker for fields (which is a third option alongside required and optional ). As you can see in Figure 4-4 , the encoding of a repeated field is just what it says on the tin: the same field tag simply appears multiple times in the record. This has the nice effect that it’s okay to change an optional (single-valued) field into a repeated (multi-valued) field. New code reading old data sees a list with zero or one elements (depending on whether the field was present); old code reading new data sees only the last element of the list. Thrift has a dedicated list datatype, which is parameterized with the datatype of the list elements. This does not allow the same evolution from single-valued to multi-valued as Protocol Buffers does, but it has the advantage of supporting nested lists. Ie., keys can be referenced with tags and their meaning does not depend on their "nest" in the schema. The Apache Avro and Parquet have also rethought schema evolution but are specific to their problem domains. Draw your own conclusions, but it seems kinda hacky to move properties around by nest to retain their meaning. since we are only evolving forward I have no problem with our chosen approach...this is about automating a version to a newer one and never having to know what was the state before Can you do this after the KinD delivery @svatwork To be honest, I don't think this story is realistic and I hope we can have a chat whether this will really solve a problem. why wouldn't it be? I observe that we create, delete and move props, and transform values. Can we not cover those use cases in the approach we designed?
gharchive/issue
2020-12-22T12:53:20
2025-04-01T06:45:37.047963
{ "authors": [ "Morriz", "svatwork" ], "repo": "redkubes/otomi-core", "url": "https://github.com/redkubes/otomi-core/issues/259", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
955663195
vault/kubernetes-external-secrets stays in CrashLoopBackOff when installing CE with chart When installing Otomi in CE mode with chart, vault/kubernetes-external-secrets pod stays in CrashLoopBackOff. cert-manager is using production certs Logs: npm info it worked if it ends with ok │ npm info using npm@6.14.8 │ npm info using node@v12.19.0 │ npm info lifecycle kubernetes-external-secrets@6.1.0~prestart: kubernetes-external-secrets@6.1.0 │ npm info lifecycle kubernetes-external-secrets@6.1.0~start: kubernetes-external-secrets@6.1.0 │ > kubernetes-external-secrets@6.1.0 start /app │ > ./bin/daemon.js │ │ {"level":30,"time":1627551639807,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"loading kube specs"} │ {"level":30,"time":1627551641414,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"successfully loaded kube specs"} │ {"level":30,"time":1627551641415,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"updating CRD"} │ {"level":30,"time":1627551641415,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"Upserting custom resource externalsecrets.kubernetes │ {"level":30,"time":1627551641687,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"successfully updated CRD"} │ {"level":30,"time":1627551641691,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"starting app"} │ Thu, 29 Jul 2021 09:40:41 GMT kubernetes-client deprecated .getStream use .getObjectStream, see https://github.com/godaddy/kubernetes-client/blob/master/mergi │ {"level":30,"time":1627551641695,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"successfully started app"} │ {"level":30,"time":1627551641695,"pid":17,"hostname":"kubernetes-external-secrets-7c6d45f95-r6cbs","msg":"MetricsServer listening on port 3001"} │ internal/streams/legacy.js:61 │ throw er; // Unhandled stream error in pipe. │ ^ │ Error: read ECONNRESET │ at TLSWrap.onStreamRead (internal/stream_base_commons.js:208:20) { │ │ errno: 'ECONNRESET', │ code: 'ECONNRESET', │ syscall: 'read' │ } │ npm info lifecycle kubernetes-external-secrets@6.1.0~start: Failed to exec start script │ │ npm ERR! code ELIFECYCLE │ npm ERR! errno 1 │ npm ERR! kubernetes-external-secrets@6.1.0 start: `./bin/daemon.js` │ npm ERR! Exit status 1 │ npm ERR! │ npm ERR! Failed at the kubernetes-external-secrets@6.1.0 start script. │ npm ERR! This is probably not a problem with npm. There is likely additional logging output above. │ │ npm timing npm Completed in 258992ms │ npm ERR! A complete log of this run can be found in: │ npm ERR! /home/node/.npm/_logs/2021-07-29T09_44_51_347Z-debug.log we do not see it happening anymore
gharchive/issue
2021-07-29T09:50:34
2025-04-01T06:45:37.051857
{ "authors": [ "j-zimnowoda", "srodenhuis" ], "repo": "redkubes/otomi-core", "url": "https://github.com/redkubes/otomi-core/issues/495", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
306412814
When to open the knowledge base automatically Spend some hours in order to make a suggestion when to open the knowledge base. Think about a metric ... Thanks for that detailed analysis.
gharchive/issue
2018-03-19T10:50:51
2025-04-01T06:45:37.060895
{ "authors": [ "ruKurz" ], "repo": "redlink-gmbh/smarti", "url": "https://github.com/redlink-gmbh/smarti/issues/226", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
269910594
search app has no function [baseurl]/search app has no function Hubzilla-Version 2.8 You don's see an input field there? Now works [baseurl]/search . Have you changed something? The navbar search isn't present on tiny screens - when the sitesearch widget was commented out you just get a search page with no way to search for anything. For large screens the widget isn't necessary, but we don't have comanche directives for responding to screen size (yet). when the sitesearch widget was commented out you just get a search page with no way to search for anything But for me there is a big input field on the search page. There is no actual need for the widget, thats why it was commented out probably... I wonder why you don't have the input field on the search page... Looks like this is to blame in the main get() section: if((! local_channel()) || (! feature_enabled(local_channel(),'savedsearch'))) $o .= search($search,'search-box','/search',((local_channel()) ? true : false)); Conditions: a small screen but I'm logged in and have savedsearch enabled. Hence no input box. So if I take out the conditionals, it will always show an input box. The only issue here seems to be that the saved searches aren't displayed if you have that feature enabled. I'll check this in at any rate but I may have to defer looking at why saved searches aren't listed (and note that if they were, they could end up using a lot of vertical space). The only issue here seems to be that the saved searches aren't displayed if you have that feature enabled. Odd, https://hub.libranet.de/search?search= request displayed one input search field for https://hub.libranet.de/profile/wallzilla Profil notwithstanding the above saved searches aren't enabled. And https://hub.libranet.de/search?search= for https://hub.libranet.de/profile/nmoplus request aren't displayed input search field notwithstanding the above saved searches aren't enabled. Both channels wallzilla and nmoplus are hosted on libranet.de with version 2.8. I have three locations for the nmoplus profile, hub.libranet.de and gerzilla.de locations has the hubzilla version 2.8, freecommunication.org has version 2.06. But only on hub.libranet.de displayed https://hub.libranet.de/search?search= request none input search field. What can I do for investigate of cause for this behavior on hub.libranet.de? In 2.8 you should get a search box at the top of the content if you are logged out. You may not get it if you are logged in. Whre can I enable saved searches for version 2.8? Additional Features > Network and Stream Filtering > Saved Searches
gharchive/issue
2017-10-31T10:39:10
2025-04-01T06:45:37.069972
{ "authors": [ "HappyPony", "git-marijus", "zotlabs" ], "repo": "redmatrix/hubzilla", "url": "https://github.com/redmatrix/hubzilla/issues/898", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1692623286
Feature: Optimize session management Current problem We create a new session for each request, it is not efficient. We could use only one session for a client. Possible solution We should not use with block for HTTP request and create a session inside the ReductClient class. Additional context No response Code of Conduct [X] I agree to follow this project's Code of Conduct +1 for this. I work with a remote system with high latency, requests as currently implemented incur the latency penalty for every request. By enabling session keep-alive things would probably be a lot faster. I tried this: In __init__ in http.py self.connector = aiohttp.TCPConnector() self.session = aiohttp.ClientSession(timeout = self.timeout, connector = self.connector) In request() in http.py, remove the creation of connector and async with ClientSession and turned it into: try: async with self.session.request( method, f"{self.url}{path.strip()}", headers=dict(self.headers, **extra_headers), expect100=expect100, **kwargs, ) as response: if response.ok: yield response else: if "x-reduct-error" in response.headers: raise ReductError( response.status, response.headers["x-reduct-error"], ) raise ReductError(response.status, "Unknown error") except ClientConnectorError: raise ReductError( 599, f"Connection failed, server {self.url} cannot be reached" ) from None I have not run any substantial tests to rule out any bad side-effects of this BUT what I did was to measure before and after. The simplified test does some basic checking of buckets and fetching k records. k=10 before: 15s after: 7s k=100 before: 100s after: 51s I did not tweak any of the keep-alive settings just the defaults coming with: TCPConnector and ClientSession. I am getting data using async for record in bucket.query("entry", start=..., stop=...) Another question, would it be possible to "batch load" records? I guess it would require backend changes to realize that there are ~k records to be sent, and batch them into m packages (with ~k/m records) of reasonable size and have client split up to individual records when iterating over it. This should probably be an option since if you dont want to get hit by the added latency. But I see it very useful if you are fetching lots of data for some use-cases. Hey @mounte, thank you for sharing this. As you've already noticed, we're already working on batching records in one request: reductstore/reductstore#236. We started it in the C++ version, but have to start from scratch in the Rust implementation. The feature will be available in ReductStore v1.5 soon. However, the session optimization is still relevant, and I see from your numbers it can improve a lot. Hey again @mounte , I've tried your suggestions.. I didn't manage to make it work... could you make PR of your working changes? I'm not sure if the client was supposed to work without context manager. P.S. The batching is ready in the development branches of reduct-py and reductstore. So if you're interested to try, you can take it from there. I'm preparing benchmarks. nice @atimin I will have a look later this week and make a PR to test with. I need to read a bit on the aiohttp because I am not sure what I did is 100% correct and if it works in various cases... @mounte I've found a way to reuse the HTTP session without breaking changes. You can use a context manager if you need better performance: async with Client(url, api_token=api_token) as client: bucket = await client.create_bucket("bucket-1", exist_ok=True) await bucket.info()
gharchive/issue
2023-05-02T15:20:14
2025-04-01T06:45:37.113211
{ "authors": [ "atimin", "mounte" ], "repo": "reductstore/reduct-py", "url": "https://github.com/reductstore/reduct-py/issues/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
262742508
Update ApiDocs.md Just a typo fix. 🐣 thanks!
gharchive/pull-request
2017-10-04T10:44:30
2025-04-01T06:45:37.114633
{ "authors": [ "bdwain", "dstodolny" ], "repo": "redux-loop/redux-loop", "url": "https://github.com/redux-loop/redux-loop/pull/157", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1610851873
resium geocoder hi there, I want to add custom geoder in resium. I found this code. it is work in sandcastle cesium but it is not work in resium. function OpenStreetMapNominatimGeocoder() {} /** * The function called to geocode using this geocoder service. * * @param {String} input The query to be sent to the geocoder service * @returns {Promise<GeocoderService.Result[]>} */ OpenStreetMapNominatimGeocoder.prototype.geocode = async (input) => { console.log(input); const endpoint = 'https://maps.googleapis.com/maps/api/geocode/json?address=${{input}}'; const resource = new Cesium.Resource({ url: endpoint, address : input, outputFormat: 'json', key: 'AIzaSyB41DRUbKWJHPxaFjMAwdrzWzbVKartNGg', }); const results = await resource.fetchText(); console.log(results.toString()); let bboxDegrees; return results?.map(resultObject => { console.log(resultObject); bboxDegrees = [ parseFloat(resultObject.lon), parseFloat(resultObject.lat) ]; return { displayName: resultObject.display_name, destination: Cesium.Cartesian3.fromDegrees( // No matter which line below is used, after searching a location // camera end position is always the same height as long as // createWorldTerrain() is used in the viewer as terrainProvider: // bboxDegrees[0], bboxDegrees[1], 0.0 bboxDegrees[0], bboxDegrees[1], 50.0 // bboxDegrees[0], bboxDegrees[1], 15000.0 // bboxDegrees[0], bboxDegrees[1], 50000.0 ), }; }); }; My code is below. <Viewer geocoder = { new OpenStreetMapNominatimGeocoder} > </viewer> When I write the code in this way, I do not know how to get an infiniteLoop error. As another method, I tried this viewer.geocoder.geocoderService =[new OpenStreetMapNominatimGeocoder()] but is not work. again querying the cesium api. can you help me please. const g = useMemo(() => new OpenStreetMapNominatimGeocoder()}, []); return <Viewer geocoder={g} />
gharchive/issue
2023-03-06T07:51:25
2025-04-01T06:45:37.171033
{ "authors": [ "onacione", "rot1024" ], "repo": "reearth/resium", "url": "https://github.com/reearth/resium/issues/585", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
54002263
Mongoid Support Hello! I would like to know if Refile already support Mongoid out of the box. Thanks! No, it does not. I think mongoid support would be best served by an external gem. It shouldn't be hard to make this work. The ActiveRecord integration is very simple. Thank you very much! Any gem? Try this https://github.com/DimaSamodurov/refile-mongoid . It's an early version, you'll need to refer github in Gemfile.
gharchive/issue
2015-01-11T19:45:16
2025-04-01T06:45:37.205740
{ "authors": [ "DimaSamodurov", "everaldo", "jnicklas", "jturolla" ], "repo": "refile/refile", "url": "https://github.com/refile/refile/issues/108", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
187229854
Remote image download in background I have the remote url being filled out after a direct upload to S3, but I want to have refile do the image work in the background. Anyone know a good way to handle this? I recommend having a background processor like sidekiq download the file and process after refile has uploaded it to s3. Refile uploads to s3 Successful saves kicks off sidekiq job sidekiq pulls down the file from s3 (Refile.attachment_url(object, :some_image)) and performs work on the image and then saves it as an absolute URL to s3. You can safely delete your refile store now This also gives you the advantage of not processing your images on every GET request (regardless of CDN) @bufordtaylor how do you get the direct S3 URL? @abitdodgy Refile.attachment_url(object, :some_image) would do the job if your backend is configured correctly. If you are going to save the record and then later migrate it to the model, you can create a string attribute on your model for that and process in the background to upload like this: model.remote_avatar_url = model.temp_remote_avatar_url model.save _Assuming that temp_remote_avatar_url is your column for that.
gharchive/issue
2016-11-04T01:11:53
2025-04-01T06:45:37.209134
{ "authors": [ "JoshuaNovak919", "abitdodgy", "bufordtaylor", "sobrinho" ], "repo": "refile/refile", "url": "https://github.com/refile/refile/issues/514", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1898740520
wait-for-checks is not working Description https://github.com/refined-github/refined-github/assets/440033/bd084604-66c8-4301-b9eb-8e468a6e0e7f How to replicate the issue + URL Enable wait-for-checks Find a PR with running status checks Click "Squash and merge" Expected checkbox to wait for status checks to complete does not appear Extension version 23.9.7 Browser(s) used Firefox Thanks for the report. I noticed this as well, it's likely because GitHub has become (even more) unreliable in updating the checks UI on the page, so a rewrite is now required: https://github.com/refined-github/refined-github/pull/5465#issuecomment-1070337400 I should probably disable the feature via hotfix
gharchive/issue
2023-09-15T16:23:39
2025-04-01T06:45:37.213181
{ "authors": [ "JohnMaguire", "fregante" ], "repo": "refined-github/refined-github", "url": "https://github.com/refined-github/refined-github/issues/6914", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
154168757
An example of how to correctly use inquiries will be greate It is not clear from the readme file what exactly should be added to the views, what should be called and rendered. Found on the internet is the following fragment: <%= render '/refinery/inquiries/inquiries/form', inquiry: @inquiry %> but a similar should also be added to readme to make it easier for people to directly use it. Could you add to the readme please with advice that would have helped you? I was not able to get the inquiry working for the time limit I have set to myself and I have postpone it for now. Because of this I do not fill confident to describe how others should use inquire. If I ever get back to it and manage to use it I will get back to this issue. @thebravoman Do you try to add the inquiry form on each pages ?
gharchive/issue
2016-05-11T06:25:00
2025-04-01T06:45:37.217145
{ "authors": [ "bricesanchez", "parndt", "thebravoman" ], "repo": "refinery/refinerycms-inquiries", "url": "https://github.com/refinery/refinerycms-inquiries/issues/170", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
219273564
Switch to binary rpc at some point JSON-RPC has been convenient so far but has limitations - we are burying a base64-encoded binary payload at the moment to take advantage of msgpack which supports NaN, inf, etc. as numeric values. We might want to switch to a binary rpc method (msgpack-rpc or something like that) when a client for javascript becomes stable and if it's not too heavyweight. As an alternative, since our API is so small (just a handful of methods) we could craft a RESTful service pretty easily using e.g. Falcon and gunicorn instead of the RPC pattern. Then switching between encodings/JSON/msgpack etc would just involve swapping middleware on the server and client sides. Might make life easier. The hug implementation used for python 3 outputs results by default in msgpack directly. We could eventually use https://github.com/lebedov/msgpack-numpy to skip the conversion to python types.
gharchive/issue
2017-04-04T14:40:25
2025-04-01T06:45:37.219345
{ "authors": [ "bmaranville" ], "repo": "reflectometry/reduction", "url": "https://github.com/reflectometry/reduction/issues/36", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
2192404734
Uploader with no files uploaded I'm currently implementing an uploader feature and need to handle the scenario where there is an error message displayed when no files are uploaded. I'm using the rx.upload_files(upload_id=id) and rx.selected_files(id) functions, but I'm unsure how to detect when no files have been uploaded. It seems that these functions don't execute when no files are uploaded. Could you please advise on how I can check for this condition and create an error message accordingly? Currently this is expected behavior given this code: https://github.com/reflex-dev/reflex/blob/ee1ff7f93fa5bfbeb48f296970977c9174fe7136/reflex/.templates/web/utils/state.js#L382-L385 The client side upload handler just returns if no files are selected for upload. I suppose this code could be changed to still submit the request, but return an empty file list. please assign @masenf tried this and repro'd ; after clicking the upload btn , what comes through to state.js (uploadFiles ) is undefined . i logged at the call-site of* uploadFile*s and it seems event.payload.files is undefined when the user sends an "empty upload" . so it throws a parse error in the eventhandler func in uploadFiles. i think a simple log(in addition to return false to notify the user to handle this , will suffice but unless some other reason to insist on sending an empty array to the backend?
gharchive/issue
2024-03-18T14:38:25
2025-04-01T06:45:37.222957
{ "authors": [ "Ernelene", "Yummy-Yums", "masenf" ], "repo": "reflex-dev/reflex", "url": "https://github.com/reflex-dev/reflex/issues/2873", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
202075823
live-coding Hi, regl is great! I would like to use regl to live-code shaders, unfortunately the screen switches to black at each keystroke, I guess it's the compilation time? Or am I missing something? Thank you Bruce @brucelane how are you constructing and using your regl command? I'm just trying it online ( http://regl.party/examples/?basic ) If you set up regl locally with a live-reloading server (webpack/browserSync) then you should be able to have your code editor window on one side, and your browser window on the other and do some live coding. You also won't have to rely on an internet connection :) If you use a tool like budo you can run it in --live mode. When you save the file locally, the page updates. There will be a flash when you save the file to do a reload, but that should be better than a flash for every keystroke. Otherwise, you (or somebody) will have to write a wrapper around regl for a tool like browserify-hmr. I really appreciate webpack, gulp or other tools to reload the page, but in live coding I need an editor which only redraws the result if the shader compiles to keep the audience "captivated" ;-) @brucelane type an error somewhere in your code, will refuse the online editor (at regl.party) from refresh i do +++ and then comment/uncomment Simple and convenient I guess it would be possible to implement such https://webpack.github.io/docs/hot-module-replacement.html maybe it could be a lib too here is an example with webpack: shaders/render.js: module.exports = regl => regl({ frag: ` precision mediump float; uniform vec4 color; varying vec2 uv; void main() { gl_FragColor = vec4(uv, color.b, color.a); }`, vert: ` precision mediump float; attribute vec2 position; varying vec2 uv; void main() { gl_Position = vec4(position, 0, 1); uv = (position + 1.0) / 2.0; }`, attributes: { position: regl.buffer([[-1, -1], [-1, 4], [4, -1]]) }, uniforms: { color: regl.prop("color") }, count: 3 }); main.js: import createREGL from "regl"; import renderShader from "./shaders/render"; ... const gl = canvas.getContext("webgl") || canvas.getContext("experimental-webgl"); const regl = createREGL(gl); let render = renderShader(regl); if (module.hot) { module.hot.accept("./shaders/render", () => { render = require("./shaders/render")(regl); }); } regl.frame(({ time }) => { regl.clear({ color: [0, 0, 0, 0], depth: 1 }); render({ color: [ Math.cos(time * 0.1), Math.sin(time * 0.8), Math.cos(time * 0.3), 1 ] }); }); it's a bit specific but i'm pretty sure it could be made more generic? anyway at least it does the job of hotswapping just the shader part that changes (when editing shaders/render.js, only that part get reloaded) sounds good, going to try it soon, once I'm less busy with a vuejs project ... Webpack is too heavy and confusing for me; I hacked this together with hot-server to update code without a hard refresh. Making tweaks is way easier if you can see the before and after right next to each other! I just released shader-reload which I'm using with budo to achieve live editing of shaders. It involves splitting them into a separate module and attaching a transform on the browserify dev server. When editing the .shader.js files, they will get live-updated, and any other files will trigger a regular browser reload. It looks a bit like this. Regl handles the updates smoothly since it will check for source code differences and re-compile if necessary before rendering. src/index.js const regl = require('regl')(document.body); const shader = require('./blue.shader'); let draw = regl({ frag: () => shader.fragment, vert: () => shader.vertex, ... }); regl.frame(() => { ... draw(); }); src/blue.shader.js module.exports = require('shader-reload')({ vertex: '... shader source string ...', fragment: '... shader source string ...' }); One minor point of annoyance with regl is that when a program fails (e.g. syntax error), it spits out console errors every frame until the program error is resolved.
gharchive/issue
2017-01-20T07:55:34
2025-04-01T06:45:37.273005
{ "authors": [ "1wheel", "alvinsight", "brucelane", "gre", "jwerle", "mattdesl", "substack", "tolkanabroski" ], "repo": "regl-project/regl", "url": "https://github.com/regl-project/regl/issues/398", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1129775261
Update quill-jdbc, quill-jdbc-zio to 3.16.3 Updates io.getquill:quill-jdbc io.getquill:quill-jdbc-zio from 3.10.0 to 3.16.3. I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "io.getquill" } ] labels: library-update, early-semver-minor, semver-spec-minor, commit-count:1 Superseded by #313.
gharchive/pull-request
2022-02-10T10:16:37
2025-04-01T06:45:37.291550
{ "authors": [ "scala-steward" ], "repo": "reibitto/podpodge", "url": "https://github.com/reibitto/podpodge/pull/279", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2745219
I treated get-raw and put-raw as private and made the arguments keyword argument style, but made the two now private raw functions require you to always pass an options map (not sure if that fits your plan for the api.... If not I can always change the two raw functions to be as they were before.) @AlexBaranosky thanks again, these are great. Some more backstory behind the API... get-raw and put-raw are meant to be used when you don't want the automatic (de)serialization based on content-type, so they should still be public. So I think all that I'd ask that you change is make them public again and leave options as an optional arg. That make sense? Yep. Will do tonight. On Jan 6, 2012 10:08 AM, "Reid Draper" < reply@reply.github.com> wrote: @AlexBaranosky thanks again, these are great. Some more backstory behind the API... get-raw and put-raw are meant to be used when you don't want the automatic (de)serialization based on content-type, so they should still be public. So I think all that I'd ask that you change is make them public again and leave options as an optional arg. That make sense? Reply to this email directly or view it on GitHub: https://github.com/reiddraper/sumo/pull/17#issuecomment-3385606 I made the put-raw and get-raw fns both have keyword arg options, so you can call like this (get-raw x y z :r 500 :d 25 :r2 300). Note: the for over result to deserialize each is duplicated in two spots. Might be a function hiding there. Thanks. I think maybe my comments weren't clear before. I was thinking get-raw and put-raw would still be still be public, ala, not defined with defn-. Also, the problem with making them take keyword args is, afaik, it's harder to pass around a hash of options that you'll call this method with several times. That's why I had it take a hash instead of keyword args before. Hit me on IRC if I can explain this better, or there is something I'm not thinking of. Mis-read your earlier post. Should be good now, no? perfect, thanks again!
gharchive/issue
2012-01-06T08:27:46
2025-04-01T06:45:37.300301
{ "authors": [ "AlexBaranosky", "reiddraper" ], "repo": "reiddraper/sumo", "url": "https://github.com/reiddraper/sumo/issues/17", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1181518802
🛑 Zignaly is down In 2519fbb, Zignaly (https://zignaly.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Zignaly is back up in 4520994.
gharchive/issue
2022-03-26T04:47:39
2025-04-01T06:45:37.309779
{ "authors": [ "reinaldoleon" ], "repo": "reinaldoleon/monitoring", "url": "https://github.com/reinaldoleon/monitoring/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2267303510
🛑 GreatNusa.com is down In 1e84665, GreatNusa.com (https://greatnusa.com) was down: HTTP code: 403 Response time: 117 ms Resolved: GreatNusa.com is back up in c795d2c.
gharchive/issue
2024-04-28T02:33:22
2025-04-01T06:45:37.316136
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/5754", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2284714216
🛑 BINUSMAYA CDN (newcontent.binus.ac.id) is down In 86f354d, BINUSMAYA CDN (newcontent.binus.ac.id) (https://newcontent.binus.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: BINUSMAYA CDN (newcontent.binus.ac.id) is back up in c08f730.
gharchive/issue
2024-05-08T05:32:17
2025-04-01T06:45:37.319424
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/6025", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2291429558
🛑 GreatNusa.com is down In a884ae8, GreatNusa.com (https://greatnusa.com) was down: HTTP code: 403 Response time: 93 ms Resolved: GreatNusa.com is back up in 9a9d90d.
gharchive/issue
2024-05-12T17:34:35
2025-04-01T06:45:37.322676
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/6162", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1763532413
More tests name: Pull request about: Request a merge between the fork and the main codebase. title: '(insert: Issue #12 | Some tests still missing in exam_test.py)' labels: ''enhancement" assignees: ''KladeRe" PR-Issue Link fixes #12 Description of PR Changes Please add a concise description of changes that this PR makes here: Added test for add_exam() Added test for list_exam() Added test for edit_exam() Added test for del_exam() Feel free to proceed whenever you feel like with this PR/issue pair!
gharchive/pull-request
2023-06-19T13:17:39
2025-04-01T06:45:37.325892
{ "authors": [ "KladeRe", "rela-v" ], "repo": "rela-v/py-study-planner", "url": "https://github.com/rela-v/py-study-planner/pull/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1379856410
Messages in a conversation are not loading Is there an existing issue for this? [X] I have searched the existing issues Current Behavior I can't scroll upwards past these messages, when on relay.cc they are viewable. Then, reloading the page and now those messages aren't loading. Expected Behavior Messages prior to 10:47 in the screenshot load. In relay.cc, the full message history of me and skydao.eth are viewable. Steps To Reproduce Only one example for now. Tried multiple page reloads, but then reloaded an hour later and it was fixed.
gharchive/issue
2022-09-20T18:53:44
2025-04-01T06:45:37.349119
{ "authors": [ "seandaopanel" ], "repo": "relaycc/receiver", "url": "https://github.com/relaycc/receiver/issues/60", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
104993332
npm install problem I get this error while running npm install on the project...anybody can help me fix it? ➜ relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40 npm install npm WARN peerDependencies The peer dependency graphql@~0.2.6 included from express-graphql will no npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly. npm WARN peerDependencies The peer dependency babel-core@* included from babel-loader will no npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly. fsevents@0.3.8 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/babel/node_modules/chokidar/node_modules/fsevents node-gyp rebuild SOLINK_MODULE(target) Release/.node CXX(target) Release/obj.target/fse/fsevents.o SOLINK_MODULE(target) Release/fse.node utf-8-validate@1.1.0 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/webpack-dev-server/node_modules/socket.io/node_modules/engine.io/node_modules/ws/node_modules/utf-8-validate node-gyp rebuild CXX(target) Release/obj.target/validation/src/validation.o SOLINK_MODULE(target) Release/validation.node bufferutil@1.1.0 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/webpack-dev-server/node_modules/socket.io/node_modules/engine.io/node_modules/ws/node_modules/bufferutil node-gyp rebuild CXX(target) Release/obj.target/bufferutil/src/bufferutil.o SOLINK_MODULE(target) Release/bufferutil.node fsevents@0.3.8 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/webpack/node_modules/watchpack/node_modules/chokidar/node_modules/fsevents node-gyp rebuild SOLINK_MODULE(target) Release/.node CXX(target) Release/obj.target/fse/fsevents.o SOLINK_MODULE(target) Release/fse.node utf-8-validate@1.1.0 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/webpack-dev-server/node_modules/socket.io-client/node_modules/engine.io-client/node_modules/ws/node_modules/utf-8-validate node-gyp rebuild CXX(target) Release/obj.target/validation/src/validation.o SOLINK_MODULE(target) Release/validation.node bufferutil@1.1.0 install /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/node_modules/webpack-dev-server/node_modules/socket.io-client/node_modules/engine.io-client/node_modules/ws/node_modules/bufferutil node-gyp rebuild CXX(target) Release/obj.target/bufferutil/src/bufferutil.o SOLINK_MODULE(target) Release/bufferutil.node npm ERR! Darwin 14.5.0 npm ERR! argv "node" "/usr/local/bin/npm" "install" npm ERR! node v0.10.26 npm ERR! npm v2.14.1 npm ERR! code EPEERINVALID npm ERR! peerinvalid The package graphql@0.4.2 does not satisfy its siblings' peerDependencies requirements! npm ERR! Please include the following file with any support request: npm ERR! /Users/niko/Downloads/relay-starter-kit-37f1d13613db732b2d924a55cecf89c255ce0f40/npm-debug.log Fixed by #21.
gharchive/issue
2015-09-05T00:52:26
2025-04-01T06:45:37.351464
{ "authors": [ "NikZar", "yungsters" ], "repo": "relayjs/relay-starter-kit", "url": "https://github.com/relayjs/relay-starter-kit/issues/29", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1271371202
🛑 UpToBox is down In 4b9fb6a, UpToBox (https://uptobox.com/) was down: HTTP code: 504 Response time: 5287 ms Resolved: UpToBox is back up in cf02cc9.
gharchive/issue
2022-06-14T21:11:46
2025-04-01T06:45:37.359466
{ "authors": [ "rem42" ], "repo": "rem42/upptime", "url": "https://github.com/rem42/upptime/issues/65", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
171177877
"unable to add mic as InputDevice" I keep getting this error I've been having the same issue. It looks like a race condition because when we initialize a CameraEngine, it performs all session initialization in background (see setupSession() in CameraEngine.swift). As soon that the initializing part is over (on our end), we ask it to change the device (going from .Back to .Front) - causing a new call to addInputDevice that happens at the same time as the background initializer one. @delannoyk I failed to follow up but never solved the problem. Where you able to? You can solve it by dispatching the call to the sessionQueue (requires modifying the source code) public func changeCurrentDevice(position: AVCaptureDevicePosition) { self.cameraDevice.changeCurrentDevice(position) // Note: to avoid multi-threading crash dispatch_async(self.sessionQueue) { () -> Void in self.configureInputDevice() } }
gharchive/issue
2016-08-15T14:10:44
2025-04-01T06:45:37.365483
{ "authors": [ "OlivierVoyer", "delannoyk", "otymartin" ], "repo": "remirobert/CameraEngine", "url": "https://github.com/remirobert/CameraEngine/issues/46", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1286880837
fix(remix-dev): do more tree-shaking in dev mode The type of dead code elimination we want to do depends on the minify syntax property: https://github.com/evanw/esbuild/issues/672#issuecomment-1029682369. Without this we have dev builds that run fine in production, but ship server code / dependencies to the browser, blowing it up due to not optimizing out imports and module level blocks like if (process.env.NODE_ENV === "test") {. This will fix the issue seen by those trying to integrate react-three-fiber into their app, as well as enable a good vitest in-source testing story. Closes: # [ ] Docs [ ] Tests Testing Strategy: This is covered by existing integration tests. @jacob-ebey Do you have any quick and easy test cases that we could run locally to see the difference here? Something hopefully less involved than making a react-three-fiber sample app :) @brophdawg11 Throw a if (process.env === "test") block in a route module scope and inspect the bundle. Without this you will have if (false) and the body of the block left in the bundle.
gharchive/pull-request
2022-06-28T07:03:12
2025-04-01T06:45:37.371224
{ "authors": [ "brophdawg11", "jacob-ebey" ], "repo": "remix-run/remix", "url": "https://github.com/remix-run/remix/pull/3588", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
629205153
breaking changes in JLine 3.15.0 Hello, JLine has released version 3.15.0, and it includes some breaking changes (some classes in org.jline.builtins were moved to org.jline.console), so picocli-shell-jline3 4.3.2 doesn't compile against it. Thank you for raising this! We will make picocli-shell-jline3 4.4.0+ depend on jline 3.15. It looks like one of the JLine authors, @mattirn, has already submitted a pull request to change the dependency and the corresponding picocli-shell-jline3 code. I will merge this shortly when I get a chance to review it. I would like to also update the documentation to show examples for the picocli-shell-jline3-4.3.2 with jline-3.14.1 combination, separately from examples for the picocli-shell-jline3-4.4.0+ with jline-3.15 combination. @thetoothpick or @mattirn, will you be able to provide a pull request for this? Thanks for the quick turnaround, @mattirn! Regarding the documentation, instead of keeping multiple versions of Example.java in the README file, would it make more sense to link to the relevant historical versions of picocli-shell-jline3/src/test/java/picocli/shell/jline3/example/Example.java? That file would then need to be kept up-to-date (which it is already). The README.md file could then contain the notes with links to the files locked to their relevant commits (e.g. JLine 3.13.2 and Picocli 4.1.2). Having multiple versions of the file in the README seems hard to maintain, and all the example changes have to be duplicated to the README. I can make a PR for this once #1099 goes in, if you'd like. @thetoothpick I merged #1099 into master. I would like for the documentation to show actual code, at least for the latest version, rather than a link. How about creating a separate documentation page for previous versions? Then we can make the README page cleaner since it only shows the most recent version, and from the README we link to the page with example code for older versions. Also, the code may be a bit long, we can look into making it shorter. Or perhaps have a short initial example and a longer full example. Thoughts? Yea, I think that makes sense to split out the older versions and keep the latest in the README. Hi @thetoothpick, have you had a chance to take a look at this? I'm hoping to do a new picocli release in a week or so, and it would be good to improve the docs in this area. @remkop I created a wiki page for the old examples. Looking at the latest PR, the example didn't actually change for the latest version, so I left the minimum versions as they were. The example for JLine 3.13.2 and Picocli 4.1.2 is now on this wiki page. I also opened a PR (#1117). I'll add some comments about cleaning up the example there.
gharchive/issue
2020-06-02T13:30:30
2025-04-01T06:45:37.380162
{ "authors": [ "remkop", "thetoothpick" ], "repo": "remkop/picocli", "url": "https://github.com/remkop/picocli/issues/1098", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1346407841
Checking import specifications for Picocli examples I tried a variant of the CheckSum example out also within the tool “Eclipse IDE for Java Developers 2022-06 (4.24.0)”. How do you think about to extend the demonstration source code? :thinking: Sorry I don’t understand the question. Can you clarify what you want me to do? The “Picocli example 4.0” is especially mentioned in one document. How do you think about to improve the awareness for details from additional test applications and related examples? :thinking: Sorry, I don't understand what you are trying to say... There is an example (called Example) in the README, there is another example (called CheckSum) in the user manual, and there are a number of examples that are not in the documentation but can be found in the picocli git repository under picocli-examples. Do you mean that the user manual should mention and link to the picocli-examples directory? :thought_balloon: I imagine that additional links (and data formats) can become helpful for better connections of some information sources. Thank you for the clarification! I added a link to picocli-examples under the CheckSum example in the user manual. The README already has many links to the user manual, I will leave the README as is for now. The change will be reflected in the published HTML of the user manual with the next release. Thank you again for raising this. Thanks for another small documentation adjustment. :thought_balloon: I became curious if further collateral evolution will happen.
gharchive/issue
2022-08-22T13:15:27
2025-04-01T06:45:37.386438
{ "authors": [ "Markus-Elfring", "remkop" ], "repo": "remkop/picocli", "url": "https://github.com/remkop/picocli/issues/1788", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2569708903
Frigate+ API Key in Frigate LXC How do I add the Frigate+ API Key to Frigate LXC, to add and activate the Plus API? I Updated the config to include Frigate+ model path. Frigate won't start. 2024-10-07 10:19:58.172528 [INFO] Preparing Frigate... 2024-10-07 10:19:58.172597 [INFO] Starting Frigate... 2024-10-07 10:19:59.473253 [2024-10-07 10:19:59] frigate.app INFO : Starting Frigate (0.14.1-) 2024-10-07 10:19:59.478520 [2024-10-07 10:19:59] frigate.util.config INFO : Checking if frigate config needs migration... 2024-10-07 10:19:59.523695 [2024-10-07 10:19:59] frigate.util.config INFO : frigate config does not need migration... 2024-10-07 10:19:59.554443 ************************************************************* 2024-10-07 10:19:59.554495 ************************************************************* 2024-10-07 10:19:59.554512 *** Your config file is not valid! *** 2024-10-07 10:19:59.554527 *** Please check the docs at *** 2024-10-07 10:19:59.554542 *** https://docs.frigate.video/configuration/index *** 2024-10-07 10:19:59.554556 ************************************************************* 2024-10-07 10:19:59.554570 ************************************************************* 2024-10-07 10:19:59.554585 *** Config Validation Errors *** 2024-10-07 10:19:59.554600 ************************************************************* 2024-10-07 10:19:59.554614 Plus API not activated 2024-10-07 10:19:59.554870 Traceback (most recent call last): 2024-10-07 10:19:59.554941 File "/opt/frigate/frigate/app.py", line 645, in start 2024-10-07 10:19:59.554960 self.init_config() 2024-10-07 10:19:59.554976 File "/opt/frigate/frigate/app.py", line 140, in init_config 2024-10-07 10:19:59.554991 self.config = user_config.runtime_config(self.plus_api) 2024-10-07 10:19:59.555007 File "/opt/frigate/frigate/config.py", line 1599, in runtime_config 2024-10-07 10:19:59.555023 config.model.check_and_load_plus_model(plus_api) 2024-10-07 10:19:59.555038 File "/opt/frigate/frigate/detectors/detector_config.py", line 91, in check_and_load_plus_model 2024-10-07 10:19:59.555067 download_url = plus_api.get_model_download_url(model_id) 2024-10-07 10:19:59.555081 File "/opt/frigate/frigate/plus.py", line 228, in get_model_download_url 2024-10-07 10:19:59.555096 r = self._get(f"model/{model_id}/signed_url") 2024-10-07 10:19:59.555110 File "/opt/frigate/frigate/plus.py", line 84, in _get 2024-10-07 10:19:59.555124 f"{self.host}/v1/{path}", headers=self._get_authorization_header() 2024-10-07 10:19:59.555139 File "/opt/frigate/frigate/plus.py", line 79, in _get_authorization_header 2024-10-07 10:19:59.555153 self._refresh_token_if_needed() 2024-10-07 10:19:59.555168 File "/opt/frigate/frigate/plus.py", line 71, in _refresh_token_if_needed 2024-10-07 10:19:59.555182 raise Exception("Plus API not activated") 2024-10-07 10:19:59.555197 Exception: Plus API not activated 2024-10-07 10:19:59.555211 2024-10-07 10:19:59.555226 ************************************************************* 2024-10-07 10:19:59.555241 *** End Config Validation Errors *** 2024-10-07 10:19:59.555255 ************************************************************* Found https://github.com/tteck/Proxmox/discussions/2711#discussioncomment-8975702 in my search Interesting! I tried that, but that does not seem to work. keen to see if we can get an answer @tteck For me, editing /etc/systemd/system/frigate.service WORKED!! I needed to put in capital letters and reboot. What I did: edit /etc/systemd/system/frigate.service Add line: Environment=PLUS_API_KEY=verylongapikey under [Service] After that: systemctl daemon-reload Then a reboot. @tteck For me, editing /etc/systemd/system/frigate.service WORKED!! I needed to put in capital letters and reboot. What I did: edit /etc/systemd/system/frigate.service Add line: Environment=PLUS_API_KEY=verylongapikey under [Service] After that: systemctl daemon-reload Then a reboot. @tteck maybe you could add this to the bottom of https://tteck.github.io/Proxmox/#frigate-lxc ?? not sure why I didn't get notified on this. Indeed adding the key to the service is the way to go (my fork already has a commented line for that. See here for that Closing this issue as resolved
gharchive/issue
2024-10-07T08:20:29
2025-04-01T06:45:37.418646
{ "authors": [ "3h50", "SubNoizey", "carlosduque-incoxe", "remz1337", "wimb0" ], "repo": "remz1337/Proxmox", "url": "https://github.com/remz1337/Proxmox/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
355315995
node bot.js isn't working When I put in node bot.js I was given this error message Error: Cannot find module 'C:\Users\REDACTED\Desktop\REDACTED\Usagi bot\bot.js' at Function.Module._resolveFilename (internal/modules/cjs/loader.js:581:15) at Function.Module._load (internal/modules/cjs/loader.js:507:25) at Function.Module.runMain (internal/modules/cjs/loader.js:742:12) at startup (internal/bootstrap/node.js:266:19) at bootstrapNodeJSCore (internal/bootstrap/node.js:596:3) REDACTED text is just my name. Make sure bot.js is not actually bot.js.txt! In file explorer go to view and check "file name extensions".
gharchive/issue
2018-08-29T20:15:29
2025-04-01T06:45:37.434967
{ "authors": [ "MrRjLeddbearl", "SLia0" ], "repo": "renesansz/discord-greeter-bot", "url": "https://github.com/renesansz/discord-greeter-bot/issues/159", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1062362174
Proposal: migrate from JavaScript to TypeScript? I see that the code uses JavaScript. The main Renovate repo uses TypeScript. Do we want to migrate from JavaScript to TypeScript? The repo uses a tech stack as close as possible to what was already in @renovatebot, but I have struck a middle ground between the simplicity of https://github.com/renovatebot/renovate-approve-bot/ and the complexity of https://github.com/renovatebot/renovate. I think the overhead of TypeScript is not worth it here. The product itself is just this index.js of ~100 lines :sweat_smile: @rarkins and @viceice What do you think? Would be grread for code safty, but i've no time to work on this. 🤷‍♂️ I'm fine to leave as is
gharchive/issue
2021-11-24T12:22:04
2025-04-01T06:45:37.444299
{ "authors": [ "HonkingGoose", "maxbrunet", "rarkins", "viceice" ], "repo": "renovatebot/renovate-approve-bot-bitbucket-cloud", "url": "https://github.com/renovatebot/renovate-approve-bot-bitbucket-cloud/issues/58", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1485700990
BREAKING: Update Troubleshoot to 0.51.1 What this PR does / why we need it: Updates Troubleshoot. Note that this update includes a change where IPv4 addresses are no longer redacted by default. Folks that want this feature will need to adjust their application to include that as a separate redactor. Which issue(s) this PR fixes: Fixes # n/a Special notes for your reviewer: Note the breaking change to the default redaction of IPv4 addresses in support bundles. Steps to reproduce Create a support bundle. Run preflights. Does this PR introduce a user-facing change? Note that this update includes a change where IPv4 addresses are no longer redacted by default. Folks that want this feature will need to adjust their application to include that as a separate redactor. Does this PR require documentation? NONE FOSSA scan is failing, but the change in Troubleshoot passed it - https://github.com/replicatedhq/troubleshoot/actions/runs/3767118196/jobs/6404338562 this has sat stale too long, will close and refresh with a new one
gharchive/pull-request
2022-12-09T00:54:41
2025-04-01T06:45:37.519443
{ "authors": [ "xavpaice" ], "repo": "replicatedhq/kots", "url": "https://github.com/replicatedhq/kots/pull/3489", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
855762317
Can't send image to the RP I try to add a screenshot/image to the report and have an error in the Logger class. I have logger utils calls: object LoggingUtils { private val LOGGER: Logger = LoggerFactory.getLogger("binary_data_logger") fun log(file: String, message: String?) { LOGGER.info("RP_MESSAGE#FILE#{}#{}", file, message) } fun log(bytes: ByteArray?, message: String?) { LOGGER.info("RP_MESSAGE#BASE64#{}#{}", Base64.getEncoder().encodeToString(bytes), message) } fun logBase64(base64: String?, message: String?) { LOGGER.info("RP_MESSAGE#BASE64#{}#{}", base64, message) } } In the test presence code for sending file: val targetContext = InstrumentationRegistry.getInstrumentation().targetContext val file = EspressoScreenshot.takeScreenshotPath("rp-test8303202978285528238.png", targetContext) Log.i("Logi", File(file).exists().toString()) LoggingUtils.log(file, "Android launch with file") So file locates at this directory: /data/user/0/com.epam.test/cache/rp-test8303202978285528238.png 2021-04-12 11:52:56.943 16833-16869/com.epam.test I/Logi: true And I have that error: java.lang.ExceptionInInitializerError at org.apache.tika.mime.MimeTypesFactory.create(MimeTypesFactory.java:69) at org.apache.tika.mime.MimeTypesFactory.create(MimeTypesFactory.java:100) at org.apache.tika.mime.MimeTypesFactory.create(MimeTypesFactory.java:189) at org.apache.tika.mime.MimeTypes.getDefaultMimeTypes(MimeTypes.java:604) at org.apache.tika.config.TikaConfig.getDefaultMimeTypes(TikaConfig.java:82) at org.apache.tika.config.TikaConfig.<init>(TikaConfig.java:247) at org.apache.tika.config.TikaConfig.getDefaultConfig(TikaConfig.java:386) at org.apache.tika.parser.AutoDetectParser.<init>(AutoDetectParser.java:55) at com.epam.reportportal.utils.MimeTypeDetector.<clinit>(MimeTypeDetector.java:37) at com.epam.reportportal.utils.MimeTypeDetector.detect(MimeTypeDetector.java:57) at com.epam.reportportal.utils.files.Utils.getFile(Utils.java:131) at com.epam.reportportal.message.HashMarkSeparatedMessageParser$MessageType$1.toByteSource(HashMarkSeparatedMessageParser.java:54) at com.epam.reportportal.message.HashMarkSeparatedMessageParser.parse(HashMarkSeparatedMessageParser.java:116) at com.epam.reportportal.logback.appender.ReportPortalAppender.lambda$append$0$ReportPortalAppender(ReportPortalAppender.java:55) at com.epam.reportportal.logback.appender.-$$Lambda$ReportPortalAppender$h9OyensytHEmozo0ozSSnM19Lbo.apply(Unknown Source:6) at com.epam.reportportal.service.LoggingContext.prepareRequest(LoggingContext.java:131) at com.epam.reportportal.service.LoggingContext.lambda$emit$1$LoggingContext(LoggingContext.java:148) at com.epam.reportportal.service.-$$Lambda$LoggingContext$K_IOQDBW1XzYIx0WSWzjuqV5Ea4.apply(Unknown Source:8) at io.reactivex.internal.functions.Functions$Array2Func.apply(Functions.java:529) at io.reactivex.internal.functions.Functions$Array2Func.apply(Functions.java:516) at io.reactivex.internal.operators.maybe.MaybeZipArray$ZipCoordinator.innerSuccess(MaybeZipArray.java:111) at io.reactivex.internal.operators.maybe.MaybeZipArray$ZipMaybeObserver.onSuccess(MaybeZipArray.java:176) at io.reactivex.internal.operators.maybe.MaybeCache.subscribeActual(MaybeCache.java:66) at io.reactivex.Maybe.subscribe(Maybe.java:4290) at io.reactivex.internal.operators.maybe.MaybeZipArray.subscribeActual(MaybeZipArray.java:62) at io.reactivex.Maybe.subscribe(Maybe.java:4290) at io.reactivex.internal.operators.maybe.MaybeToFlowable.subscribeActual(MaybeToFlowable.java:45) at io.reactivex.Flowable.subscribe(Flowable.java:14918) at io.reactivex.Flowable.subscribe(Flowable.java:14865) at io.reactivex.internal.operators.flowable.FlowableFlatMap$MergeSubscriber.onNext(FlowableFlatMap.java:163) at io.reactivex.internal.operators.flowable.FlowableOnBackpressureBuffer$BackpressureBufferSubscriber.drain(FlowableOnBackpressureBuffer.java:187) at io.reactivex.internal.operators.flowable.FlowableOnBackpressureBuffer$BackpressureBufferSubscriber.onNext(FlowableOnBackpressureBuffer.java:112) at io.reactivex.internal.operators.flowable.FlowableFromObservable$SubscriberObserver.onNext(FlowableFromObservable.java:54) at io.reactivex.subjects.PublishSubject$PublishDisposable.onNext(PublishSubject.java:308) at io.reactivex.subjects.PublishSubject.onNext(PublishSubject.java:228) at com.epam.reportportal.service.LoggingContext.emit(LoggingContext.java:148) at com.epam.reportportal.service.ReportPortal.emitLog(ReportPortal.java:218) at com.epam.reportportal.logback.appender.ReportPortalAppender.append(ReportPortalAppender.java:43) at com.epam.reportportal.logback.appender.ReportPortalAppender.append(ReportPortalAppender.java:36) at ch.qos.logback.core.AppenderBase.doAppend(AppenderBase.java:82) at ch.qos.logback.core.spi.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:51) at ch.qos.logback.classic.Logger.appendLoopOnAppenders(Logger.java:270) at ch.qos.logback.classic.Logger.callAppenders(Logger.java:257) at ch.qos.logback.classic.Logger.buildLoggingEventAndAppend(Logger.java:421) at ch.qos.logback.classic.Logger.filterAndLog_2(Logger.java:414) at ch.qos.logback.classic.Logger.info(Logger.java:587) at com.epam.test.android.espresso.junit5.LoggingUtils.log(LoggingUtils.kt:12) at com.epam.test.android.espresso.junit5.UiTest.test_short_password_error_1(UiTest.kt:88) at java.lang.reflect.Method.invoke(Native Method) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:688) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at android.com.epam.reportportal.junit5.ReportPortalExtension.interceptTestMethod(ReportPortalExtension.java:222) at org.junit.jupiter.engine.descriptor.-$$Lambda$gvLh9pTap0g17ZemJcSAjpfE7To.apply(Unknown Source:0) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.-$$Lambda$ExecutableInvoker$ReflectiveInterceptorCall$AfiRfaQW5MFAa8lovVwndaKa8l0.apply(Unknown Source:2) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.-$$Lambda$ExecutableInvoker$hmdF2IKSQSF2OJBv_amWUUg0sS8.apply(Unknown Source:6) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84) at org.junit.jupiter.engine.descriptor.-$$Lambda$gvLh9pTap0g17ZemJcSAjpfE7To.apply(Unknown Source:0) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.-$$Lambda$ExecutableInvoker$ReflectiveInterceptorCall$AfiRfaQW5MFAa8lovVwndaKa8l0.apply(Unknown Source:2) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.-$$Lambda$ExecutableInvoker$hmdF2IKSQSF2OJBv_amWUUg0sS8.apply(Unknown Source:6) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6$TestMethodTestDescriptor(TestMethodTestDescriptor.java:210) at org.junit.jupiter.engine.descriptor.-$$Lambda$TestMethodTestDescriptor$HI566US9RrpykSmg1FzjlYD6qrA.execute(Unknown Source:6) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:206) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:131) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:65) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5$NodeTestTask(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$oEtjVUBr1dzvAyO4rHrqPrrp8iU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7$NodeTestTask(NodeTestTask.java:129) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$sIF-wJyqQcSlbWMGWi7dBYWbuCo.invoke(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8$NodeTestTask(NodeTestTask.java:127) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$EAUqZA5CfD8zbN3AlxhbCoS66eU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:126) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:84) at org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185) at org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService.invokeAll(ForkJoinPoolHierarchicalTestExecutorService.java:129) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5$NodeTestTask(NodeTestTask.java:143) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$oEtjVUBr1dzvAyO4rHrqPrrp8iU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7$NodeTestTask(NodeTestTask.java:129) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$sIF-wJyqQcSlbWMGWi7dBYWbuCo.invoke(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8$NodeTestTask(NodeTestTask.java:127) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$EAUqZA5CfD8zbN3AlxhbCoS66eU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:126) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:84) at org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185) at org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService.invokeAll(ForkJoinPoolHierarchicalTestExecutorService.java:129) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5$NodeTestTask(NodeTestTask.java:143) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$oEtjVUBr1dzvAyO4rHrqPrrp8iU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7$NodeTestTask(NodeTestTask.java:129) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$sIF-wJyqQcSlbWMGWi7dBYWbuCo.invoke(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8$NodeTestTask(NodeTestTask.java:127) at org.junit.platform.engine.support.hierarchical.-$$Lambda$NodeTestTask$EAUqZA5CfD8zbN3AlxhbCoS66eU.execute(Unknown Source:2) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:126) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:84) at org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185) at java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189) at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:285) at java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1155) at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1993) at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1941) at java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:157) Caused by: java.lang.RuntimeException: problem initializing SAXParser pool at org.apache.tika.mime.MimeTypesReader.<clinit>(MimeTypesReader.java:119) ... 123 more Caused by: org.apache.tika.exception.TikaException: prooblem creating SAX parser factory at org.apache.tika.mime.MimeTypesReader.newSAXParser(MimeTypesReader.java:394) at org.apache.tika.mime.MimeTypesReader.setPoolSize(MimeTypesReader.java:378) at org.apache.tika.mime.MimeTypesReader.<clinit>(MimeTypesReader.java:117) ... 123 more Caused by: org.xml.sax.SAXNotRecognizedException: http://javax.xml.XMLConstants/feature/secure-processing at org.apache.harmony.xml.parsers.SAXParserFactoryImpl.setFeature(SAXParserFactoryImpl.java:93) at org.apache.tika.mime.MimeTypesReader.newSAXParser(MimeTypesReader.java:390) ... 125 more At the RP I see launch with that filed test. This code for adding screenshots works on my web project but fails with android - what do I do wrong? The project I've got from https://github.com/reportportal/android-kotlin-example We use Apache Tika to guess binary data MIME type. It worked well for all kinds of java machines before and our users usually work not that good, sending us incorrect MIME types and constantly open new issues about missed screenshots. Unfortunately Tika uses SAX parser for it internal needs and bypasses http://javax.xml.XMLConstants/feature/secure-processing option to it to prevent XXE attacks, but this option is not supported for Android. I'll try to find better solution but for now please manually downgrade Apache Tika to this: implementation 'org.apache.tika:tika-core:1.12' The whole JUnit 5 dependency should look like this: // Report Portal libraries implementation ('com.epam.reportportal:agent-android-junit5:5.1.0-BETA-5') { exclude group: 'org.aspectj' // AspectJ is already included by Android exclude module: 'tika-core' } implementation 'org.apache.tika:tika-core:1.12' @HardNorth thank you, it works. I'm really glad about your quick answer. Fixed in 5.1.0-RC-2
gharchive/issue
2021-04-12T09:27:12
2025-04-01T06:45:37.531115
{ "authors": [ "HardNorth", "anriijmind" ], "repo": "reportportal/agent-android", "url": "https://github.com/reportportal/agent-android/issues/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
234022831
POST converts request data from UTF-8 to ISO-Latin Posting a chunk of XML in utf-8 encoding arrives garbled (iso-latin) at the target server. It should arrive as encoded. Reproduction Steps import requests request = ''' <?xml version="1.0" encoding="UTF-8"?> <Requ> <GivenName>Özdemam</GivenName> <FamilyNameGrünzl</FamilyName> </Requ> ''' print(request) headers = {'Content-Type': 'text/xml; charset=utf-8', } req = requests.Request('POST', 'http://192.168.1.5:8888/api', headers=headers, data=request) prepped_requ = req.prepare() s = requests.Session() http_response = s.send(prepped_requ) Running `nc -l 8888' will show the garbled Umlaut characters, whereas print or command-line curl will produce the expected result. System Information Tested on CentOS7/ Python 3.4.6 (EPEL) and OSX 10.12 MacPorts Python 3.5.3 Seems to be similar to #3476, but is unresolved. You haven't encoded it. ;) On Python 3 the native string type is unicode: it has no encoding. We have to auto-encode the data, and in practice that's done by httplib with no reference to whatever you put in your header field. Try changing your code to this: import requests request = ''' <?xml version="1.0" encoding="UTF-8"?> <Requ> <GivenName>Özdemam</GivenName> <FamilyNameGrünzl</FamilyName> </Requ> '''.encode('utf-8') headers = {'Content-Type': 'text/xml; charset=utf-8', } req = requests.Request('POST', 'http://192.168.1.5:8888/api', headers=headers, data=request) prepped_requ = req.prepare() s = requests.Session() http_response = s.send(prepped_requ)
gharchive/issue
2017-06-06T20:59:17
2025-04-01T06:45:37.556789
{ "authors": [ "Lukasa", "rhoerbe" ], "repo": "requests/requests", "url": "https://github.com/requests/requests/issues/4133", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1783197144
Implement blueprint storage on web-view https://github.com/rerun-io/rerun/pull/2578 introduces the ability to persist blueprint data for desktop app. We need to provide similar implementations for the web-view where we don't have access to file-system storage. Would probably be best to do this by extending eframe: https://github.com/emilk/egui/issues/3134 @jleibs We're doing this by now, right? @jleibs We're doing this by now, right? Not to my knowledge. But also I'm not convinced it's necessary. I kind of like that there's not a magical hidden browser cache where we persist blueprint state. Either you explicitly provide blueprint via url param (or manual open/drag-and-drop),or you get the default behavior.
gharchive/issue
2023-06-30T21:47:29
2025-04-01T06:45:37.559589
{ "authors": [ "Wumpf", "jleibs" ], "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/issues/2579", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1940004079
Fix app.rerun.io deploy failing with 401 gcloud and gsutil use different authentication methods, and the google-cloud-auth action does not authenticate both right now. This makes the final publish step for deploying the web app fail. https://github.com/rerun-io/rerun/actions/runs/6495385783/job/17641411553 Fixed by https://github.com/rerun-io/rerun/pull/3926
gharchive/issue
2023-10-12T13:27:22
2025-04-01T06:45:37.561574
{ "authors": [ "jprochazk" ], "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/issues/3843", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2661180327
Improve error messages for Tensor and Text Document views What Closes #8148. This brings nicer error messages to mono component entities when multiple components were added. @emilk This also improves the API of error_*, warning_label, and success_label in UiExt. Checklist [x] I have read and agree to Contributor Guide and the Code of Conduct [x] I've included a screenshot or gif (if applicable) [x] I have tested the web demo (if applicable): Using examples from latest main build: rerun.io/viewer Using full set of examples from nightly build: rerun.io/viewer [x] The PR title and labels are set such as to maximize their usefulness for the next release's CHANGELOG [x] If applicable, add a new check to the release checklist! [x] If have noted any breaking changes to the log API in CHANGELOG.md and the migration guide PR Build Summary Recent benchmark results Wasm size tracking To run all checks from main, comment on the PR with @rerun-bot full-check. To deploy documentation changes immediately after merging this PR, add the deploy docs label. Done ✅
gharchive/pull-request
2024-11-15T08:12:13
2025-04-01T06:45:37.568759
{ "authors": [ "grtlr" ], "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/pull/8155", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2675284676
What is the testing scope of the dataset in the short video QA? Thanks for your work! I recently tried to reproduce your work on the short video question-answering task, but the performance always differs from the papaer. I wonder if it might be due to the scope of the dataset used? For the MSVD-QA dataset, the QA json downloaded from the instructions under eval_code includes three subsets: train, test, and val. May I ask if the results in your paper were tested on all of these subsets? If possible, I would also like to know if the hyperparameter settings are consistent with the code in the repository? The hyperparameter settings please refer to the paper. Can you provide the GPT version you use? Thank you for your response. I tested the GPT-3.5-turbo-0125 API on the entire MSVD dataset (including 50,505 questions), and observed a 10% drop in accuracy and a 0.3 decrease in score. We evaluate MovieChat using the test set and report the average performance across multiple turns of evaluation.
gharchive/issue
2024-11-20T10:05:35
2025-04-01T06:45:37.580250
{ "authors": [ "Espere-1119-Song", "yichufan" ], "repo": "rese1f/MovieChat", "url": "https://github.com/rese1f/MovieChat/issues/85", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
377725947
White screen while creating new need On desktop view sometimes a white screen appears after clicking on a detail picker while creating a need (e.g. "Find Band" -> "Title") can anybody reproduce this? I cannot reproduce this. Please add the browser and version you tested with, in addition to the facts @quasarchimaere asked for. Update: Could not reproduce it anymore. Maybe fixed automatically
gharchive/issue
2018-11-06T07:42:20
2025-04-01T06:45:37.584930
{ "authors": [ "maxstolze", "peacememories", "quasarchimaere" ], "repo": "researchstudio-sat/webofneeds", "url": "https://github.com/researchstudio-sat/webofneeds/issues/2497", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
178821182
moved empty descriptions from subelement connection-selection to post… fixes #826 empty views are no longer visible when there are connections present Nope. the matches and incoming request views in the need detail view are still broken sent requests is also empty (when I request contact for a match) when connected, messages are empty too
gharchive/pull-request
2016-09-23T08:57:20
2025-04-01T06:45:37.587400
{ "authors": [ "fkleedorfer", "quasarchimaere" ], "repo": "researchstudio-sat/webofneeds", "url": "https://github.com/researchstudio-sat/webofneeds/pull/830", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
192996090
chore: exit with error code 1 on build script usage section Somehow got missed out when the rest of the scripts got updated Oh, very good catch!
gharchive/pull-request
2016-12-01T23:52:59
2025-04-01T06:45:37.602126
{ "authors": [ "jviotti", "lurch" ], "repo": "resin-io/etcher", "url": "https://github.com/resin-io/etcher/pull/925", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
84168097
Fallback to nodejs binary if node was not found Deals with Ubuntu's non standard node binary, which is called nodejs. I confirm it works on a fresh Ubuntu 14.04 installation. It works in OS X as well. Doesn't work on Windows. I tried various workarounds but NPM will not make the correct symlink in Windows except the shebang is #!/usr/bin/env node.
gharchive/pull-request
2015-06-02T19:23:53
2025-04-01T06:45:37.603576
{ "authors": [ "jviotti" ], "repo": "resin-io/resin-cli", "url": "https://github.com/resin-io/resin-cli/pull/61", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
246373687
resin-sanity.bbclass: Move systemd check to image recipes With Yocto Pyro having this check in resin-sanity.bbclass was causing problems, so move it. @floion could you check if that fixes the issue for you? Thanks @resin-jenkins retest this please @resin-jenkins retest this please I've tested this commit with the odroid Pyro build and it seems to fix the issue seen there.
gharchive/pull-request
2017-07-28T15:01:48
2025-04-01T06:45:37.606650
{ "authors": [ "willnewton" ], "repo": "resin-os/meta-resin", "url": "https://github.com/resin-os/meta-resin/pull/783", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
934302
socks proxy support? feature request: would be nice to be able to use socks proxy +1 to that. +1 ! Doable by modifying request.rb:266 to allow a choice between a simple proxy and a Socks proxy! Net::HTTP.SOCKSProxy is provided by http://socksify.rubyforge.org/ but maybe proxify might be a good alternative? Personnaly i'll be modifying the gem locally to use SOCKS proxies but i don't really know how to make it possible to use both types.. GL! How can I use an https proxy with restclient? +1 +1
gharchive/issue
2011-05-20T22:58:25
2025-04-01T06:45:37.625235
{ "authors": [ "MartinVandersteen", "RoUS", "ankitrg", "psouda", "shurikk", "vnazarenko" ], "repo": "rest-client/rest-client", "url": "https://github.com/rest-client/rest-client/issues/70", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2200916422
Add context.date This PR adds a new deterministic date to the context. ctx.date.now() -> is the unix epoch time ctx.date.toJSON() -> is the JSON representation of the current time Hi Igal, for this kind of user-facing changes. Can we file an issue in the documentation to update this?
gharchive/pull-request
2024-03-21T18:46:54
2025-04-01T06:45:37.626348
{ "authors": [ "gvdongen", "igalshilman" ], "repo": "restatedev/sdk-typescript", "url": "https://github.com/restatedev/sdk-typescript/pull/288", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
239744182
[RESTEASY-1669] Deprecating ApacheHttClient*Engine constructors allow… …ing to specify a HttpContext instance and explaining potential issues with that approach in the javadoc; added a new HttpContextProvider abstraction to allow users controlling the HttpContext lifecycle. The changes look fine to me. After having reasoned on the jira comments and talked with the user, I'm merging. Thanks for the review.
gharchive/pull-request
2017-06-30T10:47:15
2025-04-01T06:45:37.636160
{ "authors": [ "asoldano", "rsearls" ], "repo": "resteasy/Resteasy", "url": "https://github.com/resteasy/Resteasy/pull/1201", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
413906300
Support 0.9.4 on Ubuntu Bionic (18.04 latest LTS) as a package I would like to see 0.9.4 of restic available as a supported package on the latest LTS (18.04.2 as of this writing) of Ubuntu. Right now 0.8.3 is the one that installs but I cannot use this with B2 (blackblaze) accounts for backups. Thank you :) Hm, we don't provide any packages for distributions, you would need find someone at Ubuntu to do that. We don't have the resources to provide packages. You could use the released binaries though, they are statically linked and don't depend on anything. Thanks for the tip! I will do that instead. Cheers! Okay, so I'm running 0.9.4 and I've configured (I think) my env vars properly B2_ACCOUNT_ID is the name of my application key example export B2_ACCOUNT_ID="myappkeyname" B2_ACCOUNT_KEY is the key example export B2_ACCOUNT_KEY="<base64 stuff copied from blackblaze web UI>" RESTIC_REPOSITORY is the name of the bucket itself prepended with b2: in front of it example export RESTIC_REPOSITORY="b2:mybucket" when I run restic init I get the following error: Fatal: create repository at b2:bigboy failed: b2.NewClient: b2_authorize_account: 401: Even restic -r "b2:bucketname" init or restic -r "b2:appkeyname" init produce the exact same error. Any thoughts on what I'm doing wrong? Nevermind. B2_ACCOUNT_ID is not the NAME of the "keyName" but rather the "applicationKeyId" found in the blackblaze UI. All is well. Thanks! Thanks for taking the time and leaving your solution here, this will help people :)
gharchive/issue
2019-02-25T02:25:44
2025-04-01T06:45:37.651173
{ "authors": [ "fd0", "stunney" ], "repo": "restic/restic", "url": "https://github.com/restic/restic/issues/2188", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
1297550675
Unable to restore - Unsupported Repository Version Three Versions at hand: My M1 mac restic 0.13.1 compiled with go1.18.3 on darwin/arm64 My centos7 container restic 0.13.1 (v0.13.0-200-gbc96879d) compiled with go1.15 on linux/amd64 The restic docker image, ran with --platform=linux/amd64 restic 0.13.1 compiled with go1.16.15 on linux/amd64 restic --repo s3:s3.amazonaws.com/${bucket}/codebase/${repo} restore latest --target restore Fatal: config cannot be loaded: unsupported repository version. Try again This isn't the restore command, however it generates the same error and I wanted to ensure my restore path wasn't weird ╭─johncasillas at John’s MacBook Pro in ~/misc/test-restic-restore 22-07-06 - 13:10:06 ╰─○ export AWS_DEFAULT_REGION= export AWS_ACCESS_KEY_ID= export AWS_SECRET_ACCESS_KEY= export AWS_SESSION_TOKEN= export RESTIC_PASSWORD= ╭─johncasillas at John’s MacBook Pro in ~/misc/test-restic-restore 22-07-06 - 13:30:55 ╰─○ restic --repo s3:s3.amazonaws.com/${bucket}/codebase/${repo} snapshots Fatal: Fatal: config cannot be loaded: unsupported repository version What backend/server/service did you use to store the repository? s3 Expected behavior Can only get this when I restore from the same container I used to backup # restic --repo s3:s3.amazonaws.com/${bucket}/codebase/${repo} latest --target restore repository d8a4bd49 opened (repo version 2) successfully, password is correct restoring <Snapshot 77bff658 of [/repos/repo] at 2022-07-06 18:22:03.523920404 +0000 UTC by root@ip.ec2.internal> to restore Actual behavior I receive the error Fatal: config cannot be loaded: unsupported repository version. when running on my Mac. When using the restic docker image, I receive Fatal: Fatal: config cannot be loaded: unsupported repository version Steps to reproduce the behavior Trying to restore on M1 mac, when repos were backed up using centos7 container. Trying to restore using restic docker image on M1 mac, where platform=linux/amd64 ╭─johncasillas at John’s MacBook Pro in ~/repos/restic-backups on hector/bitbucket✘✘✘ 22-07-06 - 14:45:14 ╰─⠠⠵ docker run -it \ --platform=linux/amd64 \ -v "$(pwd)"/restore:/restore \ -e RESTIC_PASSWORD=*** \ -e AWS_ACCESS_KEY_ID \ -e AWS_SECRET_ACCESS_KEY \ -e AWS_SESSION_TOKEN \ --entrypoint sh restic/restic / # restic --repo s3:s3.amazonaws.com/${bucket}/codebase/${repo} restore latest --target restore Fatal: Fatal: config cannot be loaded: unsupported repository version / # restic --repo s3:s3.amazonaws.com/${bucket}/codebase/${repo} snapshots Fatal: Fatal: config cannot be loaded: unsupported repository version Do you have any idea what may have caused this? ARM? Do you have an idea how to solve the issue? No. Did restic help you today? Did it make you happy in any way? I really like the software and want this to be a major component of our weekly backups - but need my devs to be able to restore to their machines. The repository was initialized with a restic version built from the master branch, so the repository format is version 2. The released versions only support version 1. You can either use the master branch to restore, or initialize the repo with either a released version of restic or by passing --repository-version 1. Here are the docs: https://restic.readthedocs.io/en/latest/030_preparing_a_new_repo.html
gharchive/issue
2022-07-07T14:31:44
2025-04-01T06:45:37.658354
{ "authors": [ "fd0", "jheck90" ], "repo": "restic/restic", "url": "https://github.com/restic/restic/issues/3817", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
2255854961
Use fd for restic find if present in PATH Output of restic version restic 0.16.4 compiled with go1.21.6 on linux/amd64 What should restic do differently? Which functionality do you think we should add? Utilise fd https://github.com/sharkdp/fd/ for restic find if present in PATH. What are you trying to do? What problem would this solve? It's much faster than anything else and supports regular expressions by default as well as glob-based patterns. Did restic help you today? Did it make you happy in any way? Nevertheless the greatest backup application I ever came across! restic find works with an internal virtual filesystem. There is nothing fd would work on. You can already use fd to search a mounted restic repository.
gharchive/issue
2024-04-22T07:57:32
2025-04-01T06:45:37.662707
{ "authors": [ "MichaelEischer", "j-lakeman" ], "repo": "restic/restic", "url": "https://github.com/restic/restic/issues/4777", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
170054594
does not restore hardlinks (linux) original files, see inode/device are equal: stat /home/jambo/go/src/local/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc /home/jambo/golang/src/github.com/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc File: ‘/home/jambo/go/src/local/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc’ Size: 2353 Blocks: 8 IO Block: 4096 regular file Device: 15h/21d Inode: 19297756 Links: 2 Access: (0444/-r--r--r--) Uid: ( 1000/ jambo) Gid: ( 1000/ jambo) Access: 2016-03-09 05:05:40.354887768 +0200 Modify: 2016-03-09 05:05:40.354887768 +0200 Change: 2016-03-09 09:10:56.556327665 +0200 Birth: - File: ‘/home/jambo/golang/src/github.com/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc’ Size: 2353 Blocks: 8 IO Block: 4096 regular file Device: 15h/21d Inode: 19297756 Links: 2 Access: (0444/-r--r--r--) Uid: ( 1000/ jambo) Gid: ( 1000/ jambo) Access: 2016-03-09 05:05:40.354887768 +0200 Modify: 2016-03-09 05:05:40.354887768 +0200 Change: 2016-03-09 09:10:56.556327665 +0200 Birth: - restored: stat /opt/big/restore/home/jambo/go/src/local/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc /opt/big/restore/home/jambo/golang/src/github.com/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc File: ‘/opt/big/restore/home/jambo/go/src/local/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc’ Size: 2353 Blocks: 8 IO Block: 4096 regular file Device: 1eh/30d Inode: 15984615 Links: 1 Access: (0444/-r--r--r--) Uid: ( 1000/ jambo) Gid: ( 1000/ jambo) Access: 2016-03-09 05:05:40.354887768 +0200 Modify: 2016-03-09 05:05:40.354887768 +0200 Change: 2016-07-31 03:00:06.907364726 +0300 Birth: - File: ‘/opt/big/restore/home/jambo/golang/src/github.com/icza/gowut/.git/objects/50/7cd5ea64c50c3df60df39dbb0d55769e755cbc’ Size: 2353 Blocks: 8 IO Block: 4096 regular file Device: 1eh/30d Inode: 16025073 Links: 1 Access: (0444/-r--r--r--) Uid: ( 1000/ jambo) Gid: ( 1000/ jambo) Access: 2016-03-09 05:05:40.354887768 +0200 Modify: 2016-03-09 05:05:40.354887768 +0200 Change: 2016-07-31 03:01:42.602489940 +0300 Birth: - Hi, thanks for pointing this out. Restoring hardlinks is a feature that is not yet implemented. The repository contains enough data to restore hardlinks in the future, but restic does not do this on restore at the moment. I'd appreciate someone taking the time and implement this. :smile: This is actually a duplicate of #152, I'm closing this issue. Comments can still be added of course.
gharchive/issue
2016-08-09T00:12:02
2025-04-01T06:45:37.665171
{ "authors": [ "djadala", "fd0" ], "repo": "restic/restic", "url": "https://github.com/restic/restic/issues/566", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
578153
Allow registration of custom JsonEncoderDecoder instances. I'd like to register some of my own codecs for RestyGWT to use. This is for times where generated code cannot work - eg: Class with no default constructor Class with no setters Class created from a factory method/builder etc.. Form example, we're using a lot of immutable objects in our code, and the classes do not have any setters and are created using a static factory method. That makes things hard for RestyGWT (and other frameworks too). But, unlike other frameworks, there is no way to inject my own implementation of JsonEncoderDecoder and register it with the RestyGWT runtime. I can have a crack at implementing this but maybe someone else has done it or is thinking about it already? Come on, 2011 and nothing on this yet? I need to send dates as unix second based timestamps and not as milliseconds. Is there nothing that allows this? Hello, I'm trying to provide a custom (de)serializer for one of my classes. I though providing a RestyJsonSerializerGenerator and JsonEncoderDecoderClassCreator would be enough but it's seems it's not working as expected. What is the difference between the existing JsonEncoderDecoderClassCreator extension and the requested feature here ?
gharchive/issue
2011-02-06T06:33:34
2025-04-01T06:45:37.683958
{ "authors": [ "aliakhtar", "blop", "jroyals" ], "repo": "resty-gwt/resty-gwt", "url": "https://github.com/resty-gwt/resty-gwt/issues/26", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1395369585
Pragmatical/1020 flux2k Fluentbit Optimization Type of PR [ ] Documentation changes [ ] Code changes [X] Test changes [X] CI-CD changes [ ] GitHub Template changes PR Checklist [ ] I have updated the documentation accordingly. [ ] I have added tests to cover my changes. [X] All new and existing tests passed. [X] My code follows the code style of this project. [X] I ran lint checks locally prior to submission. [ ] Have you checked to ensure there aren't other open Pull Requests for the same update/change? Purpose of PR move fluentbit into base/apps directory structure, also as part of this change I moved namespace creation files into base/apps directory as well Does this introduce a breaking change [ ] YES [X] NO Validation [ ] Unit tests updated and ran successfully [ ] Update documentation or issue referenced above Issues Closed or Referenced Closes #issue_number (this will automatically close the issue when the PR closes) References #issue_number (this references the issue but does not close with PR) Couple of changes: Add new line at the end of each file In VSCode you can set "files.insertFinalNewline": true in your settings json. It will auto add a new line upon saving at end the file.
gharchive/pull-request
2022-10-03T21:45:44
2025-04-01T06:45:37.690340
{ "authors": [ "kforeverisback", "pragmatical" ], "repo": "retaildevcrews/ngsa-asb", "url": "https://github.com/retaildevcrews/ngsa-asb/pull/136", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
117452676
delete_at, insert_at, splice_at, change_at, and slice use the word "index", possibly confusingly It's possible we should change the argument names in these api pages along these lines: index -> offset startIndex -> startOffset endIndex -> endOffset It's also possible this might not be super confusing, but reading sentences like If only `index` is specified, `deleteAt` removes the element at that index. are somewhat ambiguous as to whether we're talking about a table index or the array index. I'm not entirely convinced this is a problem -- these commands only operate on arrays, and "remove the element at that index" only makes sense if you're talking about an array index, AFAICT -- but I'm not against making the change. It's kind of a gray area, I could go either way but figured it was worth discussing I think the sentences are ok, but I think changing the argument names in the API signatures is a good idea. When scanning over the API signatures in the API index, users might not immediately realize that these commands are about arrays, rather than secondary indexes.
gharchive/issue
2015-11-17T21:18:44
2025-04-01T06:45:37.697438
{ "authors": [ "chipotle", "danielmewes", "deontologician" ], "repo": "rethinkdb/docs", "url": "https://github.com/rethinkdb/docs/issues/955", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
148921531
'npm install -g horizon' is broken Hi, Before diving into the code, I noticed that the installation is broken. npm ERR! 404 'horizon/server' is not in the npm registry. npm ERR! 404 You should bug the author to publish it npm ERR! 404 It was specified as a dependency of 'horizon' As an FYI, I freshly installed npm on my Ubuntu 15.10 machine ( which installed npm 1.4.21) and ran into this exact same problem trying to run npm install -g horizion. Hi all, I have just had this exact same issue on Windows 7 64bit. Node v5.10.1 & npm v1.4.9 @jakkblakk Would you mind opening a new issue with the problems you've encountered? Our windows support is more theoretical than actual at the moment, so it may be unrelated to this issue @jakkblakk @Cinderhaze Not sure if npm has the same versioning on Windows as it does non-Windows but I think you need at least npm 3.0.0 for the correct package namespacing @horizon/[client|server]. I was putting my info out there as a point of reference. Not being familiar with node/npm, having heard about horizon and wanting to give it a try, I was surprised when the getting started instructions didn't work. I'm not sure if you want to tie in some sort of getting started/FAQ to the setup instructions that would point people down the right path to having a new enough node installed. I did what I thought the instructions told me to do, and it didn't work. I know you don't control node/npm versions of packages in package managers for all distros, but a quick little pointer on the getting started page may go a long way for people who want to get up and running. On Mon, May 23, 2016 at 6:53 PM, Daniel Alan Miller < notifications@github.com> wrote: @jakkblakk https://github.com/jakkblakk @Cinderhaze https://github.com/Cinderhaze Not sure if npm has the same versioning on Windows as it does non-Windows but I think you need at least npm 3.0.0 for the correct package namespacing @horizon/[client|server]. — You are receiving this because you were mentioned. Reply to this email directly or view it on GitHub https://github.com/rethinkdb/horizon/issues/254#issuecomment-221119596 -- ~ Daryl Wiest Hey again @Cinderhaze, apologies that it's not working out. I think if you try upgrading your npm though things should be fine. Also any error messages you got would help. AFAIK there's no way currently for us to block the ability to install horizon on outdated versions of npm or else we would do so and provide you with a much more helpful message. @dalanmiller actually, in the package.json you can specify the engines field: { "name": "@horizon/server", // ... "engines" : { "node" : "<semver>", "npm" : "<semver>" } } Thanks @marshall007, I'm going to add this.
gharchive/issue
2016-04-17T05:49:13
2025-04-01T06:45:37.707374
{ "authors": [ "Cinderhaze", "dalanmiller", "deontologician", "jakkblakk", "kerbelp", "marshall007" ], "repo": "rethinkdb/horizon", "url": "https://github.com/rethinkdb/horizon/issues/254", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
413929248
Change type on export Is there a way to change the type of the entry in the postscript? Although my target is BibLaTeX, I do not use the "collection" type, and prefer to have both books and collections as "book" in my .bib file (distinguished by the presence or absence of an author). Is there a way to swap collection with book? I assume I'd have to start with if (Translator.BetterBibLaTeX && item.itemType === 'collection') { but ... I'm afraid I don't know javascript to figure it out myself :/ No worries, scripting isn't as well-documented as it should be, so I generally just do these on request. You can achieve what you want in two ways: indeed with scripting for all items at once: if (Translator.BetterBibLaTeX && item.itemType === 'collection') item.referencetype = 'book' or add this to the extrafield for individual items bibtex[referencetype=book] Excellent, this works! Oops, I closed this prematurely, I thought I had tried it but apparently that link was a book already. With collections from amazon.com I actually still get "collection" and not "book" using if (Translator.BetterBibLaTeX && item.itemType === 'collection') item.referencetype = 'book' ? My bad. The .itemType is the Zotero type. You want to either test for the Zotero itemType (export to BetterBibTeX JSON to see what itemType an item has, or test for item.referencetype === 'collection' Oh, I see, well, JSON gives me "itemType": "book", (and in Zotero itself it's also listed as "Book")? But it still becomes a "collection" in the .bib export. Is it maybe that BBT tests for books without author and translates them into collection automatically (in which case testing for "book" doesn't help, of course). I'm not sure, I'd need a BBT error report to tell, but testing on item.referencetype === 'collection' should always work for your case. Did you see the error report for the item (the same that I submitted for the other thread)? ID is WVKXIW2D-euc Thanks so much for your attention to my issues and immediate help, BTW :) Very much appreciated! That should have been if (Translator.BetterBibLaTeX && this.referencetype === 'collection') this.referencetype = 'book' // Duplicate title to shorttitle if empty if (Translator.BetterBibLaTeX && !this.has.shorttitle) this.add({ name: 'shorttitle', value: item.title }) // Duplicate title to booktitle for books if (Translator.BetterBibLaTeX && item.itemType === 'book') this.add({ name: 'booktitle', value: item.title }) // Set english as langid if empty if (Translator.BetterBibLaTeX && !this.has.langid) this.add({ name: 'langid', value: 'english' }) // Only export year to date field if (Translator.BetterBibLaTeX && item.date) { const date = Zotero.BetterBibTeX.parseDate(item.date) if (date.type === 'date') this.add({ name: 'date', value: date.year }) } // Correct single hyphens in page field if (Translator.BetterBibLaTeX) { if (this.has.pages) this.has.pages.bibtex = this.has.pages.bibtex.replace(/(\w+)-(\w+)/g, '$1--$2'); } or a little compacter if (Translator.BetterBibLaTeX) { if (this.referencetype === 'collection') this.referencetype = 'book' // Duplicate title to shorttitle if empty if (!this.has.shorttitle) this.add({ name: 'shorttitle', value: item.title }) // Duplicate title to booktitle for books if (item.itemType === 'book') this.add({ name: 'booktitle', value: item.title }) // Set english as langid if empty if (!this.has.langid) this.add({ name: 'langid', value: 'english' }) // Only export year to date field if (item.date) { const date = Zotero.BetterBibTeX.parseDate(item.date) if (date.type === 'date') this.add({ name: 'date', value: date.year }) } // Correct single hyphens in page field if (this.has.pages) this.has.pages.bibtex = this.has.pages.bibtex.replace(/(\w+)-(\w+)/g, '$1--$2'); } Does that work for you? Sorry for getting back late ... Yes! This one works. Thanks also for fixing the rest of the code :)
gharchive/issue
2019-02-25T04:36:02
2025-04-01T06:45:37.727960
{ "authors": [ "jandhau", "retorquere" ], "repo": "retorquere/zotero-better-bibtex", "url": "https://github.com/retorquere/zotero-better-bibtex/issues/1140", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
444320036
Missing unit after "Warn me when auto-exports take longer than" In BBT preferences, "Advanced" tab, a unit is missing after the field following "Warn me when auto-exports take longer than": is it seconds, minutes, etc.? It's seconds. I'll add that to the UI. It's in the new version.
gharchive/issue
2019-05-15T09:05:16
2025-04-01T06:45:37.729716
{ "authors": [ "dbitouze", "retorquere" ], "repo": "retorquere/zotero-better-bibtex", "url": "https://github.com/retorquere/zotero-better-bibtex/issues/1185", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2460958068
Android Indicator In Android side while reading nfc tag does any kind of pop up modal will be display like ios or not ? and also please let me know which android version will support this? bcoz i get error in android 11 with checking await NfcManager.isSupported(); No, for Android you have to create a UI yourself. Okay Thanks for reply @pke, which android version will support this? bcoz i get error in android 11 with checking await NfcManager.isSupported();
gharchive/issue
2024-08-12T13:06:47
2025-04-01T06:45:37.861765
{ "authors": [ "jigar-parmar-13", "pke" ], "repo": "revtel/react-native-nfc-manager", "url": "https://github.com/revtel/react-native-nfc-manager/issues/741", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
826350359
unrecognized OS: darwin ccsecret generic my-secret --from-literal key1 --from-literal key2 Enter value for key1 : Enter value for key2 : Unrecognized OS: darwin Please create an issue here, https://github.com/rewanth1997/kubectl-ccsecret/issues @ahmetb It also doesn't install view krew. @nikolay I can install it via krew on ubuntu. I haven't tested this on macos or darwin, hence I dropped the support. I will release an update soon with darwin support.
gharchive/issue
2021-03-09T17:40:33
2025-04-01T06:45:37.863500
{ "authors": [ "ahmetb", "nikolay", "rewanth1997" ], "repo": "rewanth1997/kubectl-whisper-secret", "url": "https://github.com/rewanth1997/kubectl-whisper-secret/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
660643070
React Markdown add a wrapper around table node element Requirement: Wrap all table elements with a wrapper div for styling purposes. Actual code: // Created a Wrapper function const TableWrapper = ({ content }) => { return(<div className="table-wrapper">{content}</div>); }; // Added as property to renderers object <ReactMarkdown className="mt-8 text-lg leading-snug markdown text-gray-400" escapeHtml={false} source={postData.content} renderers={{ code: CodeBlock, table: TableWrapper }} ></ReactMarkdown> Actual Result: No changes Expected Result: Table element to be wrapped with div className='table-wrapper'></div> @arindamdawn have you managed to overcome this? @iorrah Yep Here's some code const TableWrapper = ({ columnAlignment, children }) => { return ( <div className="table-wrapper"> {React.createElement("table", null, children)} </div> ); }; <ReactMarkdown className="mt-8 text-lg leading-snug markdown text-gray-400" escapeHtml={false} source={postData.content} renderers={{ code: CodeBlock, table: TableWrapper }} ></ReactMarkdown> Hope it helps! @arindamdawn appreciated!
gharchive/issue
2020-07-19T06:44:39
2025-04-01T06:45:37.870206
{ "authors": [ "arindamdawn", "iorrah" ], "repo": "rexxars/react-markdown", "url": "https://github.com/rexxars/react-markdown/issues/446", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
237982842
[Mac OS] UnicodeEncodeError Please follow the guide below You will be asked some questions and requested to provide some information, please read them carefully and answer honestly Put an x into all the boxes [ ] relevant to your issue (like that [x]) Use Preview tab to see how your issue will actually look like Make sure you are using the latest version: run youtube-dl --version and ensure your version is 2017.06.23. If it's not read this FAQ entry and update. Issues with outdated version will be rejected. [x] I've verified and I assure that I'm running youtube-dl 2017.06.23 Before submitting an issue make sure you have: [x] At least skimmed through README and most notably FAQ and BUGS sections [x] Searched the bugtracker for similar issues including closed ones What is the purpose of your issue? [x] Bug report (encountered problems with youtube-dl) [ ] Site support request (request for adding support for a new site) [ ] Feature request (request for a new functionality) [ ] Question [ ] Other The following sections concretize particular purposed issues, you can erase any section (the contents between triple ---) not applicable to your issue If the purpose of this issue is a bug report, site support request or you are not completely sure provide the full verbose output as follows: Add -v flag to your command line you run youtube-dl with, copy the whole output and insert it here. It should look similar to one below (replace it with your log inserted between triple ```): $ youtube-dl Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py", line 174, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/usr/local/bin/youtube-dl/__main__.py", line 19, in <module> File "/usr/local/bin/youtube-dl/youtube_dl/__init__.py", line 465, in main File "/usr/local/bin/youtube-dl/youtube_dl/__init__.py", line 58, in _real_main File "/usr/local/bin/youtube-dl/youtube_dl/options.py", line 899, in parseOpts File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/optparse.py", line 1402, in parse_args self.error(str(err)) UnicodeEncodeError: 'ascii' codec can't encode character u'\u2014' in position 17: ordinal not in range(128) ... <end of log> So I get this error no matter what command I give to youtube-dl. Even just typing 'youtube-dl' into terminal gives back the same error. Really not sure what to do at this point. Any help or redirection would be appreciated. Thanks :) Do not use em dash (—) in option switches.
gharchive/issue
2017-06-22T21:43:56
2025-04-01T06:45:37.885967
{ "authors": [ "Sebetter", "dstftw" ], "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/13469", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
250272575
[mixcloud] Failed to parse JSON [X] I've verified and I assure that I'm running youtube-dl 2017.08.13 [X] At least skimmed through the README, most notably the FAQ and BUGS sections [X] Searched the bugtracker for similar issues including closed ones [X] Bug report (encountered problems with youtube-dl) Downloading from Mixcloud fails with an exception in the JSON parser. Download works neither with valid login credentials nor without credentials. There's currently no open issue related to "mixcloud" or "JSON" that matches my problem. Here's the full output with --verbose directly after downloading 2017.08.13: 0 mosu@sweet-chili ~/dl] ./youtube-dl --verbose https://www.mixcloud.com/JazzStandard/jazz-beats-and-flow-004-alfa-mist/ [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: ['--verbose', 'https://www.mixcloud.com/JazzStandard/jazz-beats-and-flow-004-alfa-mist/'] [debug] Encodings: locale UTF-8, fs utf-8, out UTF-8, pref UTF-8 [debug] youtube-dl version 2017.08.13 [debug] Python version 3.6.2 - Linux-4.12.6-1-ARCH-x86_64-with-arch [debug] exe versions: ffmpeg 3.3.3, ffprobe 3.3.3, rtmpdump 2.4 [debug] Proxy map: {} [mixcloud] JazzStandard-jazz-beats-and-flow-004-alfa-mist: Downloading webpage ERROR: JazzStandard-jazz-beats-and-flow-004-alfa-mist: Failed to parse JSON (caused by JSONDecodeError('Expecting value: line 1 column 1 (char 0)',)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Traceback (most recent call last): File "./youtube-dl/youtube_dl/extractor/common.py", line 676, in _parse_json return json.loads(json_string) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) Traceback (most recent call last): File "./youtube-dl/youtube_dl/extractor/common.py", line 676, in _parse_json return json.loads(json_string) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "./youtube-dl/youtube_dl/YoutubeDL.py", line 776, in extract_info ie_result = ie.extract(url) File "./youtube-dl/youtube_dl/extractor/common.py", line 433, in extract ie_result = self._real_extract(url) File "./youtube-dl/youtube_dl/extractor/mixcloud.py", line 90, in _real_extract play_info = self._decrypt_play_info(encrypted_play_info, track_id) File "./youtube-dl/youtube_dl/extractor/mixcloud.py", line 70, in _decrypt_play_info video_id) File "./youtube-dl/youtube_dl/extractor/common.py", line 680, in _parse_json raise ExtractorError(errmsg, cause=ve) youtube_dl.utils.ExtractorError: JazzStandard-jazz-beats-and-flow-004-alfa-mist: Failed to parse JSON (caused by JSONDecodeError('Expecting value: line 1 column 1 (char 0)',)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Already fixed.
gharchive/issue
2017-08-15T10:00:06
2025-04-01T06:45:37.891212
{ "authors": [ "dstftw", "mbunkus" ], "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/13922", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }