id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2034202361
[Feature Request] Change to remote database instead of local sqlite DB Is your feature request related to a problem? Please describe. Every time app is rebuilt. database is initialized and wipes data Describe the solution you'd like Have data be persistent so when updates are applied data wont be wiped Describe alternatives you've considered N/A Additional context Allow DB connection to be used by getting host from environment variable Switching to postgresql managed DB. created write connection pool and read connection pool as well as read user and write user. Converted db.py, home.py, and auth.py to use remote postgresql db. Added evironment variables for db connection variables next, convert models to use remote db
gharchive/issue
2023-12-10T03:26:54
2025-04-01T04:33:10.152662
{ "authors": [ "Ubiquitouskiwi" ], "repo": "Ubiquitouskiwi/BOOKWRMS", "url": "https://github.com/Ubiquitouskiwi/BOOKWRMS/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
221403847
ERROR in C:/lab/quiz/src/app/app.module.ts (8,33): Cannot find module '@ultimate/ngxerrors'. Is this a feature request or a bug report? I think it´s a bug. What's gone wrong? I followed the instructions, but I got the errors below: `time: 31308ms chunk {0} polyfills.bundle.js, polyfills.bundle.map (polyfills) 266 kB {4} [initial] [rendered] chunk {1} main.bundle.js, main.bundle.map (main) 54.4 kB {3} [initial] [rendered] chunk {2} styles.bundle.js, styles.bundle.map (styles) 9.71 kB {4} [initial] [rendered] chunk {3} vendor.bundle.js, vendor.bundle.map (vendor) 3.41 MB [initial] [rendered] chunk {4} inline.bundle.js, inline.bundle.map (inline) 0 bytes [entry] [rendered] ERROR in C:/lab/quiz/src/app/app.module.ts (8,33): Cannot find module '@ultimate/ngxerrors'. ERROR in Error encountered resolving symbol values statically. Could not resolve @ultimate/ngxerrors relative to C:/lab/quiz/src/app/ap p.module.ts., resolving symbol AppModule in C:/lab/quiz/src/app/app.module.ts, resolving symbol AppModule in C:/lab/quiz/src/a pp/app.module.ts ERROR in C:/lab/quiz/src/app/app.module.ts (8,33): Cannot find module '@ultimate/ngxerrors'.` What was meant to happen? Other information @angular/cli: 1.0.0 node: 7.9.0 os: win32 x64 @angular/common: 2.4.10 @angular/compiler: 2.4.10 @angular/core: 2.4.10 @angular/forms: 2.4.10 @angular/http: 2.4.10 @angular/platform-browser: 2.4.10 @angular/platform-browser-dynamic: 2.4.10 @angular/router: 3.4.10 @angular/compiler-cli: 2.4.10 Hey, we changed the name from @ultimate/ngerrors to @ultimate/ngxerrors (added an x after ng) - can you just confirm that you've got @ultimate/ngxerrors installed? It´s what I did! Install in project´s diretory: npm i @ultimate/ngxerrors During the installation I received some warnings (bellow): PS C:\Projetos\lab\quiz> npm i @ultimate/ngxerrors quiz@0.0.0 C:\Projetos\lab\quiz `-- @ultimate/ngxerrors@0.2.0 npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules\chokidar\node_modules\fsev ents): npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.1.1: wanted {"os":"dar win","arch":"any"} (current: {"os":"win32","arch":"x64"}) npm WARN angular2-dynamic-component@0.0.50 requires a peer of @angular/core@~2.0.0 but none was installe d. npm WARN angular2-dynamic-component@0.0.50 requires a peer of @angular/common@~2.0.0 but none was instal led. npm WARN angular2-dynamic-component@0.0.50 requires a peer of @angular/compiler@~2.0.0 but none was inst alled. npm WARN angular2-dynamic-component@0.0.50 requires a peer of @angular/platform-browser@~2.0.0 but none was installed. npm WARN angular2-dynamic-component@0.0.50 requires a peer of @angular/http@~2.0.0 but none was installe d. npm WARN angular2-dynamic-component@0.0.50 requires a peer of rxjs@~5.0.0-beta.12 but none was installed . npm WARN angular2-dynamic-component@0.0.50 requires a peer of zone.js@~0.6.13 but none was installed. Angular Version: / \ _ __ __ _ _ | | __ _ _ __ / | | | | / △ \ | ' \ / _ | | | | |/ _ | '| | | | | | | / ___ | | | | (| | || | | (| | | | || | | | // __| ||_, |_,||_,|| _||| |___/ @angular/cli: 1.0.0 node: 7.9.0 os: win32 x64 @angular/common: 2.4.10 @angular/compiler: 2.4.10 @angular/core: 2.4.10 @angular/forms: 2.4.10 @angular/http: 2.4.10 @angular/platform-browser: 2.4.10 @angular/platform-browser-dynamic: 2.4.10 @angular/router: 3.4.10 @angular/compiler-cli: 2.4.10 appmodule.ts : import { NgxErrorsModule } from '@ultimate/ngxerrors'; .... ... .. imports: [ BrowserModule, // FormsModule, NgxErrorsModule, Template: Nome é requerido (minimo 8 e máximo 40 caracteres). 2017-04-12 19:21 GMT-03:00 Ryan Clark notifications@github.com: Hey, we changed the name from @ultimate/ngerrors to @ultimate/ngxerrors (added an x after ng) - can you just confirm that you've got @ultimate/ngxerrors installed? — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/UltimateAngular/ngxerrors/issues/4#issuecomment-293724011, or mute the thread https://github.com/notifications/unsubscribe-auth/AD626o54vBH9LwNbpb0EMZ76SI638wgDks5rvU5TgaJpZM4M8BvP . -- Augusto *55 21 98079-4250 - *New Should be fixed in 1.0.0 :)
gharchive/issue
2017-04-12T22:05:47
2025-04-01T04:33:10.333227
{ "authors": [ "aucevica", "rynclark" ], "repo": "UltimateAngular/ngxerrors", "url": "https://github.com/UltimateAngular/ngxerrors/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1810993452
Update sbt-mima-plugin to 1.1.3 About this PR 📦 Updates com.typesafe:sbt-mima-plugin from 1.1.1 to 1.1.3 📜 GitHub Release Notes - Version Diff Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.typesafe", artifactId = "sbt-mima-plugin" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "com.typesafe", artifactId = "sbt-mima-plugin" } }] labels: sbt-plugin-update, early-semver-patch, semver-spec-patch, version-scheme:early-semver, commit-count:1 Superseded by #451.
gharchive/pull-request
2023-07-19T01:53:17
2025-04-01T04:33:10.338351
{ "authors": [ "scala-steward" ], "repo": "UltimateSoftware/surge", "url": "https://github.com/UltimateSoftware/surge/pull/383", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1155561478
Adding requirement.txt I think there is a need of adding a requirement.txt file. There are many Python libraries and installing them one by one is not necessary. I am a GSSOC participant, can you assign this task to me. @Akshata-Gunapache Wait for the issue to be assigned then start working on it. As this issue was opened by @Rushijaviya so according to priority, this issue will be assigned to @Rushijaviya . You can create any other issue or either you can wait if @Rushijaviya is unable to complete it then will assign this issue to you. @Umesh-01 You can assign this issue to @Akshata-Gunapache as she already made pr on it. I don't want to repeat pr just for counting. @Umesh-01 Can you please assign this issue to me as I have already made a PR on it and it is my first PR as a beginner. I would be very thankful.
gharchive/issue
2022-03-01T16:41:44
2025-04-01T04:33:10.350406
{ "authors": [ "Akshata-Gunapache", "Rushijaviya", "Umesh-01" ], "repo": "Umesh-01/Python-Assistant", "url": "https://github.com/Umesh-01/Python-Assistant/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1189709099
Bug: Execution of program stops when a command is given to open a software which is not installed the program gives a file not found error, i want to implement a check to see weather the software is installed before opening it. Please assign this to me for GSSOC 22' Great idea, but till now we have used only few commands to open applications (using their paths). To go with this issue, you have to first develop a feature that can open all the applications using a single path (like global path). Then you can work on this feature. Would you like to open a new issue for the same??
gharchive/issue
2022-04-01T12:13:23
2025-04-01T04:33:10.351872
{ "authors": [ "Phoenix-031", "Umesh-01" ], "repo": "Umesh-01/Python-Assistant", "url": "https://github.com/Umesh-01/Python-Assistant/issues/59", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1609837743
the quarry does not mine anything Is there an existing issue for this? [X] I have searched the existing issues Description I found the error that the quarry no longer breaks down, it still worked on the first try. now it only consumes the coal and does nothing else. dismantling and relocating them didn't help. Multiplayer? Singleplayer Operating System? Windows Version Other Minecraft Version 1.19.2 Forge Version 43.2.0 Links https://postimg.cc/gallery/3Y4ddW9 Hey! Did you change one of the two positions before it stopped working? Yes, I changed both positions. And tried several places. Hey! Sorry for taking so long to fix this. Bug has been fixed and version 1.6.5r is now available on 1.18/1.19. https://www.curseforge.com/minecraft/mc-mods/quarry
gharchive/issue
2023-03-04T15:53:51
2025-04-01T04:33:10.380076
{ "authors": [ "Nosnate", "UnhappyCodings" ], "repo": "UnhappyCodings/vanillaquarry", "url": "https://github.com/UnhappyCodings/vanillaquarry/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
545846513
re-implement the nc_def_var_szip() function, including for parallel I/O Fixes #1546 In this PR I re-add the old nc_def_var_szip() function. The capability of using szip compression, when supported by HDF5, was already present in the library (though not mentioned in the documentation). This PR adds the old function to help users achieve this. The code has also been changed so that parallel I/O works with szip. This is fully supported by HDF5, we just need to let it work. I have deferred the question of whether this function should be added to the dispatch table. @DennisHeimbigner, do you agree that it should be? After this PR is merged I will add support in the Fortran APIs. This PR also includes open PR #1582 My travis runs are all failing because of the remote access tests. I believe you guys are working on that problem... What is the recommended method for a NetCDF client to check whether the library they are linking too support SZIP compression? Should I check based on version, or ?. I don't seen anything in netcdf_meta.h I can use or in netcdf.h or nc-config Of course you can try to invoke nc_def_var_szip and it will fail. But I assume that you want a compile time test. AFAIK I don't think we have one. (Nor do we have one for zip, butthat is assumed to be always available). Not sure we should have a compile time test because from my point of view, szip us no different than any other filter. Ed asks if szip should be in the dispatch table. As I state above, I view szip like any other filter so I would actually run szip thru nc_def_var_filter with (internally) a special check for the szip filter id. Similarly, I would put extern for nc_def_var_szip in include/netcdf_filter.h. I think Dave is right; the ifdef flags for szip are wrong. The flag in the HDF5 public header appears to be H5_HAVE_FILTER_SZIP @DennisHeimbigner actually szip is not like any other filter (except zlib). For example, the existing code handles setting of the szip filter with nc_def_var_filter() differently from setting other filters. The szip code is checked for, and H5Pset_szip() is called instead of H5Pset_filter(). Also, netcdf filters do not currently allow parallel, so if I set this through nc_def_var_filter() it would disallow parallel I/O, which works, and works very well. It's necessary, to achieve performance. So we definitely don't want to limit szip (or other filters) to sequential-only builds. Also, also, the current filter implementation seems to only allow one filter. We are intending to add a bit-grooming filter for HDF5, and if we want to use that filter with szip we need szip to be handled separately, or filters to handle multiple filters. Our config.h uses has: /* Define to 1 if you have the `H5Z_SZIP' function. */ #define HAVE_H5Z_SZIP 1 Do we think I should be using something else? Also @DennisHeimbigner WRT moving prototype to netcdf_filter.h: I will do so.
gharchive/pull-request
2020-01-06T17:31:55
2025-04-01T04:33:10.397672
{ "authors": [ "DennisHeimbigner", "edwardhartnett", "gsjaardema" ], "repo": "Unidata/netcdf-c", "url": "https://github.com/Unidata/netcdf-c/pull/1589", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2735381727
Set up integration testing using an Argus server running in Docker Compose This sets up pytest fixtures to start a full Argus server (using docker-compose) that can be used for full integration tests of the API library. This will enable us to properly verify the validity of this library against Argus. Two issues I did not realize: pytest-docker-compose-v2 will not work with Python 3.8. Not sure whether we should drop Python 3.8 compatibility from PyArgus as well (this doesn't really affect PyArgus itself, just the test suite) In order to properly test posting of incidents, it's not good enough to just operate as the admin user. We need to create a source system/user/token combo as well, and I'm not sure how we would automate that currently. Suggestions are welcome. I thought I could skip the installation of the pytest plugin only on Python 3.8, and from there skip all tests that depended on it. But being a pytest plugin, it needs to be enabled globally in the top-level conftest.py file, and there doesn't seem to be a way to continue if a requested plugin isn't available. So for the time being, I've pushed a commit that disables testing on Python 3.8.
gharchive/pull-request
2024-12-12T09:39:51
2025-04-01T04:33:10.410640
{ "authors": [ "lunkwill42" ], "repo": "Uninett/pyargus", "url": "https://github.com/Uninett/pyargus/pull/20", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
275969305
serverless | aws | python requirement | invalid volume error docker: Error response from daemon: invalid volume specification: 'C:\Users\sant_\python-s3-thumbnail:/var/task:z'. See 'docker run --help' I am using windows docker (docker toolkit). I have also tried to run the same in conda virtual environment and nodejs CLI but got the same result in all cases. Thanks for your support in advance. Interesting, I wonder if windows doesn't support the :z option to volume mounts in docker. It's only relevant to Linux systems with SELinux enabled, but it's a no-op on macOS (not an error like this). I'm not sure when I'll get to this as I don't use windows systems for development so I'd have to set one up with docker, python, nodejs, & serverless :frowning_face: Thanks for your response. But without this, I am stuck completely. Will really appreciate if you can help out. I might be able to fix this. There are two things that need to be done: Enable shared drive in Docker Taskbar settings Update lib/pip.js to replace backslash in servicePath with forward slashes. the docker run command requires windows paths to be specified in the forward-slash form of e:/folder or /e/folder. The shell that is currently used to run serverless matters a lot. cmd.exe - servicePath is in the form of 'e:\folder' (backslash) git-bash.exe - servicePath is in the form of '/e/folder' bash.exe is WSL Win10 - servicePath is in the form of '/mnt/e/folder' Note: WSL Ubuntu bash has a docker frontend that connecting to the docker backend in Windows by configuring export DOCKER_HOST='tcp://0.0.0.0:2375' See https://blog.jayway.com/2017/04/19/running-docker-on-bash-on-windows/ For usage of dockerizePip on Windows with the above patch #110, do Step 1 only if running serverless on windows, or do Step 1 & 2 if running serverless inside WSL. Enabling shared volume in Windows Docker Taskbar settings Installing the Docker client on Windows Subsystem for Linux (Ubuntu) I've also had this issue with Docker Toolbox for Windows. The instructions to enable a shared drive don't apply to it. Docker Toolbox seems to require the /c/Users/... syntax for volumes. The error disappeared once I upgraded to Docker for Windows. It'd be nice to add support for Docker Toolbox or somehow block it with a clear message. @kichik was your commit #141 simply because you had the legacy Docker version of windows installed? @heri16 Probably. But I think #144 was more of a fix for this specific issue. having a similar issue Serverless: Invoke deploy Serverless: Invoke package Serverless: Invoke aws:common:validate Serverless: Invoke aws:common:cleanupTempDir Serverless: Installing required Python packages with python3.6... Serverless: Docker Image: lambci/lambda:build-python3.6 Error -------------------------------------------------- docker: Error response from daemon: invalid volume specification: 'C:/Users/kingwill101/observerbot:/var/task:z'. See 'docker run --help'. For debugging logs, run again after setting the "SLS_DEBUG=*" environment variable. Stack Trace -------------------------------------------- Error: docker: Error response from daemon: invalid volume specification: 'C:/Users/kingwill101/observerbot:/var/task:z'.See 'docker run --help'. at ServerlessPythonRequirements.installRequirements (C:\Users\kingwill101\observerbot\node_modules\serverless-python-requirements\lib\pip.js:106:11) From previous event: at PluginManager.invoke (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:390:22) at PluginManager.spawn (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:408:17) at Deploy.BbPromise.bind.then.then (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\plugins\deploy\deploy.js:123:50) From previous event: at Object.before:deploy:deploy [as hook] (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\plugins\deploy\deploy.js:113:10) at BbPromise.reduce (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:390:55) From previous event: at PluginManager.invoke (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:390:22) at PluginManager.run (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:421:17) at variables.populateService.then.then (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\Serverless.js:157:33) at runCallback (timers.js:810:20) at tryOnImmediate (timers.js:768:5) at processImmediate [as _immediateCallback] (timers.js:745:5) From previous event: at Serverless.run (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\lib\Serverless.js:144:8) at serverless.init.then (C:\Users\kingwill101\AppData\Roaming\npm\node_modules\serverless\bin\serverless:43:50) at <anonymous> Get Support -------------------------------------------- Docs: docs.serverless.com Bugs: github.com/serverless/serverless/issues Issues: forum.serverless.com Your Environment Information ----------------------------- OS: win32 Node Version: 8.11.2 Serverless Version: 1.30.1 @glenfordwilliams please try with the latest version of the plugin. If it still happens, attach the log, version numbers, and information about your Docker installation (Docker Toolbox or Docker for Windows, versions, etc.)
gharchive/issue
2017-11-22T07:31:33
2025-04-01T04:33:10.430605
{ "authors": [ "dschep", "duttasantanuGH", "glenfordwilliams", "heri16", "kichik" ], "repo": "UnitedIncome/serverless-python-requirements", "url": "https://github.com/UnitedIncome/serverless-python-requirements/issues/105", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1955042007
Apple Silicon support So, these CPU architectures have been released for a while now, to the point that many professionals have switched to them. I found this old issue that mentions "support is coming": #336 but seems that hasn't happened yet. Any update on a timeline for Apple Silicon support? Other things break when using x64 builds on M1/M2. Hi Felix, We have been working on native Apple Silicon support and hope to have something we can share very soon. Hi, Earlier this week we released a new bundle of OpenUSD packages. Apple Silicon is supported by all packages (along with lots of other improvements). https://forum.unity.com/threads/new-openusd-packages-now-available.1524583/
gharchive/issue
2023-10-20T21:37:11
2025-04-01T04:33:10.544392
{ "authors": [ "hybridherbst", "michaeljblain" ], "repo": "Unity-Technologies/usd-unity-sdk", "url": "https://github.com/Unity-Technologies/usd-unity-sdk/issues/416", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1512867199
Added support for force_mode, freedrive and tool contact This commit includes following changes: New control mode for freedrive New control mode for when the tool is in contact Commands for stopping and starting tool contact Commands for stopping and starting force mode Added force_mode_damping and force_mode_gain_scaling to the constructor Force mode gain scaling is only available for e-series and therefore it is removed from the control script, when the robot is not e-series Tool contact example Test for script command interface Updated test for reverse interface This PR will require some changes to the ROS and ROS2 drivers, in order for them to use the added commands. This commit includes following changes: New control mode for freedrive New control mode for when the tool is in contact Commands for stopping and starting tool contact Commands for stopping and starting force mode Added force_mode_damping and force_mode_gain_scaling to the constructor Force mode gain scaling is only available for e-series and therefore it is removed from the control script, when the robot is not e-series Tool contact example Test for script command interface Updated test for reverse interface This PR will require some changes to the ROS and ROS2 drivers, in order for them to use the added commands. Thank you very much for implementing this! I've been testing some features out and would like to get some clarification about freedrive mode. I tried to call ur_driver_->writeFreedriveControlMessage(urcl::control::FreedriveControlMessage::FREEDRIVE_START);, it succeeded but went back to servo thread right after. I tried to make a noop call but did not help. So if we want to keep the robot in freedrive, do we need to actively maintain a thread that calls FREEDRIVE_START in the beginning and keeps calling FREEDRIVE_NOOP until we decide to terminate?
gharchive/pull-request
2022-12-28T14:46:54
2025-04-01T04:33:10.611684
{ "authors": [ "shuobh", "urmahp" ], "repo": "UniversalRobots/Universal_Robots_Client_Library", "url": "https://github.com/UniversalRobots/Universal_Robots_Client_Library/pull/138", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2297947593
Built-in JSON serialization without reflection Is your feature request related to a problem? Please describe. Depending on a Json serialization library that requires a custom implementation, can lead to errors. In particular, when defining a custom serialization with system.text if we forget to do stream.Position = 0; the payload will not be serialized properly, leading to errors downstream. Describe the solution you'd like Serialization should be provided by the SDK, it's not something users should override, as there's only 1 correct way of serializing the payload. The downside is that the actual users, that are using a custom serializer might cause a crash at runtime. Describe alternatives you've considered None, but open to suggestions Additional context As a sample test, I wrote a test for System.text as serializer. It's currently not working because the test compares strings to strings and the only difference is the sort order of fields. So it's the test that needs to change. On top of that, the test is not covering serialization of metrics, so it's incomplete. But it's a good enough to prove the point that serialization is dangerous if left in the hands of users and it should be handled by the SDK without giving users the ability to override it: https://github.com/Unleash/unleash-client-dotnet/compare/system-text-serialization-test?expand=1 We're making this issue visible after some issues caused by custom serializers, but it's not gonna be prioritized soon. We'd like to collect some feedback from the community (you), and why not some contribution. Why not opting for System.Text.Json rather than still using the Newtonsoft implementation? Did not glance too deeply in the code base but for the things I needed it wasn't such a hard thing, and by doing so you could tap into the power of source generated serialization as well for the newer .NET versions. Not sure how backwards compatible this is though since it's been a long time now that I used .NET framework. Why not opting for System.Text.Json rather than still using the Newtonsoft implementation? Did not glance too deeply in the code base but for the things I needed it wasn't such a hard thing, and by doing so you could tap into the power of source generated serialization as well for the newer .NET versions. Not sure how backwards compatible this is though since it's been a long time now that I used .NET framework. Yeah I think this makes sense Handled by #249 This should be out in the next major release
gharchive/issue
2024-05-15T13:35:33
2025-04-01T04:33:10.631228
{ "authors": [ "PhbsSmn", "gastonfournier", "sighphyre" ], "repo": "Unleash/unleash-client-dotnet", "url": "https://github.com/Unleash/unleash-client-dotnet/issues/221", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2387140820
decide where to put the synthetic tips generation and prepare what to put on main branch https://github.com/UoA-eResearch/ResBaz24Python/blob/338dbf157a5b68ef0141cb5e5385ab7acd1b6c62/IntroductionToPython.ipynb#L1 How to generate synthetic data exceeds content; new branching strategy in place anyways
gharchive/issue
2024-07-02T20:49:25
2025-04-01T04:33:10.638339
{ "authors": [ "jensbri" ], "repo": "UoA-eResearch/ResBaz24Python", "url": "https://github.com/UoA-eResearch/ResBaz24Python/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
848887986
[Install] Check Installation Guide with TPU v3 Remove plyfile from extra_installs. Add how to install plyfile in docs/mlsys21/tpu_v3_throughputs.md. Check the Installation section work on PyTorch XLA 1.7. These requirements are done. Close this issue now.
gharchive/issue
2021-04-02T01:10:06
2025-04-01T04:33:10.643446
{ "authors": [ "eric-zheng", "wangshangsam" ], "repo": "UofT-EcoSystem/hfta", "url": "https://github.com/UofT-EcoSystem/hfta/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
265445992
test errors when deploying to production Hi, I need help I got so many errors of the test classes when deploying to production There is not a high change of getting any help if you don't provide any details of what failed when with what error message ;-) I'm not sure if this is the same error that the OP was encountering, but I just started getting a test failure when deploying to production. System.LimitException: Apex CPU time limit exceeded Stack Trace: Class.SchemaCache.objectDescribesMap: line 56, column 1 Class.SchemaCache_Test.testSchemaCacheMaps: line 145, column 1 Any ideas?
gharchive/issue
2017-10-13T23:50:19
2025-04-01T04:33:10.645399
{ "authors": [ "MennaMagdy", "adamconde", "up2go-rsoesemann" ], "repo": "Up2Go/visualforce-table-grid", "url": "https://github.com/Up2Go/visualforce-table-grid/issues/26", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2578834358
Added PDF malware Detection Pipeline Related Issues or bug Issue : #198 Fixes: #198 Proposed Changes Pipeline is created and usage and info is mentioned in readme file Additions: Model Training notebook Pdf feature extraction notebook Data processing and model prediction notebook Dataset used Trained Model Readme.md Additional Info Anything related Issues Screenshots Original Updated original screenshot **updated screenshot ** @DarshAgrawal14 can you please show results here and also attach in the readme file
gharchive/pull-request
2024-10-10T13:18:57
2025-04-01T04:33:10.649362
{ "authors": [ "DarshAgrawal14", "UppuluriKalyani" ], "repo": "UppuluriKalyani/ML-Nexus", "url": "https://github.com/UppuluriKalyani/ML-Nexus/pull/205", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1679399345
DHGR displayed incorrectly when monochrome mode is turned on Fixed in https://github.com/V2RetroComputing/analog-firmware/commit/7c3fa52a3f9ccd4cae41440556b44ea920cd1d09
gharchive/issue
2023-04-22T07:46:05
2025-04-01T04:33:10.726149
{ "authors": [ "dkgrizzly" ], "repo": "V2RetroComputing/analog", "url": "https://github.com/V2RetroComputing/analog/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
193282629
Allow infinity training places In gitlab by @A-Lawrence on Aug 22, 2016, 23:07 Instead of having to define a fixed limit for training places, permit the use of the infinity symbol NOTE: NOT ZERO. ZERO MEANS NO SPACES AVAILABLE! In gitlab by @A-Lawrence on Aug 23, 2016, 00:01 Removed ~98472 label In gitlab by @A-Lawrence on Aug 23, 2016, 23:20 Removed ~558948 label In gitlab by @A-Lawrence on Aug 23, 2016, 23:27 mentioned in commit 925972226ae58188a74f87169a2a3f6ce5d49dc4
gharchive/issue
2016-12-03T11:53:24
2025-04-01T04:33:10.737129
{ "authors": [ "NFarrington" ], "repo": "VATSIM-UK/core", "url": "https://github.com/VATSIM-UK/core/issues/460", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
193260328
Clear Mship Cache (Old 292) In gitlab by @A-Lawrence on Aug 31, 2015, 01:07 Clear the cache with tags mship_account and the account_id when a member's account is updated via a feed/script. In gitlab by @A-Lawrence on Aug 31, 2015, 01:12 @ALawrence66 Do we currently cache these somewhere, or is this for future use? NFarrington - 2015-03-04T20:37:29.805363+00:00 In gitlab by @NFarrington on Jan 12, 2016, 23:27 Closed as we don't seem to use these tags.
gharchive/issue
2016-12-03T02:42:35
2025-04-01T04:33:10.739157
{ "authors": [ "NFarrington" ], "repo": "VATSIM-UK/core", "url": "https://github.com/VATSIM-UK/core/issues/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
751743394
Disable Secondary Forum Group Sync Due performance issues. :tada: This PR is included in version 6.5.7 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2020-11-26T17:21:14
2025-04-01T04:33:10.740883
{ "authors": [ "CalumTowers", "VATSIMUK" ], "repo": "VATSIM-UK/core", "url": "https://github.com/VATSIM-UK/core/pull/1989", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
492672480
Regional Date in VBA Json Conveter I have checked the code does not handle the regional from control system of my computer. Is there a way to change this nice converter to handle regional dates, surely this could be a good thing. Regards Chris I have sorted out this one @nectorprime Any info that you can share here that could help others with the same problem or that I could integrate into the library? Dear Timhall I have a challenge in using a "POST"method , kindly how you can help me correcting the error , the same VBA code works when using "GET method. The problem is on the get method,kindly shade light also on the issue of password & user name. Here is the code below: Private Sub CmdSales_Click() ' Const SQL_SELECT As String = "SELECT * FROM Qry3;" Dim coll As VBA.Collection Dim dict As Scripting.Dictionary Dim db As DAO.Database Dim rs As DAO.Recordset Dim fld As DAO.Field Dim qdf As DAO.QueryDef Dim prm As DAO.Parameter Dim http As Object Set http = CreateObject("MSXML2.XMLHTTP") Set db = CurrentDb Set qdf = db.QueryDefs("Qry4") http.Open "POST", "http://jsonplaceholder.typicode.com/users/?id=" & Me.txtUserID, False http.send ConvertToJson(coll, Whitespace:=3) Set JSON = ParseJson(http.responseText) For Each prm In qdf.Parameters prm = Eval(prm.Name) Next prm Set rs = qdf.OpenRecordset() Set qdf = Nothing Set coll = New VBA.Collection ' Set db = CurrentDb ' Set rs = db.OpenRecordset(SQL_SELECT, dbOpenSnapshot) If Not rs.BOF And Not rs.EOF Then Do While Not rs.EOF Set dict = New Scripting.Dictionary For Each fld In rs.Fields dict.Add fld.Name, rs.Fields(fld.Name).Value Next fld coll.Add dict rs.MoveNext Loop End If rs.Close Set fld = Nothing Set rs = Nothing Set db = Nothing Set dict = Nothing MsgBox "Post Sucess", vbOKOnly, "Audited by Chris Hankwembo" Set coll = Nothing End Sub
gharchive/issue
2019-09-12T08:53:22
2025-04-01T04:33:10.756328
{ "authors": [ "nectorprime", "timhall" ], "repo": "VBA-tools/VBA-JSON", "url": "https://github.com/VBA-tools/VBA-JSON/issues/143", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1464466029
Passing null Hi, How do I pass null value ? I need convertToJson result to be as as null instead of "null" Dim payload As New Dictionary ' How to set id to null. ' vbNullString, vbNull become 1 instead of null ' vbEmpty become 0 payload.Add "id", "null" payload.Add "provider", "something" payload.Add "name", "someone" JsonConverter.ConvertToJson(payload) vbNullStringis a string constant of zero length, a but more efficient than its equivalent literal"". vbNull is a constant = 1 returned byVarType() when the variable has no valid data. You can use it on the receiving (ParseJson) side in code like If VarType(jsonObject("error")) <> vbNull Then To send a null value , simply use Null directly or via a Variant. Sub Test() Dim payload As New Dictionary Dim nullvar As Variant: nullvar = Null payload.Add "id", nullvar payload.Add "name", "someone" Debug.Print JsonConverter.ConvertToJson(payload) ' =>{"id":null,"name":"someone"} End Sub If this solves your problem, please close your issue here. Ok it's working. Thanks.
gharchive/issue
2022-11-25T12:14:19
2025-04-01T04:33:10.759406
{ "authors": [ "Nick-vanGemeren", "wiryonolau" ], "repo": "VBA-tools/VBA-JSON", "url": "https://github.com/VBA-tools/VBA-JSON/issues/245", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1794575562
84 led driver Wrote basic LED driver. Nothing fancy, just a blinking LED to show EOS is running ok everything should be good except for the config stuff
gharchive/pull-request
2023-07-08T00:44:17
2025-04-01T04:33:10.773006
{ "authors": [ "David-Rey" ], "repo": "VIP-LES/EosPayload", "url": "https://github.com/VIP-LES/EosPayload/pull/86", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
873718364
Best_model.pdparams Hello and thank you for sharing this code. I would like to train this repo on my own dataset. I downloaded the resnet18-5c106cde.pth and converted it to this paddle's model : resnet18-torch.pdparams. I put this weight file in the pretrained folder and seems everything is ok. But when I run the train.py file , I have this error RuntimeError: Parameter file [ ./work_dir/ff_c23/Best_model.pdparams ] not exists What is the problem ? Thanks in advance. Do you have validation set ? If you don't have, you must modify function init_model in runner.py Hello and thank you for sharing this code. I would like to train this repo on my own dataset. I downloaded the resnet18-5c106cde.pth and converted it to this paddle's model : resnet18-torch.pdparams. I put this weight file in the pretrained folder and seems everything is ok. But when I run the train.py file , I have this error RuntimeError: Parameter file [ ./work_dir/ff_c23/Best_model.pdparams ] not exists What is the problem ? Thanks in advance. Hi I have the same issue. did you able to solve it?! You can remove all code in fuction init_model() of runner.py and replace by pass when you train: def init_model(): pass Từ: morteza golzan @.> Đã gửi: 28 Tháng Bảy 2021 2:08 SA Đến: VIS-VAR/LGSC-for-FAS @.> Cc: Nguyen Duc Giang 20172523 @.>; Comment @.> Chủ đề: Re: [VIS-VAR/LGSC-for-FAS] Best_model.pdparams (#36) Hello and thank you for sharing this code. I would like to train this repo on my own dataset. I downloaded the resnet18-5c106cde.pth and converted it to this paddle's model : resnet18-torch.pdparams. I put this weight file in the pretrained folder and seems everything is ok. But when I run the train.py file , I have this error RuntimeError: Parameter file [ ./work_dir/ff_c23/Best_model.pdparams ] not exists What is the problem ? Thanks in advance. Hi I have the same issue. did you able to solve it?! — You are receiving this because you commented. Reply to this email directly, view it on GitHubhttps://github.com/VIS-VAR/LGSC-for-FAS/issues/36#issuecomment-887763298, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AQY7X3TQ6JFZJZ2HJNJXQUTTZ372PANCNFSM436PPRSQ. You can remove all code in fuction init_model() of runner.py and replace by pass when you train: def init_model(): pass … ________________________________ Từ: morteza golzan @.> Đã gửi: 28 Tháng Bảy 2021 2:08 SA Đến: VIS-VAR/LGSC-for-FAS @.> Cc: Nguyen Duc Giang 20172523 @.>; Comment @.> Chủ đề: Re: [VIS-VAR/LGSC-for-FAS] Best_model.pdparams (#36) Hello and thank you for sharing this code. I would like to train this repo on my own dataset. I downloaded the resnet18-5c106cde.pth and converted it to this paddle's model : resnet18-torch.pdparams. I put this weight file in the pretrained folder and seems everything is ok. But when I run the train.py file , I have this error RuntimeError: Parameter file [ ./work_dir/ff_c23/Best_model.pdparams ] not exists What is the problem ? Thanks in advance. Hi I have the same issue. did you able to solve it?! — You are receiving this because you commented. Reply to this email directly, view it on GitHub<#36 (comment)>, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AQY7X3TQ6JFZJZ2HJNJXQUTTZ372PANCNFSM436PPRSQ. Thank you for your response. Actually, I found out that 'load_from' property from checkpoint_cfg dict in train.py file, loads the best model even if you have not trained the model previously. Therefore changing its value from './work_dir/ff_c23/Best_model' to None in your first train solves this problem. Moreover, your answer works in the same way. Thanks again.
gharchive/issue
2021-05-01T16:41:47
2025-04-01T04:33:10.788983
{ "authors": [ "giangnd1808", "mortezagolzan", "naserpiltan" ], "repo": "VIS-VAR/LGSC-for-FAS", "url": "https://github.com/VIS-VAR/LGSC-for-FAS/issues/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2060874504
Unable to Run Tutorial.ipynb Hello, I am not able to run the tutorial notebook. The error message is: I am using python 3.8.18 and pytorch 1.8.0 via (pip install torch==1.8.0+cu111 torchvision==0.9.0+cu111 torchaudio==0.8.0 -f https://download.pytorch.org/whl/torch_stable.html). A list of my conda environment packages is attached. pae_env_packages.txt Note that I have tried various versions of pytorch and I either receive the attached error, or an error regarding the mode not being boolean. Can you help resolve this error? If there is a docker container or a specification file of the exact environment you used to run the notebooks, that would be helpful. Thanks in advance. Following up on this issue.
gharchive/issue
2023-12-30T22:36:55
2025-04-01T04:33:10.813551
{ "authors": [ "dawson-miller" ], "repo": "VMBoehm/PytorchPAE", "url": "https://github.com/VMBoehm/PytorchPAE/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2433191647
GH-592 Add optional delete node confirmation dialog Fixes #592 Applied to main (=2.1) and branch 2.0.
gharchive/pull-request
2024-07-27T02:22:48
2025-04-01T04:33:10.899596
{ "authors": [ "Naros" ], "repo": "Vahera/godot-orchestrator", "url": "https://github.com/Vahera/godot-orchestrator/pull/599", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1036407628
New words in the Spelling rule Modify the Spelling rule Fix the procedure: Extending the Spelling rule Ignore some files in tools/report_vale_alerts.sh 10 days without a comment, I guess it's OK to merge :)
gharchive/pull-request
2021-10-26T14:58:25
2025-04-01T04:33:10.902902
{ "authors": [ "themr0c" ], "repo": "Vale-at-Red-Hat/vale-at-red-hat", "url": "https://github.com/Vale-at-Red-Hat/vale-at-red-hat/pull/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1129195748
Cannot walk on certain valley blocks The problem from 2.2.0 that occured on Mossy blocks, and was fixed on 2.2.1 now exists for different blocks.. as shown in the video, cannot walk over these blocks. https://cdn.discordapp.com/attachments/889682144031551548/940824662080909332/Another_Quality_Modpack_2_-_1.18.1_2022-02-08_23-16-11.mp4 Are you going to fix or should I remove the mod from my modpack?
gharchive/issue
2022-02-09T23:32:41
2025-04-01T04:33:11.197078
{ "authors": [ "itsdinkd" ], "repo": "ValleyCraft-Dev-Team/ValleyCraft", "url": "https://github.com/ValleyCraft-Dev-Team/ValleyCraft/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
796065723
Unreal: I can't include PhononProbeVolume We have buildings with large numbers of rooms with known bounding boxes. Our goal is to automatically spawn PhononReverbVolumes to the map. Unfortunately. I seem to be unable to include the PhononProbeVolume.h. I suppose this is due to it being put in the Private folder. Is it possible to move this class (and possibly other classes that might be of interest to other users with similar needs) to the public folder and expose them to the project and other plugins? We've just released Steam Audio 4.1.0, which includes an updated Unreal Engine plugin: https://github.com/ValveSoftware/steam-audio/releases/tag/v4.1.0. Try it out and let us know if you continue to face this issue in the new version. Closing pending feedback.
gharchive/issue
2021-01-28T14:38:47
2025-04-01T04:33:11.389289
{ "authors": [ "kisak-valve", "lakulish", "simongloor" ], "repo": "ValveSoftware/steam-audio", "url": "https://github.com/ValveSoftware/steam-audio/issues/145", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2249252347
Data Structure Writeup Link to Blog 1.92/2 Vance your blog is very organized, clean, and has perfect screenshots of each requirement. I quickly understood what your feature was about, and anyone who doesn't know Clash Royale would be informed and would understand quite easily. I like the demo u did with your frontend, backend, and chrome browser as it shows. Only thing that you could improve was to make the images on the blog load as some didn't initially, or make it clear which feature was involved in each screenshot.
gharchive/issue
2024-04-17T20:54:18
2025-04-01T04:33:12.330413
{ "authors": [ "SriS126", "VanceReynolds" ], "repo": "VanceReynolds/cspblog2", "url": "https://github.com/VanceReynolds/cspblog2/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1691473361
pd_ file support regression after the pdb-ng refactor we lost support for cab compressed PDB files (named pd_ in symbol servers) Workaround for now: use cabextract to extract the pdb, and then load that. You will have to manually pull the file though.
gharchive/issue
2023-05-01T22:05:10
2025-04-01T04:33:12.368391
{ "authors": [ "CouleeApps", "psifertex" ], "repo": "Vector35/binaryninja-api", "url": "https://github.com/Vector35/binaryninja-api/issues/4275", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2225559039
Better user docs on offset pointers We've had offset pointers since 3.2 but it doesn't look like have very good documentation on them: https://binary.ninja/2022/10/28/3.2-released.html#offset-pointers Duplicate of #3834
gharchive/issue
2024-04-04T13:42:24
2025-04-01T04:33:12.369535
{ "authors": [ "CouleeApps", "plafosse" ], "repo": "Vector35/binaryninja-api", "url": "https://github.com/Vector35/binaryninja-api/issues/5236", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1335282998
[NewPlugin] LittleDyingDuck BinaryNinja-RichPrecense plugin Please add my plugin to the repository. Yes, I've read the instructions, especially the part about creating an actual release and creating the tag. I either used the UI, or the hub command to create a real release. Author: Me Repo name: BinaryNinja-RichPrecense Release: 1.0 Hiya, please make a release on your plugin's GitHub repo. Hiya, please make a release on your plugin's GitHub repo. If you mean this, then there ya go. Yep, looks good. Thanks for the plug-in!
gharchive/issue
2022-08-10T22:53:45
2025-04-01T04:33:12.372822
{ "authors": [ "CouleeApps", "LittleDyingDuck" ], "repo": "Vector35/community-plugins", "url": "https://github.com/Vector35/community-plugins/issues/191", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
188743982
S3 integration not working and not throwing any error. I am trying to set-up an AWS S3 integration and it isn't working. Here is the config: https://gist.github.com/PolGuixe/0acfb2070861a563f0184789cd5bb177 When I upload a file, the document es created in the collection. Uploaded to the server and then "transferred" to S3, however it never arrives to my S3 bucket. Then the file in the server is deleted and the document points to a non existing file. Any idea why my integration isn't working? Shouldn't the server throw and error if it is unable to access the server? Knox does not supper S3 in eu-central-1. See: https://github.com/Automattic/knox/pull/273
gharchive/issue
2016-11-11T11:51:31
2025-04-01T04:33:12.383226
{ "authors": [ "PolGuixe" ], "repo": "VeliovGroup/Meteor-Files", "url": "https://github.com/VeliovGroup/Meteor-Files/issues/280", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1415308770
Unclear usage for NUnit TestCase I'm really struggling to understand how to use TestCase attributes as documented here. My test is using Bddfy alongside NUnit, with my test initialiser looking like this: [TestCase((string)null!)] [TestCase("")] [TestCase("Some string which breaks the fifty character limit on names")] public void Run(string invalidName) { _invalidName = invalidName; this.BDDfy<CreateBookStory>(); } The method to invoke Verify looks like the following: [AndThen] public Task AndThenTheRelevantValidationMessageIsReturned() { return Verify(_validationErrorResponse); } I cannot figure out how I can pass a value to Verify to parameterise the test file name, which seems simple to do with other libraries with a .UseParameters(arg) call. can u share the full class? also what package are you using? @scottyearsley bump Hi Simon, really sorry for not getting back to you sooner. Thankfully I have been able to figure out my issue and move on. Basically, I mistook an error coming out of Verify, which manifested as a Duplicate item has already been added to the Dictionary JSON serialization exception, to be an issue with the args I was passing in with the test cases and thought I was doing something wrong. I wondered if there was a clash of parameters I was sending in, i.e. null string and "" being seen mishandled. However, it actually turned out to be an issue with the object under test itself (which is my type called RestException, i.e. _validationErrorResponse in my posted code) when it was being serialized using Verify's internal serializer. Interestingly this type serializes to a string using Newtonsoft happily, but not with System.Text.Json. I accepted that the issue is with my type so I mapped the values from it into an anonymous object and verified against that instead, so all good, so please close this issue. Just want to say what a great library Verify is and how much cleaner and easier to write our acceptance tests are using it! Just a thought, it would be really great if you could somehow specify an arg to be used for 'naming' the test case rather that the test value being used for the param of the received files. So something like: [TestCase("Null string", (string)null!)] [TestCase("Empty string", "")] [TestCase("More than 50", "Some string which breaks the fifty character limit on names")] public void Test(string testName, string testValue) { _testName = testName; _testValue = testValue } return Verify(_response, testName: _testName); The received files could be clearer to read and potentially shorter - my example of "Some string which breaks the fifty character limit on names" is pretty long and I have an example where I pass in a string > 250 chars. you can fully control all parts of the verified file name https://github.com/VerifyTests/Verify/blob/main/docs/naming.md https://github.com/VerifyTests/Verify/blob/main/docs/parameterised.md#overriding-text-used-for-parameters UseTextForParameters() is perfect, thank you 👍
gharchive/issue
2022-10-19T16:58:44
2025-04-01T04:33:12.400223
{ "authors": [ "SimonCropp", "scottyearsley", "syearsley" ], "repo": "VerifyTests/Verify", "url": "https://github.com/VerifyTests/Verify/issues/678", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2001454945
运行生成的二维码太大了 已经把他拉到最大了,还是有一部分显示不出来 修改pycharm控制台的行高 已经把他拉到最大了,还是有一部分显示不出来 建议用windows terminal, 或者你试试cmd能不能直接ctrl+滚轮缩放,terminal能缩放,所以扫到了
gharchive/issue
2023-11-20T06:23:53
2025-04-01T04:33:12.402666
{ "authors": [ "Sorrow-Scarlet", "xym520520", "zx2023qj" ], "repo": "VermiIIi0n/fuckZHS", "url": "https://github.com/VermiIIi0n/fuckZHS/issues/83", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1340845893
fix: Can't replace very_good_cli generated app icon I have generated my Flutter app using very_good_cli. I'm using flutter_launcher_icon to change the default app icon. I make sure that the old icon does not exist in the project, but still, the very_good_cli icon appears after reinstalling the app. Where does it get this icon if it does not exist in the project? What are the steps I need to take to change the app's icon that was generated by very_good_cli? I have a similar issue with the splash screen, where I don't really know how to change the splash screen Drawable to something custom. Any leads on that @CartmanGD ? Using flutter_native_splash screen worked for me. However I didn't tested it with different flavors. Experiencing this same issue but only on android You may also need to uninstall and reinstall the app or restart the emulator as sometimes assets can get cached. So I checked and I realized that my mipmap-anydpi-v26 had the initial svg icon, so apparently I still had the icon and my device was building to that icon instead. So if you have the issue it's probably coming from this folder. Going to close this for now since there aren't any actionable next steps on your end. Feel free to comment with any follow-up questions and we can continue the conversation 👍 @CartmanGD were you able to fix the issue with the responses above? Yes, @k1ycee comment was really helpful. Thanks guys!
gharchive/issue
2022-08-16T20:28:12
2025-04-01T04:33:12.413318
{ "authors": [ "CartmanGD", "felangel", "k1ycee", "rajarshi-ch", "saveKenny" ], "repo": "VeryGoodOpenSource/very_good_cli", "url": "https://github.com/VeryGoodOpenSource/very_good_cli/issues/485", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
936146109
react 17 update Dependency and jest setup update for react 17. Uses unofficial @wojtekmaj/enzyme-adapter-react-17 https://github.com/VictorCazanave/react-svg-map/issues/53 [x] Fork the repository [x] Develop your code changes [x] Ensure eslint is happy: npm run lint [x] Ensure the tests pass: npm test [x] Commit your changes [x] Push to your fork [x] Submit a pull request Thanks for your contribution @RamonDonnell ! 😃 Does it solve the issue https://github.com/VictorCazanave/react-svg-map/issues/53? If yes, since the components aren't modified, can't we just update the peerDependencies to support both React 16 and 17 (^16.0.0 || ^17.0.0)? 🤔 Please resolve and merge this pull request. I'd like to use react-svg-map with react-17, and I'm sure many other people would too. Yes, other people too ;p As mentioned in https://github.com/VictorCazanave/react-svg-map/issues/53, I didn't see any issue using these components with React 17. So @sdnalloh @piotrgregor what do you need exactly? Only updating the peer dependencies to also support React 17? If there is an issue with React 17, please provide a repository with a minimal example to reproduce it. Thanks again for your contribution @RamonDonnell, but stopping React 16 support is a major change that will be done in the v3
gharchive/pull-request
2021-07-03T02:06:38
2025-04-01T04:33:12.444316
{ "authors": [ "RamonDonnell", "VictorCazanave", "piotrgregor", "sdnalloh" ], "repo": "VictorCazanave/react-svg-map", "url": "https://github.com/VictorCazanave/react-svg-map/pull/54", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1750983366
super burger landing page Super Burger Landing Page #248 added the super hamburger landing page Create Readme.md for this
gharchive/pull-request
2023-06-10T14:02:10
2025-04-01T04:33:12.457509
{ "authors": [ "Shu12388y", "Vikash-8090-Yadav" ], "repo": "Vikash-8090-Yadav/Future.WebNet", "url": "https://github.com/Vikash-8090-Yadav/Future.WebNet/pull/300", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1977620465
🛑 K-Jet is down In 058cf40, K-Jet ($KJET) was down: HTTP code: 0 Response time: 0 ms Resolved: K-Jet is back up in 0370cdd after 14 minutes.
gharchive/issue
2023-11-05T03:56:38
2025-04-01T04:33:12.459562
{ "authors": [ "Vikingfr" ], "repo": "Vikingfr/upptime", "url": "https://github.com/Vikingfr/upptime/issues/1824", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1236143186
🛑 Rue de la coloc is down In e849dd9, Rue de la coloc ($COLOC) was down: HTTP code: 0 Response time: 0 ms Resolved: Rue de la coloc is back up in 552a6c4.
gharchive/issue
2022-05-14T23:38:15
2025-04-01T04:33:12.461749
{ "authors": [ "Vikingfr" ], "repo": "Vikingfr/upptime", "url": "https://github.com/Vikingfr/upptime/issues/587", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
855498671
What if I want to train NER task only? Hello, thanks for publishing the code. I have a question, plz help me clarify this. What should I change the code if I only want to train NER task not the others? Because PhoNLP uses soft POS embedding for NER and Dependency parsing task, I think you cannot use this code for only NER task. If you only want to train NER task, you should train single NER task based on PhoBERT. I can send you the code for single NER task if you need it. Hi @thelinhbkhn2014, i want to train PhoBERT with NER task. Could you send me sample code to do it? Ok, please let me know your email. I will send my code to you. @thelinhbkhn2014 My email is natuhvnh@gmail.com. Thank you for sharing! @thelinhbkhn2014 hi can u send me the code please? I'm working on Vietnamese NER using PhoBERT for my thesis... and i'm from hust too. My email is buiphoben050199@gmail.com. Thank u! @thelinhbkhn2014 Hi Linh, can you send me your code about NER using phobert? My email is hien.trannguyenle@gmail.com. Thank you very much Hi, I'm trying to train a NER model in Vietnamese. Can you send me your code. My email is huyhoang240101@gmail.com. Thank you so much. @thelinhbkhn2014 Hi, I'm working about sequence labeling task. Can you send me your code about NER using phobert? My email is chuvanan29081998@gmail.com Thank you! Hi, I'm trying to train NER model for Vietnamese only. Can you send me the code and a document about the NER using phoBERT please. My email is vutuanhuyvth@gmail.com Thank you so much! @thelinhbkhn2014 I'm trying to train NER for the COVID-19 dataset. Can you send me your NER code using phobert? I would really appreciate it. My email is davido111200@gmail.com Thank you very much! @thelinhbkhn2014 I'm trying to train NER for a geological dataset. Can you send me your NER code using phobert? I would really appreciate it. My email is renjiangtao@cug.edu.cn Thank you very much! Ok, please let me know your email. I will send my code to you. Hello, I am trying to reproduce the NER task, Can you send me your NER code using PhoBERT? My email is shed-e@qq.com Thank you very much!
gharchive/issue
2021-04-12T03:01:06
2025-04-01T04:33:12.479808
{ "authors": [ "Davido111200", "Huyne459", "chuan298", "cugrjt", "huyhoang240101", "icyda17", "lehoangh", "natuhvnh", "sentry0501", "shed-e", "thelinhbkhn2014" ], "repo": "VinAIResearch/PhoNLP", "url": "https://github.com/VinAIResearch/PhoNLP/issues/4", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1105819924
More useful hints? I wonder if a more useful hint that had some time penalty, like "think of :nth-of-type" would help? Or maybe explaining slightly what the pattern is without saying the actual solution I'll improve the hints somewhen this week - but time penalties are not planned There are now two hints on each level - the first one gives a small tooltip with text and the second one is the MDN link
gharchive/issue
2022-01-17T12:53:54
2025-04-01T04:33:12.490344
{ "authors": [ "Haroenv", "Vincenius" ], "repo": "Vincenius/css-speedrun", "url": "https://github.com/Vincenius/css-speedrun/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
103862684
$beforeValidation and $beforeInsert I have question about validating a password before the user is inserted. I have the following insert: // Insert request User.query() .insert({name: 'Jon', 'email: email@testing.com', password: '123'}) .then(function (user) { // Validations passed and user is added }) .catch(function(error){ // Validations did not pass }); // Validation Schema User.jsonSchema = { type: 'object', required: ['name','email', 'password', 'salt'], properties: { id: {type: 'integer'}, name: {type: 'string', minLength: 2, maxLength: 255}, email: {type: 'string', minLength: 3, maxLength: 255}, password: {type: 'string', minLength: 6, maxLength: 255}, salt: {type: 'string', minLength: 3, maxLength: 255}, } }; I would like to be able to validate the password and add a salt but ONLY on an insert. I can not seem to figure if jsonScheme has the ability to pass a function to the properties.password or not. If not I need to detect when I am in $beforeValidate if I am in an insert state or an update state because the password will be an option on update, ie not change it if it is empty. Is $beforeInsert just a callback or can it cause it to reject the promise or throw an error that will be caught in the promise? If so I can probably do the password validations and the adding of the salt in there manually. Thanks! I think I got it. I removed the require for password and salt and then in $beforeInsert I test for password and I can just throw an error that gets caught by the promises.
gharchive/issue
2015-08-29T16:46:16
2025-04-01T04:33:12.492835
{ "authors": [ "jzumbrun" ], "repo": "Vincit/objection.js", "url": "https://github.com/Vincit/objection.js/issues/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1582204217
Handle Slide In Unmanaged Notification in the non-selected repo 1. As for now, only the snapshot of the repository selected in Git Machete panel is available (not all snapshots of all repositories!). The unmanaged branch notification works on the same snapshot as the one selected in Git Machete panel. Hence, we must assure that the current branch changed belongs to the same repository as the given snapshot. A handling of all repositories (not only selected) is a subject to of improvement. Originally posted by @mkondratek in https://github.com/VirtusLab/git-machete-intellij-plugin/pull/1531#discussion_r1104364628 A handling of all repositories Also, there's a question if that's really what the users are willing to see... multiple notifications popping up at once, for multiple VCS roots (some of which they might not be interested in managing under Git Machete at all) 🤔 multiple notifications popping up at once nothing like this for sure! Let's recall the (single line Xd) scenario here: Checkout an unmanaged branch from a repository that is not currently selected in Git Machete tab. (this can be done via DVCS widget or command line) So as one can perform only a single checkout at once there would be only a single notification at that time.
gharchive/issue
2023-02-13T11:50:50
2025-04-01T04:33:12.538313
{ "authors": [ "PawelLipski", "mkondratek" ], "repo": "VirtusLab/git-machete-intellij-plugin", "url": "https://github.com/VirtusLab/git-machete-intellij-plugin/issues/1542", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1652823736
Is our plugin dynamic? Why an IDE restart is always required once the plugin is reloaded? https://plugins.jetbrains.com/docs/intellij/dynamic-plugins.html IIRC we asked us this question years ago 😅 For now every time I update the plugin I must restart the IDE (the plugin tab disappears). On the other hand (however I am not 100% sure) I think that one is able to install and use the plugin without restarting. Huh plugin verifier (executed in every CI run) says: Plugin can probably be enabled or disabled without IDE restart 😅
gharchive/issue
2023-04-03T21:52:17
2025-04-01T04:33:12.540740
{ "authors": [ "PawelLipski", "mkondratek" ], "repo": "VirtusLab/git-machete-intellij-plugin", "url": "https://github.com/VirtusLab/git-machete-intellij-plugin/issues/1631", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2292811069
[Feature] 树形数据中图标的注册,与图标展示位置期望可配置 What problem does this feature solve? 可以展示一个动态生成的用户文字头像 What does the proposed API look like? 期望icon中可自定义回调函数生成,并且可以获取到当前单元格的数据。 期望树形数据中,icon可以跨越展开图标,在左侧紧紧贴合文字。 第一条中提到的,icon本身就是可以配置函数的吧 感谢您的回复,第一个确实是可以的,已经完成了,现在只需要让图标与文字贴合了,这个有什么办法吗
gharchive/issue
2024-05-13T13:21:55
2025-04-01T04:33:12.542771
{ "authors": [ "fangsmile", "liudengyan" ], "repo": "VisActor/VTable", "url": "https://github.com/VisActor/VTable/issues/1697", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1722540364
[PATTERN] Add a Binary number pattern in Java, Pattern40 Description Task for the given input n the following pattern should print (n=5 for the below pattern) Pattern Image Filename Pattern40.java @Vishal21121 finished, will send pr after assign ok @livvyaaa you can work on this issue.
gharchive/issue
2023-05-23T17:58:19
2025-04-01T04:33:12.545577
{ "authors": [ "Vishal21121", "livvyaaa" ], "repo": "Vishal21121/Patterns-different-language", "url": "https://github.com/Vishal21121/Patterns-different-language/issues/90", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
452463394
Deception IV: Blood Ties [PCSE00401] Game summary Game name: Deception IV: Blood Ties Game serial: PCSE00401 Game version: 1.00 Vita3K summary Version: v0.1 Build number: 1063 Commit hash: https://github.com/vita3k/vita3k/commit/050c22b3 Test environment summary Tested by: Zangetsu38 OS: Windows 10 CPU: AMD Ryzen 2700X GPU: NVIDIA GTX 1080 Ti RAM: 16 GB DDR4 3600 mhz Issues Crash on genreate shader. Log vita3k.log Vita3K v0.1.5 2688-6ec6eddd crashes out nearly straight away vita3k.log Vita3K summary Version: v0.1.8 Build number: 3150 Commit hash: https://github.com/vita3k/vita3k/commit/1b7e038d CPU backend: Dynarmic Test environment summary Tested by: othmanead OS: Windows 11 CPU: Intel i5-10300H GPU: RTX 3060 mobile RAM: 16 GB Issue: Game plays with audio bugs and minor graphical bugs Screenshots: Log vita3k.log Recommended labels Ingame + Audio Bug Graphics Bug
gharchive/issue
2019-06-05T12:02:17
2025-04-01T04:33:12.564395
{ "authors": [ "Zangetsu38", "benderscruffy", "othmanead" ], "repo": "Vita3K/compatibility", "url": "https://github.com/Vita3K/compatibility/issues/200", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
476441750
A Virus Named TOM [PCSE00501] Game summary Game name: A Virus Named TOM Game serial: PCSE00501 Game version: 01.00 Vita3K summary Version: v0.1 Build number: 1190 Commit hash: https://github.com/vita3k/vita3k/commit/699475a Test environment summary Tested by: IllusionMan1212 OS: Windows 10 Pro CPU: AMD Ryzen 5 1600 GPU: NVIDIA GTX 1050Ti RAM: 8 GB Issues Stuck infinitely on loading screen Screenshots Log vita3k.log Vita3K v0.1.5 2601-f803e615 goes ingame and is playable only problem so far is if you dont start a new game straight away the emulator will crash out vita3k.log Game crashes after a few seconds of running. OS: Windows 11 Pro CPU: Intel i7-6700 GPU: Nvidia RTX A2000 RAM: 16GB vita3k.log
gharchive/issue
2019-08-03T11:13:07
2025-04-01T04:33:12.573479
{ "authors": [ "IllusionMan1212", "InfernoShark", "benderscruffy" ], "repo": "Vita3K/compatibility", "url": "https://github.com/Vita3K/compatibility/issues/260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1268131414
Star Ocean Second Evolution [PCSG00714] App summary App name: Star Ocean Second Evolution App serial: PCSG00714 App version: 1.0 Vita3K summary Version: v0.1.5 Build number: 2656 Commit hash: https://github.com/vita3k/vita3k/commit/d6200b6 CPU backend: Dynarmic Test environment summary Tested by: gymzatan OS: Windows 11 CPU: AMD Ryzen 7 5800H GPU: AMD Radeon(TM) Graphics RAM: 16 GB Issues Game plays at full speed, with minor text and graphic issues: Certain menu texts are not displayed; Dialogue text disappears later than the frame; A transient yellow screen area during battle onset. Screenshots Log vita3k.log Recommended labels Ingame + text missing graphics bug The missing text issue is gone, so the game is technically fully playable now. The minor graphics issues still exist but do not affect gameplay in any way. App summary App name: Star Ocean Second Evolution App serial: PCSG00714 App version: 1.0 Vita3K summary Version: v0.1.5 Build number: 2683 Commit hash: https://github.com/vita3k/vita3k/commit/b3a44b5 CPU backend: Dynarmic Test environment summary Tested by: gymzatan OS: Windows 11 CPU: AMD Ryzen 7 5800H GPU: AMD Radeon(TM) Graphics RAM: 16 GB Issues Game plays at full speed, with minor text and graphic issues: Dialogue text disappears later than the frame; A transient yellow screen area during battle onset. Screenshots Log vita3k.log Recommended labels Playable graphics bug
gharchive/issue
2022-06-11T01:44:51
2025-04-01T04:33:12.585454
{ "authors": [ "gymzatan" ], "repo": "Vita3K/compatibility", "url": "https://github.com/Vita3K/compatibility/issues/708", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
55032524
Improve UI on iOS devices I removed the separation between the buttons and set the background to white. I believe this UI more closely aligns with what users on iOS devices would expect. Could you provide a screenshot please?
gharchive/pull-request
2015-01-21T15:23:42
2025-04-01T04:33:12.586573
{ "authors": [ "Chun-Yang", "Hirbod" ], "repo": "VitaliiBlagodir/cordova-plugin-datepicker", "url": "https://github.com/VitaliiBlagodir/cordova-plugin-datepicker/pull/55", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
984835954
Update code challenge. I have changed it a bit. I don't want to constrain them. Also, I am against giving them too much help unless they request it, it will show their ability to ask and their independence. I think its a good idea to align this with our supply chain work somewhat. In general, I would bias towards providing more information on what we're hoping for, as that will help us get more consistent results and avoid misunderstanding. e.g. MapSPAM is crop production statistics, so by analysis we're essentially asking for summary stats for these AOI? With the fires there was a real time component, so there could be trends and some sort of causal commentary. With crops they'll have to compare across technologies or crop types. Most of these comments/edits otherwise are minor details I think and won't make a much of a difference either way. So, I think this is good to merge, but I would leave ownership/final edit to @tamaramegan as she is leading the hiring process.
gharchive/pull-request
2021-09-01T08:00:46
2025-04-01T04:33:12.591641
{ "authors": [ "aagm", "fgassert" ], "repo": "Vizzuality/science-code-challenge", "url": "https://github.com/Vizzuality/science-code-challenge/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1043082180
Update sbt-scoverage to 1.9.2 Updates org.scoverage:sbt-scoverage from 1.9.1 to 1.9.2. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scoverage", artifactId = "sbt-scoverage" } ] labels: sbt-plugin-update, early-semver-patch, semver-spec-patch Codecov Report Merging #165 (abb0548) into master (dc1f3d9) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #165 +/- ## ========================================= Coverage 100.00% 100.00% ========================================= Files 1 1 Lines 41 41 ========================================= Hits 41 41 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update dc1f3d9...abb0548. Read the comment docs. Superseded by #171.
gharchive/pull-request
2021-11-03T04:45:07
2025-04-01T04:33:12.607717
{ "authors": [ "codecov-commenter", "scala-steward" ], "repo": "VladKopanev/zio-saga", "url": "https://github.com/VladKopanev/zio-saga/pull/165", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1525953849
CI: first step of CI/CD check all test pass every time a pull request is created or something is pushed into main Some small changes left, but I really appreciate this PR @Paker30 :+1:
gharchive/pull-request
2023-01-09T16:51:15
2025-04-01T04:33:12.609948
{ "authors": [ "Paker30", "VladimirMikulic" ], "repo": "VladimirMikulic/route-list", "url": "https://github.com/VladimirMikulic/route-list/pull/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2216713680
Request: option in right-click menu to do a player lookup in LFG UI Would be nice to perform a player lookup from the right-click menu when looking at players in LFG queue. Both when seeking a group, or listing one. Similar to "Copy Radier.IO URL" in the screenshot here. I am currently working on this but don't have too much time atm. Will be implemented at some point (hopefully soon) Added in alpha release https://github.com/Vloer/keycount/releases/tag/v0.8-alpha If you don't mind, please download the alpha version (you have to manually select version via curseforge) and see if this works correctly. Note: the popup only shows when data is available (you can test this on yourself) I will test this later, but initial thought is that the menu option should always be there for efficiency (and other reasons). Correct me if I'm wrong regarding the current described implementation, but it seems the addon would be performing a lookup every time a right-click (on a valid object) occurs, regardless of whether or not the user clicks the menu option. Ideally the lookup should only occur if the user actually selects the option. This may not be an immediate concern due to the size of the data being parted initially, but as it grows over time it could introduce even a small hit on resources that shouldn't otherwise be consumed. I will test this later, but my initial thought is that the menu option should always be there and the lookup should only happen if it is selected, primarily for efficiency. Correct me if I'm wrong regarding the current described implementation, but it seems the addon would need to perform a lookup every time a right-click (on a valid object) occurs in order enable the menu option, regardless of whether or not the user selects the menu option. For efficiency, ideally the lookup should only occur if the user actually selects the option. This may not be an immediate concern due to the size of the data being parsed initially, but as it grows over time this could introduce even a small hit on resources that shouldn't otherwise be consumed. Im aware of the bowl of spaghetti currently being thrown around; this was the easiest way to implement the menu only showing up when data is available. The first lookup should be performant enough until like a 100k players grouped with, the addon will require other optimizations way before that so it's not really an issue atm. I'm going to check wtih some other people if they prefer the option being always visible or just when data is avialable. Thanks for the feedback
gharchive/issue
2024-03-31T00:16:44
2025-04-01T04:33:12.619854
{ "authors": [ "Vloer", "derek-shnosh", "rideekulous" ], "repo": "Vloer/keycount", "url": "https://github.com/Vloer/keycount/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2267478906
TIFF image dont shown Describe the bug TIFF image dont shown. Expected behavior Add support for TIFF image. Desktop (please complete the following information): OS: windows 10 Browser egde Version 110.0.1587.57 Additional context Maybe use someelse js script, like tiff.js? Really don't want to add another dependency to project. If it's required I would recommend to add preprocessing code that will repack docx package with converting tiff to any other supported format.
gharchive/issue
2024-04-28T10:33:17
2025-04-01T04:33:12.627089
{ "authors": [ "VolodymyrBaydalka", "changlichun" ], "repo": "VolodymyrBaydalka/docxjs", "url": "https://github.com/VolodymyrBaydalka/docxjs/issues/111", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
875116693
Add main Add main Add a main block
gharchive/issue
2021-05-04T05:29:30
2025-04-01T04:33:12.627790
{ "authors": [ "VolodymyrIvanitskyi" ], "repo": "VolodymyrIvanitskyi/homepage", "url": "https://github.com/VolodymyrIvanitskyi/homepage/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1842140981
CHORE: Add All bots to Better Uptime for real-time tracking CHORE: Add All Robot endpoints to Better Uptime so that if anything does down, we'll know. Limit Order Bot http://check.bots.palomachain.com/limit-order-bot-script/ Momentum Order Bot http://check.bots.palomachain.com/momentum-order-bot-script/ Stop Loss Bot http://check.bots.palomachain.com/stop-loss-bot-script/ Gecko Bot http://check.bots.palomachain.com/gecko-bot-script/ Curve Stop Loss Bot http://check.bots.palomachain.com/curve-stop-loss-script/
gharchive/issue
2023-08-08T21:41:16
2025-04-01T04:33:12.630438
{ "authors": [ "rootedbox", "taariq" ], "repo": "VolumeFi/paloma", "url": "https://github.com/VolumeFi/paloma/issues/625", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
119798994
setting a custom style attribute (not through existing commands) I hoped I could do: composer.commands.exec("formatInline", "span", {styleProperty: 'fontFamily', styleValue: 'Lato'}); but unfortunately it's not that easy, http://jsfiddle.net/crl/foafka4a/72/ After checking https://github.com/Voog/wysihtml/blob/master/src/commands/formatInline.js#L139, I see that you can only set a class, why this limitation? It would be extremely pleasant to be able to manipulate styles directly composer.commands.exec("formatInline", {nodeName: "span", styleProperty: 'fontFamily', styleValue: 'Lato'}); Great, thanks pulges, I was starting to make it using classes :) other thing, why isn't toggle working? http://jsfiddle.net/crl/foafka4a/80/ I can still manage to remove it manually, but curious how this option work I have this now http://jsfiddle.net/crl/foafka4a/86/ but I notice that you lose fonts initially like mentioned in https://github.com/Voog/wysihtml/issues/262, but also when going in the 'source' view It's probably related, I guess the font-family should be made as a command for making this work well? You need to configure the parser configuration object (wysihtml5ParserRules), so it will allow font-family style in span tags. You can look at https://github.com/Voog/wysihtml/blob/master/parser_rules/advanced_and_extended.js#L311 for reference. ah, thanks :), I guessed it had to do with the parser the rules are used to configure cleanup. It is mostly a whitelist of allowed tags and parameters and everything not in in will be removed. makes sense, I saw it can remove also some wrapping, thanks again pulges, (can you have a look at the alternative version,your answer is anyway helpful since the parser also removes classes if you don't allow them, but I'm afraid I'm not understanding well how the exec/state of a command work yet) You should probably look at https://github.com/Voog/wysihtml/blob/master/examples/wotoolbar.html#L357 For an example of how to call commands on editor directly with javascript not built in toolbar (exec) In some cases selection can not be kept in editor on command inserting (for example user has to ender link address into an input). Then you have to save the selection and restore the selection later when user has inserted the value saving selection to variable var bookmark = editor.composer.selection.getBookmark(); restoring selection: editor.composer.selection.setBookmark(bookmark); Thanks I knew about the selection get and set, and about this example:) The idea is to be able to add a font-family command, and maybe others, and be able to maintain a clean DOM (not an increasing number of <span> nesting), I'm hesitating between using style for everything, or classes for everything (both are possible, but that would be more consistent for my case to keep using one way or the other. Thanks for your help @pulges, made this http://jsfiddle.net/crl/foafka4a/88/, I was surprised you don't need anymore to update active classes yourself
gharchive/issue
2015-12-01T20:18:04
2025-04-01T04:33:12.641635
{ "authors": [ "cauburtin", "pulges" ], "repo": "Voog/wysihtml", "url": "https://github.com/Voog/wysihtml/issues/265", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2666575242
🛑 AndroidStory is down In ff885a9, AndroidStory (https://androidstory.dev/) was down: HTTP code: 500 Response time: 1066 ms Resolved: AndroidStory is back up in 3816528 after 14 minutes.
gharchive/issue
2024-11-17T21:40:23
2025-04-01T04:33:12.670275
{ "authors": [ "VovaStelmashchuk" ], "repo": "VovaStelmashchuk/my-services-monitor", "url": "https://github.com/VovaStelmashchuk/my-services-monitor/issues/140", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2754858026
🛑 MixDrinks is down In fe625d7, MixDrinks (https://mixdrinks.org/) was down: HTTP code: 0 Response time: 0 ms Resolved: MixDrinks is back up in 898dfbf after 6 minutes.
gharchive/issue
2024-12-22T21:44:56
2025-04-01T04:33:12.672839
{ "authors": [ "VovaStelmashchuk" ], "repo": "VovaStelmashchuk/my-services-monitor", "url": "https://github.com/VovaStelmashchuk/my-services-monitor/issues/197", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2663812829
🛑 AndroidStory is down In 5b99e8b, AndroidStory (https://androidstory.dev/) was down: HTTP code: 502 Response time: 3603 ms Resolved: AndroidStory is back up in 2a560b0 after 16 minutes.
gharchive/issue
2024-11-16T05:31:32
2025-04-01T04:33:12.675184
{ "authors": [ "VovaStelmashchuk" ], "repo": "VovaStelmashchuk/my-services-monitor", "url": "https://github.com/VovaStelmashchuk/my-services-monitor/issues/82", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2568202489
Request for Partial Real Machine Dataset Release Dear VoxAct-B Team, I've benefited from your open-sourced simulation dataset. Would it be possible to open-source a part of the real machine dataset for reference? It would be greatly helpful for further research. Thanks! Hello, This is the Open Jar dataset used in our real-world experiment.: https://drive.google.com/file/d/1zU_O-tf9AUbyWHa5RI79YqOlCeoGav_j/view?usp=drive_link Training scripts: peract/scripts/train_open_jar_ours_vlm_real_robot_v4_acting.sh peract/scripts/train_open_jar_ours_vlm_real_robot_v4_stabilizing.sh Evaluation script: peract/scripts/eval_open_jar_ours_vlm_real_robot_v4.sh I am extremely interested in the real machine dataset you provided. Firstly, I would like to know if the collection of this real dataset involved teleoperation. Secondly, if teleoperation was used, what is the model of the teleoperation equipment? Thank you very much for your time and I look forward to your response. Yes. GELLO teleoperation framework. Please refer to Appendix.B for real-world experimental details.
gharchive/issue
2024-10-05T16:17:36
2025-04-01T04:33:12.678738
{ "authors": [ "arthur801031", "yanrihong" ], "repo": "VoxAct-B/voxactb", "url": "https://github.com/VoxAct-B/voxactb/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
135509742
Better export @afred? Either SD or HD, not both. Thumbnail URL. hmm, not failing locally... @afred? This should go in before we try to import again.
gharchive/pull-request
2016-02-22T19:02:17
2025-04-01T04:33:12.716593
{ "authors": [ "mccalluc" ], "repo": "WGBH/stock-sales-2", "url": "https://github.com/WGBH/stock-sales-2/pull/324", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1667279766
Add Docker environment checks in various API server client scripts Hi, everyone. I found some ways for a script to check if it's running inside a Docker container. I've implemented one of the methods in admin-tools/misc/getCruiseId.py. If you're satisfied with that approach, I'll apply it to a couple of other scripts that have the same need (admin-tools/misc/getLoweringId.py and admin-tools/sealog_postdive.sh). Seems like a partial fix. The .dockerenv file isn't guaranteed to exist, is it? Do we have to modify the containers to add it? It's not just https:// vs http://. Outside the containers we need to connect to http://localhost/ or https://localhost/ according to the nginx configuration on the system (the AIS requires https, not all systems do). Inside a container it matters which container we are using. If it is the one hosting the server app, then we can use localhost as the hostname. Otherwise we would have to use a different hostname to route the request to the right container. Seems like a partial fix. 1. The `.dockerenv` file isn't guaranteed to exist, is it? Do we have to modify the containers to add it? 2. It's not just https:// vs http://. Outside the containers we need to connect to `http://localhost/` or `https://localhost/` according to the nginx configuration on the system (the AIS requires https, not all systems do). Inside a container it matters _which_ container we are using. If it is the one hosting the server app, then we can use `localhost` as the hostname. Otherwise we would have to use a different hostname to route the request to the right container. To answer point 1): I haven't found any documentation explicitly stating that /.dockerenv is guaranteed to exist but I've seen it in every single container I've opened a Bash session in. Additionally, it appears in several places in the moby source code. Docker/moby adds the file; we aren't manipulating it.
gharchive/pull-request
2023-04-13T23:35:40
2025-04-01T04:33:12.749630
{ "authors": [ "jogarcia-whoi", "rgov" ], "repo": "WHOIGit/ndsf-sealog-scripts", "url": "https://github.com/WHOIGit/ndsf-sealog-scripts/pull/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1315253698
Fix missing passthrough, inconsistent track extra idxs, and improve customization handling for tnp nano This should fix the inconsistencies in the standalone muon track <-> muon matching. I guess this should still be merged? Yes...
gharchive/pull-request
2022-07-22T17:27:06
2025-04-01T04:33:12.836691
{ "authors": [ "bendavid", "kdlong" ], "repo": "WMass/cmssw", "url": "https://github.com/WMass/cmssw/pull/19", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2657655176
Resolved bug 991 : Design System > Components > Calendar > Small mobile view text and alignment issue. Description Resolve the issues on Component: Resolved issue on Components Calendar Small mobile view text and alignment issue. Screenshots : Checklist: [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have made corresponding changes to the documentation [x] My changes generate no new warnings [x] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes Resolved issue on week view
gharchive/pull-request
2024-11-14T05:37:08
2025-04-01T04:33:13.130020
{ "authors": [ "nitinsonawane123" ], "repo": "Wai-Technologies/raaghu-react", "url": "https://github.com/Wai-Technologies/raaghu-react/pull/1095", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1263908251
Add the script for checking expired translations this is a replacement for https://github.com/ppy/osu-wiki/blob/8aea832cb3111986d2fb297e334bf0827fc3008f/scripts/ci/check_outdated_tags.sh Done: lists unmodified translations that should be expired checks that the hashes added to modified translations are valid may expire the articles automatically when run with --auto TODO: environment awareness: check if https://github.com/ppy/osu-wiki/blob/8aea832cb3111986d2fb297e334bf0827fc3008f/.github/workflows/continuous-integration.yml#L91-L92 can be done from inside the script instead (I remember trying to lockpick GitHub event payloads, so that the check doesn't fail on merge commits in the master branch) when the script is used in the osu-wiki repository, it should be invoked with less arguments (--base-commit {sha} or --base-commit {sha} --outdated-since {sha}) I think we need tests for the scripts as well this may replace https://github.com/Walavouchey/osu-wiki-tools/blob/f077b5147d39a3460765ab63b576312f827785f9/wikitools_cli/outdate_translations.py -- that's something for you to check! i think all of the to-dos are pretty much done now. will add more testing and consolidate the scripts into one main console interface later. the first point of the to-do is probably unnecessary, since using --outdated-since in the gh action is fine and also useful locally
gharchive/pull-request
2022-06-07T21:40:52
2025-04-01T04:33:13.135589
{ "authors": [ "TicClick", "Walavouchey" ], "repo": "Walavouchey/osu-wiki-tools", "url": "https://github.com/Walavouchey/osu-wiki-tools/pull/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
258617427
Add a Page titled "Wallaroo and Long-running data processing and other workflows" Vid would like a page about long-running data processing. This should be added to the appendix in a section with that title that contains the following text: Wallaroo was designed to be a highly efficient and general purpose data processing engine, including event-by-event (stream) processing with microsecond response as well as other longer running tasks such as distributed logfile analysis or model training, and even real-time workflows that do not fit into standard real-time streaming topologies. In order to facilitate and simplify the deployment of such applications, we are building out additional APIs and libraries that will be released over the next few releases of Wallaroo. If you have specific needs or use-cases we encourage you to contact us. Closed by #1506
gharchive/issue
2017-09-18T20:52:16
2025-04-01T04:33:13.137436
{ "authors": [ "aturley", "slfritchie" ], "repo": "WallarooLabs/wallaroo", "url": "https://github.com/WallarooLabs/wallaroo/issues/1503", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1851688350
Tron support Hello! Thanks for the amazing package. Does it support tron network? Everything works fine until I try to sign the transaction. WalletConnect V2 is chain agnostic, so it does support tron, are you sure tron_sign is the method you should use? I don't know much about Tron. If you do some sleuthing in their packages, you could likely figure out how to send the request successfully: https://github.com/tronprotocol/tronwallet-adapter/tree/main WalletConnect V2 is chain agnostic, so it does support tron, are you sure tron_sign is the method you should use? I don't know much about Tron. If you do some sleuthing in their packages, you could likely figure out how to send the request successfully: https://github.com/tronprotocol/tronwallet-adapter/tree/main Its a pity, but tronwallet-adapter contains references to build-in method wallet.signTransaction({ transaction }) only. I could not find working example about making and signing transaction on wallet connect v2 while using dart. When we initialize WalletConnectModalService it accepts any method. Neither "tron_signTransaction" nor "tron_sign" does not work with request. This method does not return any error. Does it use correct params or method name? I don't know. final res = await _service.web3App?.signEngine.request( topic: topic, chainId: _chainId, request: SessionRequestParams( method: 'tron_sign', params: json.encode(trx), ), );
gharchive/issue
2023-08-15T15:40:32
2025-04-01T04:33:13.142196
{ "authors": [ "KabaDH", "Luzzotica" ], "repo": "WalletConnect/WalletConnectModalFlutter", "url": "https://github.com/WalletConnect/WalletConnectModalFlutter/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2222043091
MetaMask connecting don't reactive Describe the bug I try to connect MetaMask wallet connecting but don't reactive MeraMask wallet To Reproduce Steps to reproduce the behavior: I installed Modal Package and build I use the app and try to connect MetaMask MeatMask app active but don't ask connecting's question Expected behavior MataMask app ask about connecting. Screenshots Just don't reactive Environment WalletConnectUnity Modal v1.1.2 Is this happening in Unity Editor or on a player like Android, iOS, Windows? Android Wallets used for testing Metamask Android Additional context OKX wallet connecting is normal. When Modal verion 1.1.1, i can connect MetaMask. but When Package upgrade later, i can't MetaMask wallet Hey, @matthew2099, MetaMask has recently released an update that improves WalletConnect integration. Could you please update the MetaMask app and try again to see if the issue persists? @skibitsky Sorry i am too late. I try that and i see the connecting message. but now i'm not sure. but i think maybe it's ok
gharchive/issue
2024-04-03T06:44:49
2025-04-01T04:33:13.147342
{ "authors": [ "matthew2099", "skibitsky" ], "repo": "WalletConnect/WalletConnectUnity", "url": "https://github.com/WalletConnect/WalletConnectUnity/issues/198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1411602284
Set up alerts with Infura/Pokt We should know when we reach out quotas. @chadyj is this something you could help us with? Sorry missed this ping. But I don't have a good idea for this other than manual checking.
gharchive/issue
2022-10-17T13:38:28
2025-04-01T04:33:13.148424
{ "authors": [ "arein", "chadyj" ], "repo": "WalletConnect/rpc-proxy", "url": "https://github.com/WalletConnect/rpc-proxy/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2541029535
ho aggiornato wasabi e adesso nn mi fa piu accedere si è bloccato General Description How To Reproduce? Go to '...' Click on '...' Scroll down to '...' See ERROR Screenshots Operating System Logs Insert Your Logs Here Wasabi Version hi, Please provide more info. Where did you download it from. What exactly happens, is there an error in the logs, etc. Ho scaricato da GitHub si mi esce un errore dicendomi che il mio file è già in uso in un altro processo Il dom 22 set 2024, 15:06 Marnix Croes @.***> ha scritto: hi, Please provide more info. Where did you download it from. What exactly happens, is there an error in the logs, etc. — Reply to this email directly, view it on GitHub https://github.com/WalletWasabi/WalletWasabi/issues/13443#issuecomment-2366779695, or unsubscribe https://github.com/notifications/unsubscribe-auth/BLPWM7AHSSBWUKZLMJNHZBLZX26D5AVCNFSM6AAAAABOUPEVIKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNRWG43TSNRZGU . You are receiving this because you authored the thread.Message ID: @.***> Try to restart your laptop or simply close the old Wasabi from the update manager or close tor. If it doesn't work please provide your logs. Ho provato ma nn funziona l'ho anche disinstallato e poi reinstallato. Come vado ad aggiornarlo si blocca. Nn ho capito cosa vi devo fornire ? Il dom 22 set 2024, 17:05 Turbolay @.***> ha scritto: Try to restart your laptop or simply close the old Wasabi from the update manager or close tor. If it doesn't work please provide your logs. — Reply to this email directly, view it on GitHub https://github.com/WalletWasabi/WalletWasabi/issues/13443#issuecomment-2366828914, or unsubscribe https://github.com/notifications/unsubscribe-auth/BLPWM7EARBTHWHF7BYPDAEDZX3MDZAVCNFSM6AAAAABOUPEVIKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNRWHAZDQOJRGQ . You are receiving this because you authored the thread.Message ID: @.***> Please delete your file <wasabi_data_folder>/Client/BitcoinStore/Main/IndexStore/IndexMain.sqlite and relaunch Wasabi. <wasabi_data_folder> is %APPDATA%/WalletWasabi on Windows or $HOME/.walletwasabi on Linux and MacOs closing as it's a support request. Please comment if you need any more assitance Try new versio E proprio quando faccio l'aggiornamento di questa versione che si blocca e nn mi fa più accedere Il lun 28 ott 2024, 14:33 Turbolay @.***> ha scritto: Try new version 2.3.0.0 — Reply to this email directly, view it on GitHub https://github.com/WalletWasabi/WalletWasabi/issues/13443#issuecomment-2441603974, or unsubscribe https://github.com/notifications/unsubscribe-auth/BLPWM7CII7HQHBMK6E7MSPTZ5YVJBAVCNFSM6AAAAABOUPEVIKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDINBRGYYDGOJXGQ . You are receiving this because you authored the thread.Message ID: @.***>
gharchive/issue
2024-09-22T11:27:42
2025-04-01T04:33:13.162881
{ "authors": [ "MarnixCroes", "Nico-Bit10", "turbolay" ], "repo": "WalletWasabi/WalletWasabi", "url": "https://github.com/WalletWasabi/WalletWasabi/issues/13443", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1428349219
[VDG] Minor release notification We only notify users about major (X.X.X.0) releases in the UI, the main reason is to not have "upgrade fatigue". However, even minor (X.X.X.X) releases are important, and some users want to stay always up to date. Two possible solutions: Show minor release notification only on status icon hover, just like we show Tor disruption on hover (this might still annoy default users who hover on status icon leading to upgrade fatigue). Or add a setting to include minor releases in the regular release notification UI, so it'll look exactly the same as now. The versioning process has changed, is this still something to consider @lontivero ? We discontinued the concept of minor releases. The issue makes it clear how it makes no sense
gharchive/issue
2022-10-29T16:36:21
2025-04-01T04:33:13.165282
{ "authors": [ "Kruwed", "MaxHillebrand", "turbolay" ], "repo": "WalletWasabi/WalletWasabi", "url": "https://github.com/WalletWasabi/WalletWasabi/issues/9425", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
559780522
How to add a reference list at the end of the document and cite them in the main text? Hi, I am trying to add a reference list at the end of the document and cite them in the main text. I have a bib file named "biblio.bib" which has the following content and put into the src folder of the project: @inproceedings{wrigstad2017mastery, Author = {Wrigstad, Tobias and Castegren, Elias}, Booktitle = {SPLASH-E}, Title = {Mastery Learning-Like Teaching with Achievements}, Year = 2017 Then I add the following line to the "compile_pdf.sh" file: --bibliography=src/biblio.bib \ The effect I would like to have is: by writing the following line in the md file, the achievement-driven learning methodology [@wrigstad2017mastery] and running compile_pdf.sh, the line show up in the pdf file as the achievement-driven learning methodology [1] and the pdf has a "Reference" section at the end showing (just as an example, don't need to be exactly the same format as below) [1] Tobias Wrigstad and Elias Castegren. Mastery learning-like teaching with achievements. In SPLASH-E, 2017 However, it doesn't have the effect I wanted. Do you have any suggesstion on this? An example with this kind of references and citations would be really helpful. Thank you very much! The ability to reference literature would greatly add to this neat little project, allowing it to use it for any kind of academic writing. did you also use --filter pandoc-citeproc? (see https://pandoc.org/MANUAL.html "Citations") Cite keys are successfully being resolved using the following command: pandoc test.md -o example.pdf --from markdown --template eisvogel --listings --bibliography=src/biblio.bib (Ensure pandoc-citeproc is installed.) With the example MarkDown file: # Hello the achievement-driven learning methodology [@knuthwebsite] and the following biblio.bib: @misc{knuthwebsite, author = "Donald Knuth", title = "Knuth: Computers and Typesetting", url = "http://www-cs-faculty.stanford.edu/\~{}uno/abcde.html" } gives rise to the following pdf (screenshot): However, I have not yet figured out how to... ... adapt the citation style of both the citation key and reference. ... put references under a dedicated references section. ... have the right tooling and workflow, e.g. auto-complete for citation keys, renaming citation keys Well, https://pandoc.org/MANUAL.html#pandocs-markdown states that including # References in your document will give rise to a references section. That would solve point 2) I'm a little surprised, that pandoc test.md -o example.pdf --from markdown --template eisvogel --listings --bibliography=src/biblio.bibshould work, would have expected something like pandoc test.md -o example.pdf --from markdown --filter pandoc-citeproc --template eisvogel --listings --bibliography=src/biblio.bib (i.e. with extra --filter pandoc-citeproc) ... Regarding citations in markdown and pandoc see https://pandoc.org/MANUAL.html#citation-rendering and https://pandoc.org/MANUAL.html#citations and https://github.com/jgm/pandoc-citeproc/blob/master/man/pandoc-citeproc.1.md Pandoc uses per default the "Chicago style". You find other styles here: https://www.zotero.org/styles (download and use with --csl xyz.csl as stated in https://github.com/Wandmalfarbe/pandoc-latex-template/issues/151#issuecomment-687640102) Pandoc would add the bibliography at the end of the document. If you want it to appear at some point (not the end), you would need to insert a div with id refs like ::: {#refs} ::: or <div id="refs"> </div>
gharchive/issue
2020-02-04T15:06:29
2025-04-01T04:33:13.175858
{ "authors": [ "cagix", "helenxu", "pykong" ], "repo": "Wandmalfarbe/pandoc-latex-template", "url": "https://github.com/Wandmalfarbe/pandoc-latex-template/issues/151", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2376017069
Update Booking Page Update Booking &Reservation Page. Update Destination Page good ok looks ok
gharchive/pull-request
2024-06-26T18:31:00
2025-04-01T04:33:13.177702
{ "authors": [ "Chiashen72", "NorDayana", "Wang-572", "hh1889" ], "repo": "Wang-572/GTGMblog", "url": "https://github.com/Wang-572/GTGMblog/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2612993015
VSCode VSCode vscode 如何对vite + react + typescript 项目进行断点调试 在 VS Code 中对 Vite + React + TypeScript 项目进行断点调试需要一些配置,主要是设置 launch.json 文件以支持调试环境。以下是步骤: 1. 安装必要的扩展 确保已安装以下 VS Code 扩展: Debugger for Chrome(或适用于您浏览器的调试器扩展) ESLint(可选,确保代码质量) 2. 创建 launch.json 首先,您需要创建调试配置文件 launch.json,这是 VS Code 调试的核心配置文件。执行以下步骤: 打开 VS Code,按 Ctrl+Shift+P 并搜索 Debug: Add Configuration。 选择 Chrome,这是一个常用的调试环境。 配置文件 launch.json 可能类似如下: { "version": "0.2.0", "configurations": [ { "type": "chrome", "request": "launch", "name": "Vite + React", "url": "http://localhost:5173", // Vite 默认端口 "webRoot": "${workspaceFolder}/src", "sourceMaps": true, "trace": true, "skipFiles": ["<node_internals>/**"] } ] } 3. 配置 Vite 进行调试 Vite 默认支持调试,但需要确保它生成了可以映射到 TypeScript 源代码的 Source Maps。Vite 中的 sourceMap 通常默认开启,但可以在 vite.config.ts 文件中明确配置: import { defineConfig } from 'vite' import react from '@vitejs/plugin-react' export default defineConfig({ plugins: [react()], build: { sourcemap: true // 确保生成 source maps } }) 4. 启动 Vite 开发服务器 在终端中运行以下命令启动 Vite 开发服务器: npm run dev 5. 启动调试 确保在 src 文件夹中的 TypeScript 或 React 代码上放置断点。 点击 VS Code 左侧调试图标,选择 Vite + React 调试配置,点击“启动调试”按钮。 VS Code 会在 Chrome 浏览器中自动启动调试。 6. 确保断点可以正常命中 在调试器启动后,加载您的应用并导航到断点所在的页面。VS Code 应该会在合适的地方命中断点,您可以在调试器中查看变量、调用堆栈等。 7. 调试技巧 在 .tsx 文件中放置断点,以调试 React 组件。 使用浏览器中的调试控制台与 VS Code 调试控制台同步进行调试。 使用 Ctrl+Shift+D 快捷键快速打开调试面板。 按照这些步骤,您应该能够在 VS Code 中对 Vite + React + TypeScript 项目进行顺利的断点调试。
gharchive/issue
2024-10-25T03:39:40
2025-04-01T04:33:13.184373
{ "authors": [ "WangShuXian6" ], "repo": "WangShuXian6/blog", "url": "https://github.com/WangShuXian6/blog/issues/206", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1917352
initial demo of using jsonp to allow the client.js to get a page from ano initial demo of using jsonp to allow the client.js to get a page from another wiki-server - add alink of the style [[remoteserver:1111|welcome-visitors]] - but you must make sure both servers are running a jsonp enabled wiki note that this makes the /remote uri redundant.. (though i've only tested viewing) please don't merge into master - this is really just to show you what I'm talking about wrt using JSONP for REST and Fork or remote wiki pages. alot of client code still points to the wrong url's - show source for eg - a little refactoring to extract out the build URL code should simplify that tho I like this. I think it would be sweet if a server and the client code it offers could have the freedom to choose either approach: client goes straight to server with the desired page client asks server to retrieve the desired page As you point out, to make this work all servers would need to understand the jsonp protocol. I am a little afraid of jsonp when applied to urls found on the internet. When I've used jsonp I've been in control of all sides of the dialog so I had no issues with eval of untrusted code. Allen Wirfs-Brock tells me that he has been trying to get browser authors to relax the same-origin policy for application/json text as simply parsing this is much safer than eval of random jsonp expressions. That would be an ideal. Well, I've talked to some Mozilla people about my idea, but it isn't something I've really pushed on. Also, Crockford said he liked the idea. We'd probably need to get broader community support before the browser implementers would really jump on it. The basic idea would be to recognize application/json or application/jsonp as distinct mime types in script tags. So to load a json data file you would say something like: <script src="whatever" type="application/json"> </script> and the user agent would recognize "application/json" as a distinct scripting language which it would process using the JavaScript JSON.parse functionality it already has. The resulting objects with no cross site restrictions would then be hung off of the DOM node corresponding to the scripting tag. I have a writeup that I have never broadly circulated that describes this in more detail and address issues like how to make it work with existing jsonp servers that expose files that would normally not parse with JSON.parse. goodness me, lets see. non-cross site restricted access to import data from anywhere, into your local browser's persistent datastore. I would love it, and so would the blackhats - yes, it's much better than the random code injection, it is still random data injection. so Allen - oh, yes please, making a web data -> DOM mechanism would really improve data application developers lives. PS: if we're talking about pushing browsers into a sanity-land where they actually implement what should have been there in the first place, how about fixing authentication? why are browsers not essentially using my ssh keys (though i don't quite grok why ssh keys are separate from gpg keys ...) Note that my proposal is far safer than current jsonp approaches which allow complete arbitrary JavaScript code to be injected and executed. My proposal only loads valid JSON data trees to be loaded. Such trees are complexly inert. No execution would be involved. At worse you get random strings, numbers, and interior tree node objects hanging of specific script nodes. You are going to get that from any form of json data load, so I don't see where your concern lies. If it is that the data becomes accessible by anyone via dom traversal then stash the object reference in a local var and delete the script node as soon as it is loaded. However, you probably aren't really protecting much. If anybody can already access the data using jsonp and a script node with type="text/javascript". yes, exactly, it'll be safer, but still as unsafe as the rest of the web. Way too many of us do DSL / data is code like work, so 'no execution of js' does not mean no execution. Really, it all comes back to the sad fact that web technology has not worked on webs of trust (gpg style), so that I can use type=application/json , and have some assurance (by checking the signature against corruption and that it comes from a source that I have decided to trust) that its data form someone i know. imo we don't really want to end up with a git-like collaboration method - it leads back to single webmaster syndrome, with the added confusion of many duplicate forks with spelling changes. instead, if I sign your public key as 'i trust you', then your changes could be auto-merged... I do wish this was rw federated wiki accessible, that way it'd be easer to refactor. Ward - I'm playing with my static-server implementation, and I think I can make it serve jsonp using mod_include (ie, server side includes). if thats the case, can we change the definition of a federated server to only talk jsonp? from there, we can later add mirror/read non-federated wiki sources separately.. We'd need to work out a way to prevent javascript injection before I'd feel comfortable with building on jsonp. I'm sorry for ignoring this for so long - I need to get some serous foswiki work done before the Foswiki Camp and general assembly at the end of the month - After that, well - I'm somehow expecting that figuring out how to give the girls their first snowy xmas will hit, and then it'll be time to jetlag back to Brisbane :) I really don't feel I can put in enough time to work though my fuzzy ideas - but while they're drawing on new paper together (collaboration between young twins is fascinating) my focus on what is currently the static site has more to do with simplifying - if we can use zero code to serve reading, we leverage the nature of http. To add write, all we need to do is write a simple server code to allow PUT (er, or use DAV?) This then makes me feel that we can focus on the meat of the innovation - the interaction and data dynamics that allow users to innovate. so when we say 'static' i just think fast and lightweight, not read only. btw, I keep staring at the elephant - what happens if you have millions of edit events on a page? Especially as iirc, we only get ~5M in the browse datastore. I was hoping that we would have a federation already, so I could work on our combined, and separate visions, but real life makes quite moments harder :) mmm, times up. We'll have good stuff waiting for you.
gharchive/issue
2011-10-15T12:25:13
2025-04-01T04:33:13.228769
{ "authors": [ "SvenDowideit", "WardCunningham", "allenwb" ], "repo": "WardCunningham/Smallest-Federated-Wiki", "url": "https://github.com/WardCunningham/Smallest-Federated-Wiki/issues/47", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2097185975
Jigsaw JSON UI I made the Jigsaw UI look like the vanilla jigsaw UI and its fully functional :) Thanks Keke! Enjoy the 🍽️⬆️
gharchive/pull-request
2024-01-24T00:12:15
2025-04-01T04:33:13.235239
{ "authors": [ "KekeCreations", "Warhead51707" ], "repo": "Warhead51707/Jigsaw-Block-Bedrock", "url": "https://github.com/Warhead51707/Jigsaw-Block-Bedrock/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1473265606
Set up test server Separate server for testing the infrastructure internally vs the server that collects data. [ ] record whole video instead of components [ ] push recordings to a different server [x] push data to a different server [ ] add "record" true or false to the config The 'testing' branch will push to a separate 'testing' server. This means we can save the production server for true data collection, and not mix up the data collection server with test data.
gharchive/issue
2022-12-02T18:13:59
2025-04-01T04:33:13.258712
{ "authors": [ "JamesPHoughton" ], "repo": "Watts-Lab/deliberation-empirica", "url": "https://github.com/Watts-Lab/deliberation-empirica/issues/357", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1361924414
We are using an old version of issue templates. Please update to the new issue template workflow With the new issue templates, we can provide a better experience to our contributors Expected Behavior Issue templates are helpful when you want to guide opening issues, while allowing contributors to specify the content of their issues. If you want contributors to provide specific, structured information when they open issues, issue forms help ensure that you receive your desired information. Current Behavior We currently have the older approach, there is a new fancy approach! Possible Solution Create a new version based on the PowerShell Team their version: https://github.com/PowerShell/PowerShell/blob/master/.github/ISSUE_TEMPLATE/Bug_Report.yaml Not completed, reopened. Fixed in #41
gharchive/issue
2022-09-05T13:00:47
2025-04-01T04:33:13.278555
{ "authors": [ "Manbearpiet" ], "repo": "WeAreInSpark/AzureDevOpsPowerShellAPI", "url": "https://github.com/WeAreInSpark/AzureDevOpsPowerShellAPI/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2744057265
🛑 Bunkr is down In 2909729, Bunkr (https://bunkr.co.uk/) was down: HTTP code: 403 Response time: 617 ms Resolved: Bunkr is back up in c6f33d5 after 16 minutes.
gharchive/issue
2024-12-17T06:36:44
2025-04-01T04:33:13.280947
{ "authors": [ "WeArePanda" ], "repo": "WeArePanda/Uptime", "url": "https://github.com/WeArePanda/Uptime/issues/90", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1125601650
Error injecting constructor, java.io.FileNotFoundException: conf/global.properties (No such file or directory) 启动不了 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - com.google.inject.ProvisionException: Unable to provision, see the following errors: Error injecting constructor, java.io.FileNotFoundException: conf/global.properties (No such file or directory) at azkaban.execapp.FlowRunnerManager.(FlowRunnerManager.java:175) at azkaban.execapp.FlowRunnerManager.class(FlowRunnerManager.java:108) while locating azkaban.execapp.FlowRunnerManager for the 3rd parameter of azkaban.execapp.AzkabanExecutorServer.(AzkabanExecutorServer.java:126) at azkaban.execapp.AzkabanExecutorServer.class(AzkabanExecutorServer.java:97) while locating azkaban.execapp.AzkabanExecutorServer 1 error 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at com.google.inject.internal.InjectorImpl$2.get(InjectorImpl.java:1028) 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at com.google.inject.internal.InjectorImpl.getInstance(InjectorImpl.java:1054) 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at azkaban.execapp.AzkabanExecutorServer.main(AzkabanExecutorServer.java:178) 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - Caused by: java.io.FileNotFoundException: conf/global.properties (No such file or directory) 2022-02-07 15:10:07.606 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at java.io.FileInputStream.open0(Native Method) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at java.io.FileInputStream.open(FileInputStream.java:195) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at java.io.FileInputStream.(FileInputStream.java:138) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at azkaban.utils.Props.(Props.java:84) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at azkaban.utils.Props.(Props.java:74) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at azkaban.execapp.FlowRunnerManager.(FlowRunnerManager.java:212) 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at azkaban.execapp.FlowRunnerManager$$FastClassByGuice$$f5329b23.newInstance() 2022-02-07 15:10:07.607 ERROR [main] azkaban.utils.StdOutErrRedirect 65 write - at com.google.inject.internal.DefaultConstructionProxyFactory$FastClassProxy.newInstance(DefaultConstructionProxyFactory.java:89) cp conf bin/ 下即可。 这是一个BUG吧。 修改internal/internal-start-executor.sh第20行为azkaban_dir=$(cd $(dirname $0)/../..;pwd)可以正常启动了 这里就是读取exec路径,应该是它这个路径读取的有问题,所以找不到conf目录 提供最新版DataSphereStudio全家桶安装服务。 包括(DataSphereStudio, Linkis, Exchangeis(Datax, Sqoop), Scriptis, Streamis(Flink), Visualis, DolphinScheduler) 另外包含 元数据,数据安全,数据质量,数据血缘。(这些都是DataSphereStudio不自带的服务) 包含hadoop集群搭建(hive, spark, flink, kafak, kylin) 提供上述所有系统二次定制开发服务 提供系统搭建运维疑难问题解决 添加备注,DSS大数据 持续为DataSphereStudio开源贡献力量
gharchive/issue
2022-02-07T07:14:53
2025-04-01T04:33:13.291769
{ "authors": [ "YinDonglai", "aileennian", "tuigerphkeeper" ], "repo": "WeBankFinTech/Schedulis", "url": "https://github.com/WeBankFinTech/Schedulis/issues/80", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1479017286
Al and ML Description if any developer is expert in ai ml can you just add a roadmap of ai ml roadmap and also good resources. Screenshots No response Additional information No response can't we use this playlist of Kunal on AI & ML as a starting point. we all need this repository kind of roadmap and resources in ai and ml. industry's best creators of youtube resources like web development already have. Hello guys, any one interested to collaborate w me to create the ML roadmap? @Vishvsalvi, I would love to work with you. Please assign this to me. I would also like to collaborate. That's great guys, we need to inform maintainers first coz they are busy and barely pay any attention in this roadmap can i contribute a roadmap for ML I also want to contribute on this isssue Hey @1nevil can you assign this issue to me would love to contribute how I will assign it @KaKashi1210 @sonustar @prakashvishal93 ? sorry by mistake closed issue sorry for that. Is this issue still open? I would like to contribute my part.
gharchive/issue
2022-12-06T12:47:06
2025-04-01T04:33:13.301746
{ "authors": [ "1nevil", "AniketNS", "KaKashi1210", "Vishvsalvi", "abhigyan631", "ankkol2011", "prakashvishal93", "sonustar", "suryanshsoni120" ], "repo": "WeMakeDevs/roadmaps", "url": "https://github.com/WeMakeDevs/roadmaps/issues/314", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1676976065
Updating Geek Around Community Discord Name Changes proposed Updating Geek Around Community Discord Name to "Geek Around Community X Cloud Native Nashik " Screenshots Also, this could be changed in other roadmaps too. You can go ahead with updating those if you want :) Hello @kaiwalyakoparkar, I have updated the link, please review this PR and check if everything is ok. And I'll be changing this in other roadmaps too so please keep an eye on those too. Thanks😊.
gharchive/pull-request
2023-04-20T15:39:14
2025-04-01T04:33:13.303895
{ "authors": [ "Sandeep-RN", "kaiwalyakoparkar" ], "repo": "WeMakeDevs/roadmaps", "url": "https://github.com/WeMakeDevs/roadmaps/pull/515", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
266498187
my name Adding myself @salman-bhai many files have been changed. Should we close this? Yes I will send a new one.
gharchive/pull-request
2017-10-18T13:43:06
2025-04-01T04:33:13.384163
{ "authors": [ "MJ10", "YonatanHanan" ], "repo": "WebClub-NITK/Hacktoberfest-2k17", "url": "https://github.com/WebClub-NITK/Hacktoberfest-2k17/pull/505", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
320273306
如何自定义回复消息 bi'ru比如 暂时关闭,如有问题进一步回复即可
gharchive/issue
2018-05-04T13:22:56
2025-04-01T04:33:13.888075
{ "authors": [ "binarywang", "lzwzzy" ], "repo": "Wechat-Group/weixin-java-tools", "url": "https://github.com/Wechat-Group/weixin-java-tools/issues/574", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
228829836
[Multiple Suggestions] Tiered accounts / limits / expiration This is just a suggestion and it's specific to my wants, so it's probably not something that would be seen as a necessity. I have limited space on my vps with multiple websites, backups, etc. I want to open it to the public, but I don't want them to have 100mb upload limit, I want it at 10mb for non registered users, but 100mb for registered users. For people with accounts, I don't want to delete their files, but I don't want public uploads to stay forever. I want it to expire after x amount of days ("pubExpire: 30" // public files expire after 30 days variable). (I think this is a request already?) People with accounts I want them to be able to keep their uploads permanent but have a limit on how much they can upload. So registered users can have permanent uploads for say 5gb storage ("quota: 5GB"). Basically my suggestions are: Public non registered max upload size Registered max upload size Account storage quotas Also while I'm here, does anyone have a script I can have to add accounts to database so I don't have to change allow registrations in config and restart the service? Or any other scripts (deleting after x amount of days). Thanks :) Thanks for the suggestion! I'm making a public kanboard soon to organize the new features lolisafe is gonna have on it's next version, a few weeks from now. Maybe I can squeeze in the storage quota as well. this goes completely against the pomf design idea, and this thought deserves to die Good thing we are not a pomf clone. Sorry :( just something I would have liked to see from having multiple users on a small vps. Was not aware of specific guidelines. Repo owner can close this if they'd like. No @notstevek , I believe those would be nice suggestions. Will keep them in mind, I'm not following those old pomf guidelines. @Pitu I am happy with how this is now and none of the other pomf like upload projects have met my needs as much as this has. Like I said in the op, this was just a suggestion based on my wants. If the community disagrees with it, I would not want this to cause any issues. Anyway, thanks for the response and for your work. Pinging @BobbyWibowo for this one 🙈 Pinging @BobbyWibowo for this one 🙈
gharchive/issue
2017-05-15T20:09:23
2025-04-01T04:33:13.893542
{ "authors": [ "Pitu", "alucard0134", "ghost", "notstevek" ], "repo": "WeebDev/chibisafe", "url": "https://github.com/WeebDev/chibisafe/issues/39", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1346892159
API Link returns 404 The API Link returns a 404 error. Hi, I am aware I don't think it's something in the code that's faulty. It's probably the API link itself. Yes, they have changed the API endpoint for TTS How would this be fixed? Or is it unfixable? Following Is this fixable? I don't know. Maybe if like a new link can be found or something? For the api cuz the api link returns a 404, meaning it doesn't exist anymore, maybe it got moved. i believe if you sign up to become a tiktok devolper they give you the api info @ItsFreshify I looked around and it's a very limited subset of their API that they give you access to. It does not include the TTS engine. so does this mean that tiktok tts rip? On Mon, Aug 29, 2022 at 6:03 PM Weilbyte @.***> wrote: @ItsFreshify https://github.com/ItsFreshify I looked around and it's a very limited subset of their API that they give you access to. It does not include the TTS engine. — Reply to this email directly, view it on GitHub https://github.com/Weilbyte/tiktok-tts/issues/10#issuecomment-1230909905, or unsubscribe https://github.com/notifications/unsubscribe-auth/AZRSAYXVDVCN6J2MED7JN3TV3UXSHANCNFSM57I3PHCQ . You are receiving this because you authored the thread.Message ID: @.***> If I recall correctly, oscie57's python CLI tool. Generally yes but I have not been successful in intercepting the HTTPS traffic from the app, because it crashes on every emulator I tried it on, now, the TikTok Lite app works but It doesn't seem to have TTS support so it's not making those requests anyway. And I don't have a rooted Android phone at hand. Even if I did I'm pretty sure if this was a deliberate action towards tools like this, that they would have also added authentication requirements to the endpoint. I have a device with root, and properly hiding it from tiktok, but can't intercept the requests too, it does work triggerint the tts when editing a video tho @xfrcc which android version? iirc if above 6.0, you need to patch the APK to accept the MITM proxy's CA @xfrcc which android version? iirc if above 6.0, you need to patch the APK to accept the MITM proxy's CA 7.1.2, i've tried patching with apk-mitm, but not working because apktool it's parsing bad the xml's Got news, new endpoint is api22-normal-c-useast1a.tiktokv.com, however, you must replicate now the headers like if it was in the app, i will test which one is necessary and which one us redundant Thank you! I will attempt to reproduce and see how this can be implemented within this tool. does this mean that the tts could work again? couldn't you use the session id from the web on pc? I get the Couldn't load speach, maybe it requires it to be a sessionid generated by logging in on phone? Not too sure, https://i.imgur.com/oPSNJiI.png A solution is in the works and expected to be out by Friday, if everything goes smoothly. @AdamAtomus the aid parameter is wrong, it needs to be 1233 A solution is in the works and expected to be out by Friday, if everything goes smoothly. @AdamAtomus the aid parameter is wrong, it needs to be 1233 yeah same issue, what SID did you use? from the app or from website on pc Thank you! @xfrcc and @Weilbyte ! Cheers!
gharchive/issue
2022-08-22T19:39:27
2025-04-01T04:33:13.905849
{ "authors": [ "AdamAtomus", "ItsFreshify", "Petter1515", "Weilbyte", "alexgarciab", "elidigital", "rschoolu", "xfrcc" ], "repo": "Weilbyte/tiktok-tts", "url": "https://github.com/Weilbyte/tiktok-tts/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
134443508
Exception thrown after page reload. In admin panel, from time to time, reload using reload button in web browser throws following exception: LogicException in RequestHelper.php line 92: Cannot set session attributes without valid session. @rmartonik You have to provide more information: With OS, webserver + versions are you running on? What do you use as the session handler? http://symfony.com/doc/2.8/cookbook/session/sessions_directory.html Also check the logs in app/logs or var/logs for additional information about the error. Or you webserver/php logs. @hacfi It happens since we've upgraded Symfony dependencies to 3.x. Every time the cache is cleared or invalidated, without warming it up (cache:warmup), first request will always throw an exception. Consecutive requests are fine so that must be something with classes.php or dumped container imho. Will do an investigation ;) Cloning the repo now..will have a quick look! FYI: - Installing wellcommerce/composer-plugin (dev-master 9a72afd) Cloning 9a72afddecd9b86136afffc1060e2e8eb0c94e24 The "wellcommerce/composer-plugin" plugin requires composer-plugin-api 1.0.0, this *WILL* break in the future and it should be fixed ASAP (require ^1.0 for example). Not sure why specifying a specific version seems to be an issue but the ^1.0 version constraint doesn’t sound bad to me. @hacfi Thanks! Issue created: https://github.com/WellCommerce/WellCommerce/issues/41
gharchive/issue
2016-02-18T00:28:01
2025-04-01T04:33:13.910043
{ "authors": [ "adampiotrowski", "hacfi", "rmartonik" ], "repo": "WellCommerce/WellCommerce", "url": "https://github.com/WellCommerce/WellCommerce/issues/39", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
141328288
force reload data how can i force reload of data and dots in a calendar view? oc [calendar reloadData]; swift calendar.reloadData() Does not work [_calendar reloadData]; This is my code: NSURLSession *session = [NSURLSession sharedSession]; NSURLSessionDataTask *dataTask = [session dataTaskWithRequest:urlRequest completionHandler:^(NSData *data, NSURLResponse *response, NSError *error) { NSHTTPURLResponse *httpResponse = (NSHTTPURLResponse *)response; if(httpResponse.statusCode == 200) { NSError *parseError = nil; _response = [NSJSONSerialization JSONObjectWithData:data options:0 error:&parseError]; NSMutableArray *eventos = [[[_response objectForKey:@"RESPONSE"] objectAtIndex:0] objectForKey:@"DATA"]; int result = [[[[_response objectForKey:@"RESPONSE"] objectAtIndex:0] objectForKey:@"RESULT"] integerValue]; if(result == 1){ NSDate *now = [NSDate date]; NSDictionary *evento; int cont = 0; for(evento in eventos) { [_datesWithEvent insertObject:[[evento objectForKey:@"EVENTO"] objectForKey:@"FECHA"] atIndex:cont]; cont = cont + 1; } [_calendar reloadData]; } } else { NSLog(@"Error"); } }]; [dataTask resume]; Are you sure that reloadData is being called and _calendar is not deallocated yet? It looks like a really bad code, most likely the issue is not in the library.
gharchive/issue
2016-03-16T16:33:18
2025-04-01T04:33:13.913082
{ "authors": [ "WenchaoD", "awdesarrollo", "lucabecchetti", "numen31337" ], "repo": "WenchaoD/FSCalendar", "url": "https://github.com/WenchaoD/FSCalendar/issues/230", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2468206689
Spotify Plugin This PR integrates Spotify to the play your favorite songs on your devices. You can select or mix playlists by placing figures on one of the large areas. Furthermore, by placing another figure in the middle, you can specify a device on which the music should be played. I would like to take this #21
gharchive/pull-request
2024-08-15T14:45:12
2025-04-01T04:33:13.916930
{ "authors": [ "yannickschrade" ], "repo": "Wendelpunkt/ToypadChallenge", "url": "https://github.com/Wendelpunkt/ToypadChallenge/pull/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2035041220
Possible to implement CSDI as a forecasting model for POTS data 1. Feature description Although, in the CSDI paper, the authors discussed using the model for time series forecasting, their experiments and scenarios are for complete time series. However, the model is possible for POTS forecasting. 2. Motivation Make CSDI a forecasting model in PyPOTS. 3. Your contribution I'm currently overloaded. If anyone is interested in this task, please discuss it below to see if you can take it. The CSDI model has been included in PyPOTS as an imputation model, so necessary modules are all available already. I start working on it, to get familiar with code and understand its infrastructure. Many thanks, Ahmad. I've assigned this to you and me. I can help you at any time, please just start the discussion here if you need any information or assistance. And once you make the PR, I'll review the code. Is this the Algorithm that should be implemented, could you describe more, As I understand now it will be in forecasting algorithms in PyPOTS. @ahmaddroobi99 The graph you posted is an overview of how CSDI works: input, process, and output. Here are some clues and suggestions from me: read the paper and figure out the model itself first. I believe you've done well with this; walk through the code under this path https://github.com/WenjieDu/PyPOTS/blob/main/pypots/imputation/csdi and figure out how CSDI is implemented as an imputation model here. The CSDI code in PyPOTS keeps the same logic as the official implementation https://github.com/ermongroup/CSDI but is in a more unified PyPOTS way; read the code in https://github.com/ermongroup/CSDI/blob/main/exe_forecasting.py and the code it imports. This script will help you learn how CSDI works on the forecasting task; get your hands on turning the imputation CSDI in PyPOTS into a forecasting model. Basically, you need to do some refactoring work to adapt the code for the forecasting task; Let me know if you have any further questions. Hi @WenjieDu and @ahmaddroobi99 maybe I can help on this as well, I have some experience in CSDI code and had also familiar with their code and paper. Thanks, @LINGLONGQIAN, good to know you'd like to help. I'm going to check with Ahmad to see how things are going on his side. Then we can determine which part you can take care of. Ahmad @ahmaddroobi99 recently has been very busy with his own work. After communication, Linglong @LINGLONGQIAN and I will take care of this issue.
gharchive/issue
2023-12-11T07:53:39
2025-04-01T04:33:13.923178
{ "authors": [ "LINGLONGQIAN", "WenjieDu", "ahmaddroobi99" ], "repo": "WenjieDu/PyPOTS", "url": "https://github.com/WenjieDu/PyPOTS/issues/264", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1521505864
The provided document for implementation of the proxy is so vague the provided manual is so vague and after doing all steps, we could not get connected to our servers. I think the manual should be more accurate what is status of your server ? mostly skip to clone this repo, Dockerfile and docker-compose file is in this repo. first step Clone the repository to your local machine I have done all the steps correctly. Finally I got this message: Starting whatsapp_proxy ... done Showing that the proxy is working properly. but when I set the IP in proxy filed of the whatsapp I get " can not connect" error but when I set the IP in proxy filed of the whatsapp I get " can not connect" error This unfortunately will be completely dependent on your connectivity and hosting parameters. Can you probe the stats port from your device? That should tell you if you can contact your proxy or not. Do you have right firewall ? if you are in AWS open below ports : Or adjust your provider policy to open below ports . you can reduce base on your need 80 443 5222 8080 8443 8222 8199 I am on Hetzner and firewall is inactive Is there any need to edit yml files? or set some parameters more than what's provided in the manual? Hetzner has a additional firewall for incoming traffic, you can open the ports on the robot config website.
gharchive/issue
2023-01-05T22:01:26
2025-04-01T04:33:13.947959
{ "authors": [ "Melbar666", "msalehy", "slawlor", "xxamr" ], "repo": "WhatsApp/proxy", "url": "https://github.com/WhatsApp/proxy/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
731597423
Add TrustedSite Crawler to bots list See github issue: #540 TrustedSite TrustedSite Crawler Link: https://www.trustedsite.com/a/crawler Coverage increased (+0.02%) to 99.97% when pulling 8d81ea805540420e708c79a17893259bab8b9210 on ayumi-cloud:TrustedSite into 880b9fa797401d14b28956442944c3daa70240ff on WhichBrowser:master. Thanks! Applied using a clean commit: e15880b3624ebdc5403738891ac01227cd03d462 Thanks! Applied using a clean commit: e15880b3624ebdc5403738891ac01227cd03d462
gharchive/pull-request
2020-10-28T16:25:10
2025-04-01T04:33:13.962110
{ "authors": [ "NielsLeenheer", "ayumi-cloud", "coveralls" ], "repo": "WhichBrowser/Parser-PHP", "url": "https://github.com/WhichBrowser/Parser-PHP/pull/614", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
383205582
perfectly renders single recipe view single recipe view now uses the real data when clicked on single recipe from the all recipes view please avoid the error of travis if it's related to timeout issue regarding neo4js !
gharchive/pull-request
2018-11-21T16:44:02
2025-04-01T04:33:13.965304
{ "authors": [ "ddilano" ], "repo": "Whisk1809/whisk", "url": "https://github.com/Whisk1809/whisk/pull/53", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1798077566
[BUG] when fetch all groups When i try to fetch all groups on whatsapp using this function sock.groupFetchAllParticipating() it not only returns the value to the variable i created but also returns the value to the groups.update event what #160 this issue? #160 check my pull request Thanks!! You saved me!!
gharchive/issue
2023-07-11T04:35:49
2025-04-01T04:33:13.966884
{ "authors": [ "Hendra827", "JunaGanz", "SheIITear" ], "repo": "WhiskeySockets/Baileys", "url": "https://github.com/WhiskeySockets/Baileys/issues/238", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
625645650
nos expected result with gs1_128 When create barcode from string : ` def generate_barcode(self): code = u"(01)01234567890128(15)057072" GS1 = barcode.get_barcode_class('gs1_128') fp = io.BytesIO() gs1 = GS1(code, writer=ImageWriter()).write(fp) a = base64.b64encode(fp.getbuffer()) ` The expected result its: The result obtanides its: I tried a couple of barcode readers (including this one online), and the former barcode doesn't match the input string, but the generated one reads properly. Can you clarify a bit more what's wrong? @WhyNotHugo According to GS1 specs for GS-128, the parentheses enclosing the Application Identifiers (AIs) should be visible in the Human Readable Interpretation (HRI) (the string below the barcode) but should NOT be encoded in the barcode itself. Oh, I'd no idea. Do you have a link to this spec? I guess the right fix then would be for this lib to add the parenthesis to the generated label, but not to the barcode itself. [https://www.gs1.org/docs/barcodes/GS1_General_Specifications.pdf](GS1 General Specifications) For GS1-128 see Pages 272-286 Regarding the parentheses in HRI see page 229: chapter 4.15, rule 3 Keep the specs at hand. It's a great source of information and the fact that it is freely available makes it invaluable. I'll be around if you need anything else and I can help! Finally y use external string with the barcode to generate the final image. The problem was the parentesis . Thx for aswer. Please reopen. This does not work as it should. I believe #100 (which has just been merged) should have fixed this. Can you confirm this works fine? Unfortunately it doesn't. I skimmed through the code and the PR has to do with EAN13 not GS128 (aka code128) and adding guard bars. Parentheses are still encoded in the barcode. Am I missing something? My bad, mixed up barcodes 😅 Any news on this one? Hi, if you do : gs1 = barcode.get_barcode_class('gs1_128') barcode_img = gs1('90502087191121904229589045190400011980', writer=ImageWriter()) if you compare the image generated with the right image below, you can see they are diferent. You read codebar generated, but in some cases it doesn't work.
gharchive/issue
2020-05-27T12:18:20
2025-04-01T04:33:14.032423
{ "authors": [ "WhyNotHugo", "ctc87", "diggy128", "jmarxuach" ], "repo": "WhyNotHugo/python-barcode", "url": "https://github.com/WhyNotHugo/python-barcode/issues/75", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2102450274
🛑 submissions is down In bce6d56, submissions (peamail.net) was down: HTTP code: 0 Response time: 0 ms Resolved: submissions is back up in db524fc after 15 minutes.
gharchive/issue
2024-01-26T15:55:27
2025-04-01T04:33:14.035184
{ "authors": [ "Wicloz" ], "repo": "Wicloz/upptime", "url": "https://github.com/Wicloz/upptime/issues/156", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }