added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:10:07.889317
2021-07-28T15:48:21
954999308
{ "authors": [ "daverodgman", "mpg", "yuhaoth" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13269", "repo": "ARMmbed/mbedtls", "url": "https://github.com/ARMmbed/mbedtls/issues/4820" }
gharchive/issue
TLS 1.3: simultaneous support for TLS 1.2 and 1.3 Enable support for building the library with support for both 1.2 and 1.3 available. prototype issue: https://github.com/hannestschofenig/mbedtls/issues/297 From the above comment linking to PR in the prorotype's repo, and according to @yuhaoth's memories, this was about doing the work in the prototype, so it doesn't seem relevant for upstream, much less for the upstream MVP. Hence, moving out of the MVP epic.
2025-04-01T04:10:07.893655
2022-02-25T10:19:39
1150278245
{ "authors": [ "gilles-peskine-arm", "stevew817" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13270", "repo": "ARMmbed/mbedtls", "url": "https://github.com/ARMmbed/mbedtls/pull/5579" }
gharchive/pull-request
Erase secrets in allocated memory before freeing said memory Description The PSA Crypto storage module uses dynamically-allocated buffers to buffer sensitive key material during the process of loading said key material from NVM, or storing it to NVM. To avoid leaking the key material through the heap, zero out the buffer before freeing it. Status READY Requires Backporting Unsure, maintainer can make a decision Migrations NO Additional comments Todos [x] Tests [x] Documentation [x] Changelog updated [ ] Backported Steps to test or reproduce @daverodgman does this need a changelog entry? Yes, please add a changelog entry. grep '\* Zeroize' ChangeLog If this needs backporting, can anybody tell me which branches this would be expected on? There is currently only one long-time support branch: 2.28. So please backport to mbedtls-2.28. Backported.
2025-04-01T04:10:07.897014
2017-01-10T13:39:00
199823498
{ "authors": [ "NWilson", "andresag01", "daverodgman" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13271", "repo": "ARMmbed/mbedtls", "url": "https://github.com/ARMmbed/mbedtls/pull/745" }
gharchive/pull-request
DRAFT: Threefish implementation This PR is an early prototype implementation of the Threefish tweakable block cipher. The implementation is based on the specification at https://www.schneier.com/academic/paperfiles/skein1.3.pdf. This PR includes: Block cipher implementation and header file (threefish.c and threefish.h) Changes to the make and CMake build files A test suite test_suite_threefish Documentation Outstanding work: Additional tests (both: selt test and the test suite) Integration with mbed TLS cipher abstraction: Need to accommodate the mbedtls_threefish_settweak() function with the existing API Hey, looks cool! Nicer than AES (in my opinion). Do you expect Threefish to be used over ChaCha20 - or is it losing the popularity contest as the successor to AES? Closing as we don't appear to have demand for this algorithm.
2025-04-01T04:10:07.900813
2018-06-05T13:05:47
329447557
{ "authors": [ "adbridge", "artokin", "ciarmcom" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13272", "repo": "ARMmbed/nanostack-border-router", "url": "https://github.com/ARMmbed/nanostack-border-router/issues/114" }
gharchive/issue
Thread commission fails when Border Router is built with using IAR or ARMC6 Thread commissioning fails when Border Router is build with IAR or ARMC6 compiler. Commissioning works fine when builds are made by using ARM or GCC_ARM. Commissioning device (mobile phone) does not enter to screen where QR-code can be scanned. Therefore I'm not able to add new devices to the network. Tests were made against mbed-os-5.9.0-oob branch. @artokin ARMC6 is not officially supported yet. Does it work with ARMC5 ? ARM Internal Ref: IOTTHD-2595 This is working fine witn ARMC5 and GCC__ARM. It looks like the problem exists only with IAR/ARMC6. I was using IAR ANSI C/C++ Compiler V<IP_ADDRESS>/LNX for ARM (from CI) and IAR Workbench 8.11.2 PR https://github.com/ARMmbed/nanostack-border-router/pull/116 will adjust mbedtls MBEDTLS_SSL_MAX_CONTENT_LEN size to 4096 (from 16k). Then IAR BR commissioning works as memory allocation succeeds. PR merged to mbed-os-5.9.0-oob branch. Issue with ARMC6 commissioning still remains but as ARMC6 is not officially supported no further actions will be made. Closing this issue.
2025-04-01T04:10:07.903761
2017-08-21T15:11:46
251688718
{ "authors": [ "deepikabhavnani", "soramame21" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13273", "repo": "ARMmbed/sd-driver", "url": "https://github.com/ARMmbed/sd-driver/pull/56" }
gharchive/pull-request
Add SD card reader support over DISCO_L476VG platform for running Cloud Client. @MarceloSalazar Add SPI pins setting for support SD card reader/writer over DISCO_L476VG, which is required and tested by using ARMmbed/mbed-cloud-client-example-sources-internal. @soramame21 - Please verify the SD card specific test, you can have more details of that in "Build the mbed OS Test Cases" section of Readme. Required tests to be successful: sd-driver-tests-block_device-basic sd-driver-tests-filesystem-basic sd-driver-tests-filesystem-fopen @deepakvenugopal, @MarceloSalazar Thanks for your comments. I had run and passed the test. please check the attached test result. sd-driver-test-result-Ren0822.txt
2025-04-01T04:10:07.919985
2023-02-22T21:21:31
1595849147
{ "authors": [ "asjohnston-asf", "jtherrmann" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13274", "repo": "ASFHyP3/gis-services", "url": "https://github.com/ASFHyP3/gis-services/pull/8" }
gharchive/pull-request
Refactoring for make-hand-service.py Makes the following changes: Logging with print statements. Use tempfiles where appropriate. Add cli options. I'm currently running the script with: python make_hand_service.py /home/arcgis/jtherrmann/ --rasters-filter='.*Copernicus_DSM_COG_10_N08_00_W01*' But it fails; full log output: CreateFileGDB CreateMosaicDataset AddFields AddRastersToMosaicDataset CalculateFields BuildFootprints Traceback (most recent call last): File "make_hand_service.py", line 90, in <module> update_boundary='UPDATE_BOUNDARY', File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/management.py", line 14697, in BuildFootprints raise e File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/management.py", line 14694, in BuildFootprints retval = convertArcObjectToPythonObject(gp.BuildFootprints_management(*gp_fixargs((in_mosaic_dataset, where_clause, reset_footprint, min_data_value, max_data_value, approx_num_vertices, shrink_distance, maintain_edges, skip_derived_images, update_boundary, request_size, min_region_size, simplification_method, edge_tolerance, max_sliver_size, min_thinness_ratio), True))) File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/geoprocessing/_base.py", line 512, in <lambda> return lambda *args: val(*gp_fixargs(args, True)) arcgisscripting.ExecuteError: ERROR 999999: Something unexpected caused the tool to fail. Contact Esri Technical Support (http://esriurl.com/support) to Report a Bug, and refer to the error help for potential solutions or workarounds. No mosaic dataset item selected. This operation cannot be performed. Failed to execute (BuildFootprints). Start Time: Wednesday, February 22, 2023 9:16:36 PM ERROR 999999: Something unexpected caused the tool to fail. Contact Esri Technical Support (http://esriurl.com/support) to Report a Bug, and refer to the error help for potential solutions or workarounds. No mosaic dataset item selected. This operation cannot be performed. Failed to execute (BuildFootprints). Failed at Wednesday, February 22, 2023 9:16:38 PM (Elapsed Time: 2.59 seconds) I wonder if perhaps the regex filter is not matching any rasters? This will be my next line of investigation. Drop the .s from the filter, I suspect. Per https://pro.arcgis.com/en/pro-app/latest/tool-reference/data-management/add-rasters-to-mosaic-dataset.htm by default you're doing simple wildcard matching, e.g. '*Copernicus_DSM_COG_10_N08_00_W01*'. It's only if you put the word REGEX` in the filter that you get full perl regular expressions. Drop the .s from the filter, I suspect. Per https://pro.arcgis.com/en/pro-app/latest/tool-reference/data-management/add-rasters-to-mosaic-dataset.htm by default you're doing simple wildcard matching, e.g. '*Copernicus_DSM_COG_10_N08_00_W01*. It's only if you put the word REGEX in the filter that you get full perl regular expressions. Ah, I see, I added REGEX because I thought you just left that part out of the filter you gave me. So I thought you had just forgotten the .. I ran it as python make_hand_service.py --dataset-name GLO30_HAND_ASJ --rasters-filter '*Copernicus_DSM_COG_10_N08_00_W01*' /home/arcgis/asjohnston/gis-services/image_services/glo_30_hand/ and everything worked how I expected. The .sd file published with no issues.
2025-04-01T04:10:07.927300
2019-08-28T19:00:53
486554890
{ "authors": [ "chrsengel", "heldyboy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13275", "repo": "ASRG/asrg.io", "url": "https://github.com/ASRG/asrg.io/issues/24" }
gharchive/issue
Calendar Function Need to find a calendar program. Could be the same as what the auto-ISAC uses. Then we should collect all the conferences from the calendar sources and put them into a single database. This database should be used to show the conference calendar. Additional Ideas Webcrawlers to find new sources Chris, can you see if you can start this? The best way from my point of view is to take an existing calendar and copy / paste.... then we can add our own content. Sure. I'll look for possible solutions. Hey, can we discuss this one as well? We could embed a public google calendar into the site. You only need to create one and make it public. Here's a guide. Otherwise I don't know of any reliable solution for static pages. If we transition to a CMS like wordpress or build up our own infrastructure there will be more options.
2025-04-01T04:10:07.937543
2024-08-19T19:44:15
2474074462
{ "authors": [ "ASzc", "adambkaplan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13276", "repo": "ASzc/fbc-utils", "url": "https://github.com/ASzc/fbc-utils/issues/4" }
gharchive/issue
Allow OCI Manifests as Valid MIME Type OCI Manifest lists are the equivalent/replacement of Docker manifests. Konflux produces operator bundles with this MIME type. MIME: application/vnd.oci.image.manifest.v1+json Done, thanks!
2025-04-01T04:10:07.940873
2023-02-05T11:55:42
1571400089
{ "authors": [ "SuelaCark", "asynchroza" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13277", "repo": "AUBGTheHUB/spa-website-2022", "url": "https://github.com/AUBGTheHUB/spa-website-2022/issues/212" }
gharchive/issue
Change "Welcome to The Hub" to something else in landing section. Reference design in Adobe Forgot to say that this issue is covered within another ticket #341
2025-04-01T04:10:07.974863
2020-10-13T01:56:22
719790363
{ "authors": [ "codecov-io", "siangernlow" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13278", "repo": "AY2021S1-CS2103T-T13-1/tp", "url": "https://github.com/AY2021S1-CS2103T-T13-1/tp/pull/147" }
gharchive/pull-request
Alert user when SHN has been violated When a visit is being added, a warning should be shown if the person has made an illegal visit. A visit is illegal if: The visiting person is infected or in quarantine or both. The location that the person visited is not the address stated in the person. Subsequently, the visit will still be added, but the relevant warning will be generated as well. Close #143 Codecov Report Merging #147 into master will increase coverage by 0.04%. The diff coverage is 85.71%. @@ Coverage Diff @@ ## master #147 +/- ## ============================================ + Coverage 73.98% 74.03% +0.04% - Complexity 789 794 +5 ============================================ Files 113 113 Lines 2464 2476 +12 Branches 304 308 +4 ============================================ + Hits 1823 1833 +10 - Misses 551 552 +1 - Partials 90 91 +1 Impacted Files Coverage Δ Complexity Δ .../address/logic/commands/visit/AddVisitCommand.java 88.57% <85.71%> (-2.74%) 13.00 <6.00> (+5.00) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update aab6c10...e8471c0. Read the comment docs.
2025-04-01T04:10:07.981357
2021-11-08T00:48:49
1046902396
{ "authors": [ "codecov-commenter", "ruoyann" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13279", "repo": "AY2122S1-CS2103T-T09-1/tp", "url": "https://github.com/AY2122S1-CS2103T-T09-1/tp/pull/252" }
gharchive/pull-request
Reordering of clear before help feature Thought it makes more sense to list out all the commands before help and exit. Changed the summary to match the flow as well. Codecov Report Merging #252 (511d9c8) into master (7f8dd2c) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #252 +/- ## ========================================= Coverage 68.19% 68.19% Complexity 745 745 ========================================= Files 94 94 Lines 2581 2581 Branches 346 346 ========================================= Hits 1760 1760 Misses 763 763 Partials 58 58 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 7f8dd2c...511d9c8. Read the comment docs.
2025-04-01T04:10:07.985493
2021-10-30T05:34:57
1040057848
{ "authors": [ "angnobel", "nus-pe-script" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13280", "repo": "AY2122S1-CS2103T-W13-1/tp", "url": "https://github.com/AY2122S1-CS2103T-W13-1/tp/issues/199" }
gharchive/issue
[PE-D] Consistent sequencing of features Minor point again but would be good if the sequence of features are consistent. Eg since it starts with the job position features in the table of contents, start with job position features in command summary. Labels: severity.Low type.DocumentationBug original: yucheng11122017/ped#17 Split into 3 tables
2025-04-01T04:10:07.988389
2021-10-30T05:17:56
1040054805
{ "authors": [ "huizhuansam", "nus-pe-script" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13281", "repo": "AY2122S1-CS2103T-W13-2/tp", "url": "https://github.com/AY2122S1-CS2103T-W13-2/tp/issues/155" }
gharchive/issue
[PE-D] Find Command "--phone" flag issues Input : find --phone 99272758 Output: Invalid command format! find: Finds all persons whose fields contain any of the specified keywords (case-insensitive) and displays them as a list with index numbers. Parameters: FLAG KEYWORDS [MORE_FLAGS] Example: find -n alex -a serangoon User Guide Suggests that this is an acceptable command. Labels: type.FunctionalityBug severity.Medium original: Jai2501/ped#3 Related to #150 Update: this issue is actually related to the lack of support of long-form prefixes by the Find command. @Zhou-Jiahao-1998 please look into this.
2025-04-01T04:10:08.023350
2022-10-26T11:28:43
1423867914
{ "authors": [ "codecov-commenter", "teoyuqi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13282", "repo": "AY2223S1-CS2103T-T10-1/tp", "url": "https://github.com/AY2223S1-CS2103T-T10-1/tp/pull/107" }
gharchive/pull-request
Update DG Reduces steps in add-task Add rationale for design decisions of Task Update section on Logic, Ui, Model and Storage to be in line with our current implementation: Change links from SEEDU repo to our repo Update diagrams Remove unused diagrams Codecov Report Base: 77.84% // Head: 77.55% // Decreases project coverage by -0.29% :warning: Coverage data is based on head (5a62927) compared to base (e41d3c0). Patch has no changes to coverable lines. Additional details and impacted files @@ Coverage Diff @@ ## master #107 +/- ## ============================================ - Coverage 77.84% 77.55% -0.30% - Complexity 741 745 +4 ============================================ Files 111 111 Lines 2221 2245 +24 Branches 264 265 +1 ============================================ + Hits 1729 1741 +12 - Misses 398 408 +10 - Partials 94 96 +2 Impacted Files Coverage Δ ...eedu/address/logic/commands/DeleteLinkCommand.java 85.71% <0.00%> (-5.72%) :arrow_down: ...c/main/java/seedu/address/model/module/Module.java 84.84% <0.00%> (-5.32%) :arrow_down: ...a/seedu/address/model/module/UniqueModuleList.java 80.32% <0.00%> (-4.29%) :arrow_down: src/main/java/seedu/address/ui/TaskCard.java 0.00% <0.00%> (ø) src/main/java/seedu/address/ui/ModuleCard.java 0.00% <0.00%> (ø) src/main/java/seedu/address/ui/TaskListPanel.java 0.00% <0.00%> (ø) ...c/main/java/seedu/address/model/person/Person.java 100.00% <0.00%> (ø) src/main/java/seedu/address/model/AddressBook.java 93.33% <0.00%> (+0.15%) :arrow_up: ...a/seedu/address/model/person/UniquePersonList.java 90.19% <0.00%> (+0.83%) :arrow_up: src/main/java/seedu/address/model/person/Name.java 90.90% <0.00%> (+0.90%) :arrow_up: ... and 1 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. :umbrella: View full report at Codecov. :loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
2025-04-01T04:10:08.025567
2022-10-29T01:58:44
1428018166
{ "authors": [ "Eclipse-Dominator", "nus-se-script" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13283", "repo": "AY2223S1-CS2103T-T11-1/tp", "url": "https://github.com/AY2223S1-CS2103T-T11-1/tp/issues/102" }
gharchive/issue
[PE-D][Tester A] App has no error message when add command pairs with invalid email format No error message when inputing email address without @ command. It remains at the output of previous input Labels: severity.Medium type.FeatureFlaw original: PokezardVGC/ped#8 This issue will be addressed by #153
2025-04-01T04:10:08.026758
2022-09-27T13:58:33
1387818686
{ "authors": [ "chao890", "kevinchangjk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13284", "repo": "AY2223S1-CS2103T-T15-1/tp", "url": "https://github.com/AY2223S1-CS2103T-T15-1/tp/pull/39" }
gharchive/pull-request
Update delete function for User Guide Resolves #31 Added and modified the existing delete function in the User Guide. Good.
2025-04-01T04:10:08.027636
2022-09-19T10:57:07
1377770162
{ "authors": [ "jialatteo", "wongyewjon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13285", "repo": "AY2223S1-CS2103T-W09-2/tp", "url": "https://github.com/AY2223S1-CS2103T-W09-2/tp/pull/3" }
gharchive/pull-request
Add parameter hints Parameter hints are displayed in ResultDisplay when a command word is typed Resolve #67 Feature would have to update new commands in the future
2025-04-01T04:10:08.039567
2022-10-01T20:58:48
1393549397
{ "authors": [ "alvinjiang1", "codecov-commenter", "jq1836" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13286", "repo": "AY2223S1-CS2103T-W12-3/tp", "url": "https://github.com/AY2223S1-CS2103T-W12-3/tp/pull/55" }
gharchive/pull-request
Update Use Cases and DG Format Things done: Update Use cases to follow the format as stated in Issue #44. Update DESIGN section of DG. Before: After: Update headers: "(OUTDATED)" appended to the end when that section is outdated. Update Glossary to include CSV. Codecov Report Merging #55 (b296751) into master (3fb88d6) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #55 +/- ## ========================================= Coverage 72.15% 72.15% Complexity 399 399 ========================================= Files 70 70 Lines 1232 1232 Branches 125 125 ========================================= Hits 889 889 Misses 311 311 Partials 32 32 Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. Documentation [X] User Guide [X] Developer Guide [X] Javadocs Sections updated under UG NIL Sections updated under DG DESIGN Appendix: Requirements -> Use cases Classes with Javadocs update NIL Here are some guidelines I followed while updating the DG, it would be good to follow these in future commits to DG, to ensure consistency: I intend to follow closely the HighWater DG for most aspects of our DG as I feel that it is very well-written, and if our DG is similar in quality to that, there should be no issues to our documentation. I have also referred to the AB3 DG a lot regarding when to use callouts and markups. Use of present tense to signify work in progress (e.g. developers are to add additional classes...) Make use of Cause - Effect Structure (e.g. CLI gives the user an easy way to input commands... Therefore ...) Liberally make use of Markups (e.g. when describing relevant classes) When coming across jargon, do remember to add it in the Glossary section in alphabetical order. As a follow up of point 4, when referencing words from the Glossary, do add a clickable link that takes the reader back to the glossary section. I have decided that for this particular purpose, the format will just be a bolded, clickable word. Our DG is online so the word will show up in blue. For example, "RC4" is a word in the glossary, and should be written as [**RC4**](#glossary) IMPORTANT!!! When writing the description in your PRs in the future, do add the section I have written at the top. I will check the documentation based on this section of the PR description. Checking the tasklist signifies that that portion of the documentation is ready to be checked (even if it may not have been changed at all). The "Sections updated under..." portion will tell me which part of the UG/DG/java files to look at/compare. As of now, please DO NOT change the headers with the (OUTDATED) word. I intend to check through those sections again before we submit v1.2, and so I will remove it myself :)) Oh one more thing that would be of great help is to deploy the docs on your repo and include a link to the DG. Would help a lot with seeing how the actual documentation will look like.
2025-04-01T04:10:08.049111
2023-10-08T09:04:56
1931719365
{ "authors": [ "codecov-commenter", "derekjxtan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13287", "repo": "AY2324S1-CS2103T-T08-1/tp", "url": "https://github.com/AY2324S1-CS2103T-T08-1/tp/pull/68" }
gharchive/pull-request
Derek/add generic uniqueitemlist Added a generic UniqueItemList class, which mirrors the functionality of uniquePersonList. To be used for patients, doctors and appointments. Meant to replace UniquePersonList. Added interface Listable for all classes that are to be managed using a UniqueItemList. Codecov Report Merging #68 (31bc6c5) into master (b772575) will increase coverage by 0.30%. The diff coverage is 84.90%. :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. @@ Coverage Diff @@ ## master #68 +/- ## ============================================ + Coverage 75.85% 76.15% +0.30% - Complexity 440 459 +19 ============================================ Files 73 76 +3 Lines 1379 1430 +51 Branches 133 140 +7 ============================================ + Hits 1046 1089 +43 - Misses 303 307 +4 - Partials 30 34 +4 Files Coverage Δ ...odel/person/exceptions/DuplicateItemException.java 100.00% <100.00%> (ø) ...model/person/exceptions/ItemNotFoundException.java 100.00% <100.00%> (ø) ...va/seedu/address/model/person/patient/Patient.java 96.55% <100.00%> (+0.25%) :arrow_up: ...u/address/model/person/patient/UniqueItemList.java 82.60% <82.60%> (ø) :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
2025-04-01T04:10:08.054087
2024-10-23T09:14:26
2607866274
{ "authors": [ "InfinityTwo", "codecov-commenter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13288", "repo": "AY2425S1-CS2103T-T10-2/tp", "url": "https://github.com/AY2425S1-CS2103T-T10-2/tp/pull/144" }
gharchive/pull-request
Modify ModelClassDiagram and Diagram Colours Describe your changes Remove tags from ModelClassDiagram Add Department, Role and ContractEndDate to ModelClassDiagram Modify the Diagram colours a bit (for extra lines of code) Issue ticket number and link Builds on #138 Checklist before requesting a review [x] I have performed a self-review of my code. [x] If it is a core feature, I have added thorough tests. Codecov Report All modified and coverable lines are covered by tests :white_check_mark:
2025-04-01T04:10:08.056004
2016-12-14T20:42:00
195640359
{ "authors": [ "Jfairfield1" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13289", "repo": "AZMAG/map-Employment", "url": "https://github.com/AZMAG/map-Employment/issues/33" }
gharchive/issue
Legend doesn't always load properly From @Jfairfield1 on December 14, 2016 20:39 Sometimes the legend will not properly load into the page. My best guess is that a race is going on between the cbr request and the boundaries request. Copied from original issue: AZMAG/map-DemographicState#42 This issue was moved to AZMAG/map-DemographicRegional#40
2025-04-01T04:10:08.072669
2023-04-06T14:50:37
1657574625
{ "authors": [ "lauritapio", "mhinkkan", "murgui" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13290", "repo": "Aalto-Electric-Drives/motulator", "url": "https://github.com/Aalto-Electric-Drives/motulator/issues/85" }
gharchive/issue
mt.BaseValues() examples and documentation don't match expected value in 0.1.2 Hi, I'm using v0.1.2, I tried running 'plot_vector_ctrl_pmsm_2kw.py' and it reports the error Traceback (most recent call last): base = mt.BaseValues( TypeError: BaseValues.__init__() got an unexpected keyword argument 'n_p' I checked the reference at https://aalto-electric-drives.github.io/motulator/autoapi/motulator/index.html#motulator.BaseValues and indeed the example used the suggested arguments. I removed the argument and then the error was that it was missing argument 'p', so it looks like n_p has been renamed to p, so either the examples should be changed to p or the function should be changed to n_p. Regards Thanks for pointing this out! The polepair number was indeed recently named to n_p, but it seems that there are some places still where this has not been changed. We'll get this fixed shortly! Thank you for the valuable feedback! The reason for the problem might be that we have updated the repository since last release quite a lot, including the change in the variable name (p -> n_p). To fix the problem, new release 0.1.3 was published a moment ago. Furthermore, the documentation regarding installation was updated: https://aalto-electric-drives.github.io/motulator/installation.html You might be interested in Option 2, i.e. cloning the repository. It allows you to modify the existing system models and the example controllers. The example instructions are for VS Code, but quite similar steps could be used with other IDEs. Please inform us if the problem still appears after installing the latest release.
2025-04-01T04:10:08.092661
2021-12-17T09:01:41
1083031677
{ "authors": [ "AaronErhardt", "euclio" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13291", "repo": "AaronErhardt/relm4", "url": "https://github.com/AaronErhardt/relm4/issues/57" }
gharchive/issue
Inconsistent behavior between FactoryVec and FactoryVecDeque I'm trying to create a widget that adds and removes an arbitrary number of children at the position before some hard-coded widgets. I tried modifying the factory example like so: gtk::ApplicationWindow { set_default_width: 300, set_default_height: 200, set_child = Some(&gtk::Box) { set_orientation: gtk::Orientation::Vertical, set_margin_all: 5, set_spacing: 5, factory!(model.counters), append = &gtk::Button { set_label: "Add", connect_clicked(sender) => move |_| { send!(sender, AppMsg::Add); } }, append = &gtk::Button { set_label: "Remove", connect_clicked(sender) => move |_| { send!(sender, AppMsg::Remove); } }, append = &gtk::Button { set_label: "I'M LAST", }, } } If I use FactoryVec for my storage, the new widgets are added and removed at the last position (after the "I'M LAST" label). However, if I use FactoryVecDeque for my storage, the widgets are added at the first position (before the "Add" button). The latter behavior happens to be what I want, but I wanted to open this issue to see if this is something I can rely on, or if there's a better way to accomplish what I want. The reason why this happens is because there are different traits used behind the scenes. FactoryView is used by both but FactoryListView is only required by FactoryVecDeque to enable more flexibility. That's because a Vec can only push and pop at the end whereas the VecDeque needs to take care about positioning. Since not all widgets support arbitrary positions both traits exist to at least allow FactoryVec to be used. This also causes FactoryVec to use append to insert widgets and FactoryVecDeque uses insert_after unless it's the first widget. Of course it's better to use a separate container only for your factory but I'm pretty sure that this behaviour won't change anytime soon. Of course it's better to use a separate container only for your factory Agreed, but in this case I need the children to all be children of the same widget (a PanelPaned from libpanel). I'd suggest then to document the current behaviour. Any change to this (which is not very likely to happen anyway) would then be avoided or at least considered a breaking change. The behavior is now documented: https://aaronerhardt.github.io/docs/relm4/relm4/factory/collections/index.html
2025-04-01T04:10:08.115400
2019-02-26T09:47:45
414519947
{ "authors": [ "nre-ableton", "rco-ableton" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13292", "repo": "AbletonDevTools/groovylint", "url": "https://github.com/AbletonDevTools/groovylint/pull/55" }
gharchive/pull-request
Add logging support to run_codenarc.py This PR adds logging support to run_codenarc.py, and also adds a custom log formatter so that the CodeNarc output matches the output of this script. With these changes, a sample --verbose run (with errors) looks like this: DEBUG Executing CodeNarc command: java -Dorg.slf4j.simpleLogger.showThreadName=false -Dorg.slf4j.simpleLogger.defaultLogLevel=debug -classpath /Users/nre/Code/AbletonDevTools/groovylint/jars:/usr/local/opt/groovysdk/libexec/lib/*:/Users/nre/Code/AbletonDevTools/groovylint/jars/CodeNarc-1.2.1.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/GMetrics-1.0.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/slf4j-1.7.25/slf4j-api-1.7.25.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/slf4j-1.7.25/slf4j-simple-1.7.25.jar org.codenarc.CodeNarc -rulesetfiles=ruleset.groovy -report=xml:codenarc-report.xml -includes=./Jenkinsfile,**/*.groovy INFO org.codenarc.ruleregistry.PropertiesFileRuleRegistry - Loaded properties file in 43ms; 362 rules INFO org.codenarc.ruleset.GroovyDslRuleSet - Loading ruleset from [ruleset.groovy] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/basic.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/braces.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/convention.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/design.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/exceptions.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/formatting.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/generic.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/groovyism.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/imports.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/naming.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/security.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/size.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/unnecessary.xml] INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/unused.xml] INFO org.codenarc.ruleset.PropertiesFileRuleSetConfigurer - RuleSet configuration properties file [codenarc.properties] not found. DEBUG org.codenarc.CodeNarcRunner - results=DirectoryResults(null) [DirectoryResults() [DirectoryResults(vars) [], DirectoryResults(jars) [], FileResults(Jenkinsfile) [Violation[rule=VariableTypeRequiredRule[name=VariableTypeRequired, priority=3], lineNumber=59, sourceLine=def foo = 'bar', message=The type is not specified for variable "foo"], Violation[rule=UnusedVariableRule[name=UnusedVariable, priority=2], lineNumber=59, sourceLine=def foo = 'bar', message=The variable [foo] in class None is not used]]]] INFO org.codenarc.report.AbstractReportWriter - No custom message bundle found for [codenarc-messages]. Using default messages. INFO org.codenarc.report.AbstractReportWriter - Report file [codenarc-report.xml] created. INFO CodeNarc completed: (p1=0; p2=1; p3=1) 7291ms DEBUG CodeNarc returned with code 0 DEBUG Reading report file codenarc-report.xml DEBUG Removing report file codenarc-report.xml DEBUG Parsing report XML INFO Scanned 4 files DEBUG Parsing violations in package: . DEBUG Parsing violations in file: ./Jenkinsfile ERROR ./Jenkinsfile:59: VariableTypeRequired: The type is not specified for variable "foo" ERROR ./Jenkinsfile:59: UnusedVariable: The variable [foo] in class None is not used ERROR Found 2 violation(s) ping @AbletonDevTools/gotham-city Hmm, it seems that pylint does not like the fact that --verbose appears in two separate files. 🙄 How should I fix this? Can I just # noqa it? Sorry for the force-push, but I made some pretty big changes on this PR and the fixups would have gotten out of control quickly. Namely, I decided to abandon the idea of using a custom Formatter class and now instead introduced a function to examine each line of CodeNarc's output and re-log it appropriately. I realized that previously we were not respecting CodeNarc's logging levels and were just logging everything it produced as logging.DEBUG, which is not very nice. I think that it's better to look at the actual log level (which is pretty easy to parse) and use the same when calling logging.log(). @rco-ableton @mst-ableton do either of you have an opinion on this PR? @nre-ableton please rebase and I'll review! @rco-ableton Done!
2025-04-01T04:10:08.124695
2019-07-19T19:48:07
470511030
{ "authors": [ "AnshumanDwivedi", "geethab123", "realraviteja", "satish765001", "yruslan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13293", "repo": "AbsaOSS/cobrix", "url": "https://github.com/AbsaOSS/cobrix/issues/140" }
gharchive/issue
multi copybook or not I have a copybook which has multiple 01 levels. Can this copy book is parsed as variable length. can I parse like below cobolDataframe = spark .read .format("cobol") .option("copybook", v_copybook) .option("schema_retention_policy", "collapse_root") //removes the root record headerc .option("drop_group_fillers", "false") .option("generate_record_id", false) // this adds the file id and record id .option("is_record_sequence", "true") // reader to use 4 byte record headers to extract records from a mainframe file .option("is_rdw_big_endian", "true") //.option("is_rdw_part_of_record_length", false) .option("rdw_adjustment", -4) .load(v_data) Please confirm. After parsing I am not getting correct parsed data. Yes it can. If your level 1 entries represent segments they should also redefine each other. 01 RECORD1. ... 01 RECORD2 REDEFINES RECORD1. ... They are not redefining. Still can be treated as variable length Seems Cobrix is not able to process multiple copybooks like syncsort is doing. @CobrixTeam : could you please check and suggest in this regards ? Sure, we can help. Could you provide an example set of copybooks and a small example data file? @yruslan we are also having similar issue: Quick summary About the issue: In my project i have a requirement to read multiple copybooks as One main Copybook using Cobrix. Example: TRANSACTION.CPY 01 CUSTOMERS 05 PURCHASES FIRST_NAME PIC X (04) LASTNAME_NAME PIC X (04) 01 ORGANIZATION 05 DEPARTMENTS ORG_NAME PIC X (04) VENDOR_NAME PIC X (09) 01 MEDICARE 05 BILLS TREATMENT_TYPE PIC X (04) LOCATION_NAME PIC X (09) 01 MEMBERSHIP 05 PARTNERS PARTNER_TYPE PIC X (04) PARTNER_NAME PIC X (09) I have similar kind of layout copybook and i'm using cobrix to parse the data. In my research what i understood is i cant read any copy which will have multiple main headers . Cause Cobrix can handle only one Main header copybook no multiple. Please provide your thoughts and resolutions for the above issue. i have similar requirement where copybook contains three 01 levels . One header , one trailer and the another for data with occurs depending on . Header and trailer has length 30 and data levels varies as its occurs depending on. Any suggestion would be greatly appreciated .
2025-04-01T04:10:08.130384
2018-10-25T15:46:54
374014011
{ "authors": [ "Jack12816", "mungler" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13294", "repo": "Absolventa/emarsys-rb", "url": "https://github.com/Absolventa/emarsys-rb/pull/61" }
gharchive/pull-request
Improved the Contact.query method - What is it good for This PR ships an enhanced Contact.query method which allows to query all contacts without requiring any filter. Furthermore it transforms the key_id and key_value when passed. - What I did I added the transformation of the passed key_id and the possibility to pass in custom filters. The YARD documentation was updated to reflect this with more examples and for the new usages specs were added. - A picture of a cute animal (not mandatory but encouraged) Looks good 👍 and kudos on the cute animal 😍 😁 FYI @Jack12816 published new version 0.3.13 to rubygems.org with these changes 👍
2025-04-01T04:10:08.133358
2024-08-14T03:10:13
2464798660
{ "authors": [ "jstone-lucasfilm", "ld-kerley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13295", "repo": "AcademySoftwareFoundation/MaterialX", "url": "https://github.com/AcademySoftwareFoundation/MaterialX/pull/1970" }
gharchive/pull-request
Refine MaterialX render module build configuration Add more refined control over building MaterialX render modules. Specifically USD only needs the core render module in some build configurations (iOS). Also add a guard to ensure that MaterialXGenGlsl is built if MaterialXGenMsl is built, because of the dependency in the data library. USD uses MaterialX::GlslShaderGenerator::TARGET to locate those resources. @ld-kerley I like the idea behind this proposal, though the split between MATERIALX_BUILD_RENDER and MATERIALX_BUILD_RENDER_CORE seems counterintuitive to me, as there's no corresponding module named MaterialXRenderCore. As an alternative, what if we added a MATERIALX_BUILD_RENDER_PLATFORMS flag, defaulting to ON, with all platform-specific render modules gated by this flag? This would allow USD developers to opt out of platform-specific render modules such as MaterialXRenderGlsl and MaterialXRenderOsl, without losing access to the MaterialXRender module that they still need. Thanks for the updates, @ld-kerley, and this looks really promising! I'll take a closer look in upcoming days, and let's see if we can get this change ready to include in MaterialX v1.39.2.
2025-04-01T04:10:08.136263
2021-04-07T22:28:10
852875935
{ "authors": [ "lgritz", "pellerington" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13296", "repo": "AcademySoftwareFoundation/OpenShadingLanguage", "url": "https://github.com/AcademySoftwareFoundation/OpenShadingLanguage/pull/1355" }
gharchive/pull-request
Added farmhash for lower OpenImageIO versions Older versions of OpenImageIO don't have the farmhash which is now nessisary for optix 7 strings. To fix this, I added the file into into OpenShadingLangauge wrapped around version guards. Some existing functions had to be included again with OIIO_HOSTDEVICE added (Uint128Low64, Uint128High64, Uint128, CopyUint128, Hash128to64) so I wrapped them in the "inlined" namespace to prevent confilcts. I made these changes so that I could build OSL with my version of OIIO, I am opening this pull request incase there is any interest in supporting it in the official branch. Potentially the VFX platform might have moved far enough by the time of the OSL release. Hi, sorry for the delay in responding to this (the last couple weeks have been really busy for me). I've backported these farmhash fixes to the OIIO release branch and it's included in OIIO <IP_ADDRESS>. The in-development OSL (and in particular the OptiX support) is technically considered "experimental/optional", it's a pre-release, so I think it's not unreasonable to expect it to be built against the latest OIIO release. I think I would prefer to require using a sufficiently recent OIIO release than to have to duplicate this code in the OSL codebase as well. But if people are really inconvenienced by that, I can certainly reconsider.
2025-04-01T04:10:08.139044
2024-09-02T22:44:05
2501651867
{ "authors": [ "cary-ilm" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13297", "repo": "AcademySoftwareFoundation/OpenTimelineIO", "url": "https://github.com/AcademySoftwareFoundation/OpenTimelineIO/issues/1791" }
gharchive/issue
Consider signing release artifacts The OpenSSF Best Practices Badge suggests signing release artifacts. Consider using OpenEXR's release-sign.yml workflow as a template. It's triggered on release creation and does these steps: Runs get archive to generate a <release>.tar.gz artifact Signs the <release>.tar.gz via sigstore Uploads the resulting sigstore signature file along with the tarball. Oops, this duplicates #1782, but with a bit more explicit suggestions!
2025-04-01T04:10:08.154579
2024-11-12T14:07:32
2652299381
{ "authors": [ "evelynnesher", "sgbaird" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13298", "repo": "AccelerationConsortium/ac-training-lab", "url": "https://github.com/AccelerationConsortium/ac-training-lab/issues/113" }
gharchive/issue
Autotrickler v4 - custom RS232 control with A&D scale As a follow up to #20, the goal is to get a similar interface and IolT setup to that of the US Solid scale. Some related material at https://ac-microcourses.readthedocs.io/en/latest/courses/robotics/3.2-serial-communication.html. In particular: ✅ Read HR Analytical Balance Literature This document provides an overview of the analytical balance we'll be interfacing with. Understanding the capabilities and specifications of the instrument is crucial for effective integration. ✅ Skim HR Analytical Balance Manual. Read Sections 15 and 17 closely. Sections 15 and 17 detail the RS232C interface and the commands used to communicate with the balance. This information is essential for programming our microcontroller to interact with the balance correctly. The connection at the back of the scale is similar. We have two scales (the 1 mg and the 0.1 mg resolution). The 0.1 mg resolution scale is at NRC. Let's focus on the 1 mg at UoT. Iset up the RS232 connection with the scale. As shown below, I needed to unplug the auto trickler as it was plugged into the RS-232C serial interface. Currently, it prints “ST” when the weight measurement is stable and “US” when the weight is actively changing and unstable. That was fast! Great. There should be a second pico 2ch rs232 module available (I think in the same room?), so we can have one for the US solid scale and one for the A&D scale in the office. Got HiveMQ set up and publishing with no issues, so I’m moving on to the app now. Do you want this one to have the autotrickler feature too? I’ll have to figure out a way to make it work since they both use the same port. Nice! You'll use the same port, but adjust the MQTT topic structure to differentiate between the two devices. I think a separate app might be better, but I'm open to the idea of integrating it into the existing app. I think you misunderstood, I meant the physical port on the scale, where both the Raspberry Pi (for the RS232 connection) and the AutoTrickler are connected. But as you can see, only one can connect at a time. Ah, my bad 😅 The autotrickler can be left disconnected for now. I set up the app in a private space on my personal account to test it out first. It is currently identical to the previous one but ready to be implemented into the AC Hugging Face Spaces. App is public and scale is embedded into GatherTown. Done! I added a connection status bar that is identical to the one for the continuous output scale. I set the connection status refresh rate to 10 seconds, matching the other scale. The graph now skips unstable data inputs, keeping it cleaner and only showing weights of properly placed items—not the weight of your hand as you place them. Thanks! It's not urgent, but I'd rather that the unstable values be included to be more realistic to if someone had it in front of them. Likewise, there may be other uses for what would be considered unstable data.
2025-04-01T04:10:08.156509
2016-05-05T16:50:28
153273577
{ "authors": [ "dsingh07", "nickdgriffin" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13299", "repo": "Accenture/adop-docker-compose", "url": "https://github.com/Accenture/adop-docker-compose/pull/65" }
gharchive/pull-request
Adding minor bug fixes Bug fixes: Certificate generation will now exit with a helpful message if bash is not run as admin (openssl command to create a cert will fail since it needs to source a .srl file) Fix to allow "source ./conf/env.provider.sh" to be run if CONF_PROVIDER_DIR is not set, where we just use the current directory as a reference LGTM.
2025-04-01T04:10:08.158444
2022-10-04T11:25:42
1396101078
{ "authors": [ "DannyCastroVentura", "JoernBerkefeld" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13300", "repo": "Accenture/sfmc-devtools-copado", "url": "https://github.com/Accenture/sfmc-devtools-copado/pull/112" }
gharchive/pull-request
#111: set skipInteraction properly to avoid mcdev's interactive wizard mode PR details What is the purpose of this pull request? (put an "X" next to an item) [x] Bug fix What changes did you make? (Give an overview) closes #111 Is there anything you'd like reviewers to focus on? the lines where ['npm install ${installer} --foreground-scripts', CONFIG.mcdev_exec + ' --version'] gets changed are just line break changes caused by a recent update to eslint/prettier Ok, I thought I was changing the issue I've created, but instead was changing this one. Sory, mind bug
2025-04-01T04:10:08.167888
2024-02-19T15:55:14
2142689272
{ "authors": [ "DataTriny", "mwcampbell" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13301", "repo": "AccessKit/accesskit", "url": "https://github.com/AccessKit/accesskit/pull/352" }
gharchive/pull-request
fix: Factor out core AT-SPI translation layer This will allow us to re-use the logic for translating between AccessKit and AT-SPI, in both the current Unix adapter and a new client-side library (for ATs like Orca) that is being developed as part of the new GNOME accessibility stack. I don't expect to make any major changes to the accesskit_atspi_common API at this point, though I can't rule out needing to make a few more tweaks similar to my last few pushes. Now I'm ready to show what I'm doing with this refactor. Note that the new GNOME accessibility stack is code-named Newton. See this repository with a newton_atspi_compat crate and a Python binding on top: https://gitlab.gnome.org/mwcampbell/newton_atspi_compat and this proof-of-concept Orca modification that uses the Python module: https://gitlab.gnome.org/mwcampbell/orca/tree/newton It's still kind of a mess. I might be doing too much directly in the Python binding. And I do have to write a lot of conditionals in Orca, mostly because Orca is hard-coded to call functions on Atspi.Accessible, so I can't just take advantage of duck typing. And I need to document the steps to set up a complete working system, including my fork of Mutter (the compositor). But the Newton protocols are now working end to end, without passing through the AT-SPI D-Bus interface as I did in an earlier demo. And, back to this PR, my hope is that as we continue to develop the AccessKit AT-SPI backend, most of that work can be reused for the Newton project. As soon as I add support for the action and component interfaces to the new "simplified" API I just pushed, I will be done adding things to this PR. I don't plan to add or change any more in this PR. Once it passes review, I'm ready to merge it. I wonder, though, if we should classify this as a refactor instead of a fix, even though the only way to do that is to use refactor!, which will trigger a breaking version increment on accesskit_unix. Really though, the main user of accesskit_unix so far is accesskit_winit, and the automatic dependency bump shouldn't cause a breaking version increment on accesskit_winit. Yes, but then accesskit_unix would have been broken on the main branch in the meantime, leading to failing CI. I'm not sure that release-please really gives us a good option here. I do have a breaking change in mind for accesskit_unix. To prepare the API to support both Newton and legacy AT-SPI in a single adapter, I believe we'll have to modify Adapter::new to take a RawWindowHandle, since the Newton backend requires a Wayland display and surface. I don't see why this would have broken accesskit_unix? We'd just have a lot of duplicate code for a brief period of time. Oh, you're right, of course. Then I can still pull the accesskit_unix changes out into a separate PR. Doing it in two steps would also make it a bit simpler for me to rebase my unix-text branch. Then I'll back out the accesskit_unix changes in this PR, and open the second one when this PR is merged. Before I do that, do you have any feedback about the design as a whole (that would require me to change both sides at once)? I think you already addressed my concerns. The last remaining one would be from an outsider point of view: we know the dependency count of accesskit_unix is an issue for some downstream projects, yet here you are adding another one. I haven't looked into this, but I wouldn't be surprized if build times would slightly decrease because of that though. It bothers me that Bevy isn't enabling the Unix adapter by default. But I can't let perceptions related to dependency count and binary size dictate all design decisions, especially when it comes to adding just one small crate. Anyway, once Newton is stabilized, the accesskit_unix adapter could let users disable AT-SPI and only enable Newton via features, thus eliminating the dependency on zbus. OK, the changes to accesskit_unix are now in the unix-use-atspi-common branch, which I'll squash and rebase once this is merged. Oh, one final comment but it's not a blocker: I wonder if the platforms directory really is the best place for this crate. I now have the same feeling for accesskit_winit actually. I agree it's not ideal, but nothing better immediately comes to mind. We can always move it later without breaking anything user-visible. Are we ready to merge this first PR?
2025-04-01T04:10:08.214859
2024-04-30T16:21:48
2271962172
{ "authors": [ "dhensle", "i-am-sijia" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13302", "repo": "ActivitySim/activitysim-prototype-mtc", "url": "https://github.com/ActivitySim/activitysim-prototype-mtc/issues/12" }
gharchive/issue
Full Scale Performance: Sharrow On This is the issue to report on memory usage and runtime performance when using sharrow... First ran sharrow compile with the following settings: households_sample_size: 100 sharrow: test Run completed in 76 minutes. log_sh_compile.zip Then ran in production mode households_sample_size: 0 100 percent sample sharrow: require Run completed in 7.7 hours with a memory peak at about 163 GB in trip destination. logs_sh_full.zip Followed by multiprocessing households_sample_size: 0 100 percent sample sharrow: require multiprocessing: True num_processors: 24 Run completed in 110 minutes (1.8 hours). log_sh_full_mp.zip Ran with 100% households and sharrow on, single process. Run completed in 1090.3 minutes (18.2 hours). This is much longer than the previous time posted above of 7.7 hours. Current run was performed using PR #867 commit c9d4205. log.zip Timing statements comparing the old run above to this current run show large differences mainly in the destination models: Will try again with the main branch of ActivitySim instead of PR 867 to see if that makes a difference. Ran using an older environment that uses the current version of ActivitySim (main@bd48d3db), but has sharrow v2.8.2 instead of the previous run's main@8d63a66 (> v2.9.1). Numba was also older using 0.56.4 compared to 0.59.1. The run results were pretty much exactly the same -- run time was 1080.3 minutes. log.zip One difference between these current set of runs and the 7.7 hour run above is the server. The 7.7 hour run was done on SANDAG's 1TB RAM, 40 Core machine. These were done on RSG's 500 GB RAM, 24 core machine. Sharrow, single process, MTC extended model ran in 10.7 hours on WSP's 512 GB RAM, AMD server. Using everything the latest as of June 26. Memory peak 145 GB in trip destination. ActivitySim: pr/867@c9d4205 Sharrow: v2.10.0 MTC: extended@a3da8bd activitysim.log timing_log.csv Running the same tests as above and on the same machine, but using multiprocessing instead of multi-threading: Comments: 16 and 24 core runs are incomplete due to https://github.com/ActivitySim/activitysim/issues/876 Saw roughly linear decreases in runtime for computationally intensive models going from 4 to 12 cores, but after that the gains decreased. 20 cores took longer than 12 cores. This is due to some models being slower (school escorting, school location, joint tour scheduling, etc), and increased time spend apportioning and coalescing all of the cores. However, this runtime difference was pretty minimal. The runtime in the final activitysim.log file is slightly longer than the total in the timing_log.csv file across all runs. The difference increases with the number of cores.
2025-04-01T04:10:08.249868
2017-12-18T16:13:21
282934161
{ "authors": [ "CheezItMan", "mcgmar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13303", "repo": "Ada-C8/ada-trader", "url": "https://github.com/Ada-C8/ada-trader/pull/38" }
gharchive/pull-request
Carets - Maria - ada-trader Ada Trader Congratulations! Comprehension Questions Question Answer How do Backbone Views help structure your code? The views help with separation of concerns and acted as a facilitator between data, like functions in the models, and between the DOM. Did you use jQuery directly in your Views? How, and why? I did via the $el which is a jQuery object. I did so so that the views would render. What was an example of an event you triggered? Why did you need to trigger it? An event I triggered was listing the quotes. That list change was triggered when a user clicked the buttons. In what was is unit testing in JavaScript similar to unit testing in Ruby? In what ways is it different? The tests are similar in that they had validations and you tested for each function that was built. Ada Trader What We're Looking For Feature Feedback Core Requirements Git hygiene Not very many commits, good commit mesages Comprehension questions Check, however you didn't actually manually trigger any events. Organization Models and collections are defined in separate files Check Code that relies on the DOM is located in or called by $(document).ready Check Functionality Quote prices change when clicking Buy and Sell Check The Trade History updates when buying and selling a quote Check A user can create an open order using the Order Entry Form NOT WORKING An open order removes itself from the open orders and updates the Trade History when fulfilled MISSING General Has separate views for different parts of the app Check Uses events (listening/handling and triggering) to manage different behavior in views You have the QuoteListView listening for changes in the collection and models and re-rendering. However You don't have the OrderListView working or listening to the form, or any component listening for changes in the Quote instances. Practices good standards for views (ES6 syntax, doesn't directly use jQuery, returns this in render) Check, but you're not utilizing the form for the OrderListView. Error handling for the Order Entry Form MISSING Testing Has unit tests for models MISSING Overall You've got a lot incomplete here. You set up a view for the Quotes and QuoteListView perfectly and append trades to the list well. You didn't complete the OrderView or OrderListView or complete any testing. You can see some notes I put in your code as to what could be done to get it working. The biggest error is the fact that both QuoteListView and OrderListView are both tied to the same main html element which means they could theoretically interfere with each other. See my comments in your code.
2025-04-01T04:10:08.279264
2016-02-10T03:19:38
132597357
{ "authors": [ "SirePi", "brogine", "ilexp" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13304", "repo": "AdamsLair/duality", "url": "https://github.com/AdamsLair/duality/issues/276" }
gharchive/issue
Rectangle shape for rigid collision detection I'm currently testing the framework, which i consider very simple and easy to learn as about now. I'm trying to make a rpg-styled game but i think polygon is not very accurate in this case. I think Rectangle must be added when editing rigid bodies. That'd solve a few of my problems. Thanks Adam for your work. Well, a Rectangle is in fact a Polygon.. if you check the Rigidbody in the object inspector panel, you will see that you can edit the vertices individually, so it's easier to position them to form a perfect rectangle. Basically what SirePi said: You can express a rectangle using the existing polygon shape. Still, it would of course be more convenient if there was an actual "Rect" control, so you would be able to quickly create rectangular polygons. I'm aware that the current RigidBody Editor is not ideal in some cases (see issue #54), but this is currently a rather low-priority task, since all the typical use cases can be accomplished - just sometimes not in the most convenient way. Other than that, I'm currently working on a Tilemaps plugin, which you can find more about in issue #249 or in the development report forum thread. It won't be released (or officially supported) anytime soon, but in case you're curious, the latest state is available in the tilemap_wip branch on GitHub. From a more practical perspective, there are a lot of tilemap-based projects on the forum, so I imagine you could get more advice there, if you need help. Yes, but if you have a scene with 50 polygons, you have to draw everyone of them by hand and then edit them individually, there's some time there. If you try to check collisions between hand-drawed polygons, you don't get what you are expecting. I just think this as a minor but useful feature. Yes, but if you have a scene with 50 polygons, you have to draw everyone of them by hand and then edit them individually, there's some time there. Hmm, I think I already answered this, see here: Still, it would of course be more convenient if there was an actual "Rect" control, so you would be able to quickly create rectangular polygons. I'm aware that the current RigidBody Editor is not ideal in some cases (see issue #54), but this is currently a rather low-priority task, since all the typical use cases can be accomplished - just sometimes not in the most convenient way. So yes, you definitely have a point that I acknowledge and it would be more convenient to have a distinct Rect tool, but we already have an issue for this - even though it's currently a low-priority task. :) For now, you could maybe copy-paste your RigidBodies (Perform a right-click-drag from the source to the target GameObject), use Prefabs, write a custom editor plugin with a script for rect shape generation or write a custom Component that automatically generates a RigidBody for tile-like objects - whichever suits your use case best.
2025-04-01T04:10:08.283001
2024-04-14T01:12:27
2241848170
{ "authors": [ "illuhad", "nilsfriess", "ouankou" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13305", "repo": "AdaptiveCpp/AdaptiveCpp", "url": "https://github.com/AdaptiveCpp/AdaptiveCpp/pull/1436" }
gharchive/pull-request
Replace deprecated APIs in the src and tests While building ACPP from source and running the built-in tests, I noticed a few warnings caused by deprecated APIs. bool llvm::StringRef::startswith(llvm::StringRef) const (in the source) warning: ‘bool llvm::StringRef::startswith(llvm::StringRef) const’ is deprecated: Use starts_with instead [-Wdeprecated-declarations] Atomic accessors in SYCL 2020 (in the tests) warning: 'operator[]' is deprecated: Atomic accessors are deprecated as of SYCL 2020 [-Wdeprecated-declarations] 'get_count' in SYCL 2020 (in the tests) warning: 'get_count' is deprecated: renamed to 'size' in SYCL 2020 Specification [-Wdeprecated-declarations] I'm still new to SYCL, so please let me know if the proposed change contains any errors. Thanks! Re the deprecated SYCL functions: Since these are only deprecated and not (yet) removed from the SYCL spec, I think we should still include them in the tests. Ideally we should test both the deprecated function and the "replacement". Maybe we can consider disable deprecation warnings for the tests to reduce the noise Re the deprecated SYCL functions: Since these are only deprecated and not (yet) removed from the SYCL spec, I think we should still include them in the tests. Ideally we should test both the deprecated function and the "replacement". Yes. They are officially still part of SYCL, and SYCL CTS also tests for deprecated APIs. Thanks for the clarifications! I'll close the PR since those deprecated APIs must still be kept and tested.
2025-04-01T04:10:08.445221
2021-11-11T20:16:18
1051326512
{ "authors": [ "ameshkov", "dnmTX", "krystian3w", "sfionov", "slavaleleka" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13306", "repo": "AdguardTeam/CoreLibs", "url": "https://github.com/AdguardTeam/CoreLibs/issues/1550" }
gharchive/issue
Add regexp support for $domain modifier now there is no chance to create single cosmetic rule for hiding the same element on such domains: example1.com, example2.com, example3.org it would be nice if $domain modifier supports regular expressions, just like $path does something like this: [$domain=/example[0-9]\.(com|org)/]##.ad Why limit just to the cosmetic rules? If we do it here, we should extend it to basic rules as well. @sfionov can we include this in v1.10 for instance? Yes, assigned to 1.10 https://github.com/AdguardTeam/tsurlfilter/issues/41 https://github.com/AdguardTeam/tsurlfilter/issues/58 https://github.com/AdguardTeam/AdguardKnowledgeBase/pull/1109 @AdguardTeam maybe unrelated question but what about redirects? For example links from voe.sx are allways redirected to another domain and they change those domains very frequently. Here is a example: Original Link: https://voe.sx/ffgecklhux79 redirects(as of now) to: https://tubelessceliolymph.com/ffgecklhux79 So basically rules that applied for voe.sx simply don't work. And domains like tubelessceliolymph.com are disposable and replaced every week or so and user(s) like me have to reaply same rules to the new domain. Any workaournd(s) for this? uBo implemented regex, if someone trust something like: /bad-js.$domain=/[a-z-]{18}\.[\w]{2,20}/ /[a-z-]{18}\.[\w]{2,20}/##.bad-class but fail for , in cosmetic/scriptlet filters and | in network filters if gorhill rejected repair: https://github.com/uBlockOrigin/uBlock-issues/discussions/2234#discussioncomment-5403472 https://github.com/AdguardTeam/AGLint/issues/161
2025-04-01T04:10:08.446357
2018-10-17T12:56:32
371059906
{ "authors": [ "sfionov", "szolin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13307", "repo": "AdguardTeam/CoreLibs", "url": "https://github.com/AdguardTeam/CoreLibs/issues/545" }
gharchive/issue
Don't send onBrowserApi callback on wrong token Also, it will be better if, in case of invalid token, message will say about this and will have DEBUG log level, not ERROR. ERROR AGWebSocketHandler *7: json request parse: (6) bad value last value:d83cfbe7-bd3b-4d27-9364-a949c7462ba4 /core/pull-requests/747
2025-04-01T04:10:08.449726
2019-06-26T21:56:47
461199688
{ "authors": [ "ameshkov" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13308", "repo": "AdguardTeam/CoreLibs", "url": "https://github.com/AdguardTeam/CoreLibs/issues/986" }
gharchive/issue
Wrong web page encoding or Automatic download page files @Chandler-Lu commented on Wed Jun 26 2019 When I visit a page without HTTPS, just like http://bbs.pcbeta.com/forum-win10-1.html, the web page may display garbled characters or Automatic download page file (forum-win10-1.html). Then I need to refresh some times, the page can display normally. Expected behavior display normally Actual behavior garbled characters, lose css style, automatic download page file Your environment Environment name and version: (e.g. Chrome 59): safari 12.1.1 (146<IP_ADDRESS>.1) Any specific potentially conflicting software installed: (e.g. antiviruses, firewalls, traffic counters, cleaners): ClashX Cannot reproduce this anymore, seems to be fixed
2025-04-01T04:10:08.502932
2024-05-19T16:42:41
2304691110
{ "authors": [ "Admonstrator" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13309", "repo": "Admonstrator/glinet-uptime", "url": "https://github.com/Admonstrator/glinet-uptime/issues/65" }
gharchive/issue
⚠️ GLDDNS Server (ns2.glddns.com) has degraded performance In b982428, GLDDNS Server (ns2.glddns.com) (ns2.glddns.com) experienced degraded performance: HTTP code: 200 Response time: 495 ms Resolved: GLDDNS Server (ns2.glddns.com) performance has improved in b5674ea after 6 minutes.
2025-04-01T04:10:08.537482
2020-05-01T16:53:01
610852042
{ "authors": [ "joerghoh" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13310", "repo": "Adobe-Consulting-Services/acs-aem-commons", "url": "https://github.com/Adobe-Consulting-Services/acs-aem-commons/issues/2282" }
gharchive/issue
Do not start services on AEMaaCS instances if they are not working there In the documentation a number of features are listed already as "do not use them when you are deploying on AEM as a CloudService". Right now this is not technically enforced, but with the RequireAem feature (see https://github.com/Adobe-Consulting-Services/acs-aem-commons/blob/b391b2455bb1aac018183ae4b76794379ebb8a59/bundle/src/main/java/com/adobe/acs/commons/util/RequireAem.java) it can be enforced quite easily. Right now these features are tagged as "do not run within AEMaaCS": bulk-workflow-manager ensure-oak-index healthchecks/smtp healthchecks/status-emailer healthchecks json-event-logger versioned-clientlibs workflow-processes/assets-rendition-matter workflow-processes/assets-watermark-process workflow-processes/brand-portal file-fetcher mcp-tools/asset-ingestion/asset-ingestor mcp-tools/asset-ingestion mcp-tools/asset-ingestion/s3-asset-ingestor mcp-tools/asset-ingestion/url-asset-ingestor mcp-tools/refresh-folder-thumbnails redirect-map-manager remote-assets (I have extracted this list from the official documentation) see #2285 #2285 has been merged, closing this one.
2025-04-01T04:10:08.544968
2020-03-25T14:07:20
587722286
{ "authors": [ "adamcin", "davidjgonzalez", "justinedelson", "kwin" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13311", "repo": "Adobe-Consulting-Services/acs-aem-commons", "url": "https://github.com/Adobe-Consulting-Services/acs-aem-commons/pull/2233" }
gharchive/pull-request
Feature/2232 - Removes META-INF/vault files in SCM; Minor sonarqube fixes Removes the old properties.txt and friends (cnd, etc.) from SCM. Fixed a few CM related quality gate checks to allow it to pass/deploy to CM. @davidjgonzalez this is a good idea. have you done a comparison between the generated META-INF/vault/* files between the last release and with these changes? If not, I'll try to script that and post the results here. @justinedelson adding the generated files for ui.apps and ui.content here's the ui.apps properties.txt for a quick look <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd"> <properties> <comment>ACS AEM Commons UI.Apps Package</comment> <entry key="acHandling">merge</entry> <entry key="created">2020-03-25T11:31:00.424-0400</entry> <entry key="groupId">com.adobe.acs</entry> <entry key="description">ACS AEM Commons content package containing code (/apps).</entry> <entry key="version">4.4.3-SNAPSHOT</entry> <entry key="requiresRoot">false</entry> <entry key="dependencies">day/cq60/product:cq-content:6.3.64</entry> <entry key="path">/etc/packages/adobe/consulting/acs-aem-commons-ui.apps.zip</entry> <entry key="createdBy">davidg</entry> <entry key="name">acs-aem-commons-ui.apps</entry> <entry key="artifactId">acs-aem-commons-ui.apps</entry> <entry key="group">adobe/consulting</entry> </properties> ui.apps.meta-inf-vault.zip ui.content.meta-inf-vault.zip side by side diff: https://gist.github.com/justinedelson/8392242adf39ef7e746bf273e4a5df98 base don the diff, looks like a benign change but keeps the code repo a little cleaner. Am planning on cutting a release for @wimsymons since we've been putting him off for too long, and didnt want to get us caught w/ that possible properties.txt issue. I think this breaks oakpal. I now get errors like [INFO] --- oakpal-maven-plugin:1.5.1:scan (oakpal-scan) @ acs-aem-commons-ui.apps --- [INFO] Found a new index node [reference]. Reindexing is requested [INFO] Reindexing will be performed for following indexes: [/oak:index/uuid, /oak:index/reference, /oak:index/nodetype] [INFO] Indexing report - /oak:index/uuid*(0) - /oak:index/reference*(0) - /oak:index/nodetype*(1258) [INFO] Reindexing completed [INFO] Reindexing will be performed for following indexes: [/oak:index/principalName, /oak:index/authorizableId, /oak:index/acPrincipalName, /oak:index/repMembers] [INFO] Indexing report - /oak:index/principalName*(2) - /oak:index/authorizableId*(2) - /oak:index/acPrincipalName*(0) - /oak:index/repMembers*(0) [INFO] Reindexing completed [ERROR] Error during processing of /apps/acs-commons/touchui-widgets/icon-picker: javax.jcr.nodetype.NoSuchNodeTypeException: Node type cq:ClientLibraryFolder does not exist [ERROR] E /apps/acs-commons/touchui-widgets/icon-picker (javax.jcr.nodetype.NoSuchNodeTypeException: Node type cq:ClientLibraryFolder does not exist) That is an issue though in oakpal which should be fixed there, as we shouldn't include any irrelevant metadata in packages (like the cnd for nodetypes which are always existing already). @kwin Sorry for just now seeing your question, as it looks like this issue has already been resolved. There are a couple ways to provide the platform CNDs to oakpal other than maintaining redundant CNDs in the package artifacts. One way is to use the cndNames parameter to reference specific CND resources on the test classpath. Another is to enable resolution by detecting Sling-Nodetypes manifest headers in test-scope dependencies. This enabled by setting slingNodeTypes to true. This option might be more convenient in the future now that Adobe's uber-jars (starting with 6.5.0) include the platform nodetypes. Finally, it is also possible to serialize the nodetypes into the jcrNodetypes json property of plans and checklists, but this method makes it harder to manually maintain those files if we are talking about the full set of aem nodetypes, rather than just nodetypes directly referenced by progress check implementations or forcedRoot primary/mixin types.
2025-04-01T04:10:08.551647
2023-02-01T20:06:03
1566747100
{ "authors": [ "synox", "theManikJindal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13312", "repo": "AdobeDocs/adobe-dev-console", "url": "https://github.com/AdobeDocs/adobe-dev-console/issues/65" }
gharchive/issue
The docs don't mention where to find the api_key Issue in /src/pages/guides/authentication/APIKeyIntegration.md When the API has been successfully connected, you will be able to access the newly generated API Key. But there is no field with the label API key on the console page. Maybe you mean CLIENTID? Yes API key and client id mean the same thing
2025-04-01T04:10:08.553570
2024-05-08T01:02:33
2284469293
{ "authors": [ "Alicesnk", "jtwood426" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13313", "repo": "AdobeDocs/campaign.en", "url": "https://github.com/AdobeDocs/campaign.en/issues/7" }
gharchive/issue
Text Presentation Error Issue in ./help/v8/connect/ac-ajo.md The bottom of the page shows this raw text: [Learn more in Journey Optimizer documentation]https://experienceleague.adobe.com/en/docs/journey-optimizer/using/orchestrate-journeys/about-journey-building/using-adobe-campaign-v7-v8){target=“_blank”}. Thanks for raising this - we'll update shortly Internal reference > DOCAC-10478
2025-04-01T04:10:08.559786
2021-09-07T15:37:07
990113140
{ "authors": [ "3phase", "DaveInMatrix" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13314", "repo": "AdobeDocs/uxp-xd", "url": "https://github.com/AdobeDocs/uxp-xd/issues/109" }
gharchive/issue
Manifest link leads to old manifest version Issue in /src/pages/develop/plugin-development/index.md Hi, I'd like to ask if (or just point out) there's any specific reason why Plugin Development > Plugin Structure > Manifest leads to Manifest v3 link (which would even explicitly point to v3 in the URL), while the latest is v4? This was little obfuscating at least to me to me. Regards, Petko This is deliberate as XD currently only works with the v3 manifest. We are in the process of adding support for v4 features.
2025-04-01T04:10:08.562021
2019-12-31T17:14:38
544233533
{ "authors": [ "ericdrobinson", "pklaschka" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13315", "repo": "AdobeXD/typings", "url": "https://github.com/AdobeXD/typings/issues/72" }
gharchive/issue
Suggestion: Call out custom type name overrides in ReadMe PR #70 changed the Selection type declaration to be XDSelection. It may be helpful for users to have this explicitly called out (with explanation for why) in the ReadMe. Suggestion: Add a new section to the ReadMe file (called something like "Notes" or "Errata") that calls out customizations like this and includes a blurb for their raison d'être. @ericdrobinson The funny thing is: I wanted to implement this tomorrow, anyway. You're too fast at creating issues :stuck_out_tongue_winking_eye:. I'll also (before creating the release) rename the LinearGradient (I've waited for too long for the docs to update this :neutral_face:), cf. #41, and will then include these as "exceptions" in the section where we've specified that the typings follow the docs.
2025-04-01T04:10:08.563489
2018-07-03T14:59:19
337937555
{ "authors": [ "sxa555" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13316", "repo": "AdoptOpenJDK/openjdk-build", "url": "https://github.com/AdoptOpenJDK/openjdk-build/issues/394" }
gharchive/issue
Add arm32 into the nightly openjdk10 pipelines Now that https://github.com/AdoptOpenJDK/openjdk-build/issues/332 is no longer a problem, we can add these back into the nightly pipelines :-) Added in https://github.com/AdoptOpenJDK/openjdk-build/commit/03128d78f590d09c9fa2bb8ac8146cff9df3d60c
2025-04-01T04:10:08.567401
2018-04-11T18:24:40
313436379
{ "authors": [ "karianna", "vsebe" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13317", "repo": "AdoptOpenJDK/openjdk-infrastructure", "url": "https://github.com/AdoptOpenJDK/openjdk-infrastructure/pull/340" }
gharchive/pull-request
Add xz compression to CentOS build tools Signed-off-by: Violeta Sebe<EMAIL_ADDRESS> What's this required for? It is required to unpack git source: git-2.15.0.tar.xz (in Extract git source task.
2025-04-01T04:10:08.575649
2021-04-17T08:11:03
860355019
{ "authors": [ "Gaminguide1000", "Silfryi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13318", "repo": "Advanced-Rocketry/AdvancedRocketry", "url": "https://github.com/Advanced-Rocketry/AdvancedRocketry/issues/2062" }
gharchive/issue
Advanced Rocketry planet data save bug Version of Advanced Rocketry AdvancedRocketry-1.12.2-2.0.0-242-universal Have you verified this is an issue in the latest unstable build [x] Y Version of LibVulpes LibVulpes-1.12.2-0.4.2-83-universal Version of Minecraft 1.12.2 Does this occur without other mods installed [x] Y If Y, what is the MINIMUM set of mods required. Idk, but i think only AdvRocketry&LibVulpes Crash report and log, or VisualVM snapshot (if applicable) http://pastebin.com is a good place to put them crash reports that are put in the issue itself are impossible to read If it's an issue involving lag, please submit a VisualVM sampler snapshot. Mods like lag goggles cannot provide all of the information needed to debug lag in any useful capacity. For more information on how to create a snapshot, please see "How to Run CPU Sampling" here. Then export the snapshot as an .nps. You may have to zip it since github is weird, then please upload it here. Description of the problem First of all: I have a custom PlanetDefs.xml (not mine). So, always after a while, the server console says: "[Server thread / FATAL] [advancedrocketry]: An error has occurred saving planet data, this can happen if another mod causes the game to crash during game load. If the game has fully loaded, then this is a serious error, Advanced Rocketry data has not been saved. "There are no conflicts between different mods, nor does the game / server crash when I start it. should that be a "serious error". Yes, because it means your planetDefs has a high chance of becoming so corrupt you cannot load your world.
2025-04-01T04:10:08.580412
2023-09-13T11:30:45
1894330245
{ "authors": [ "StefanGreve" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13319", "repo": "Advanced-Systems/anonpy", "url": "https://github.com/Advanced-Systems/anonpy/issues/33" }
gharchive/issue
Improve Publishing Workflow for Development Discussed in https://github.com/Advanced-Systems/anonpy/discussions/4 Originally posted by StefanGreve August 9, 2023 Preamble Editing the workflow file responsible for the publishing process in a live environment can be a little bit disheartening if there are errors, to put it mildly. Luckily for us, PyPI provides the facilities to test builds on https://test.pypi.org/. Documentation about that can be found here: https://packaging.python.org/en/latest/guides/using-testpypi/ Requirements Essentially, we need a workflow file that can be manually triggered with the workflow_dispatch event on GitHub actions which requires that the committer updates the __version__ variable in src/anonfile/anonfile.py to major.minor.patch-buildnumber. This makes it much easier to test the release workflow during development, and avoids scenarios in which we have to publish a patch because something went wrong with python-publish.yml. The secrets for test.pypi.org need to be configured by the project administrator (@StefanGreve) through the repository settings panel. I would propose calling them TEST_PYPI_USERNAME and TEST_PYPI_PASSWORD, respectively. Planning Status Scheduled for milestone 1.0.0. Closed as duplicate, see also #30
2025-04-01T04:10:08.612119
2016-07-24T09:50:49
167226012
{ "authors": [ "Grover-c13", "Spl3en", "rubenvereecken" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13320", "repo": "AeonLucid/POGOProtos", "url": "https://github.com/AeonLucid/POGOProtos/pull/52" }
gharchive/pull-request
2025-04-01T04:10:08.614814
2024-08-16T22:12:43
2471108036
{ "authors": [ "593Coding", "Aeonss" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13321", "repo": "Aeonss/BubbleBlaster", "url": "https://github.com/Aeonss/BubbleBlaster/issues/35" }
gharchive/issue
I've this error I ran pip install -r requirements.txt and python bubbleblaster.py Try the newest 2.0.1 version first, see if that works. Run the "INSTALL_DEPENDENCIES.bat" first and then the "bubbleblaster.bat" I included a exe version... try if it doesn't work I included a exe version... try if it doesn't work I found the problem, you need to put this dll on system32 in windows11 https://www.dllme.com/dll/files/libomp140_x86_64/00637fe34a6043031c9ae4c6cf0a891d/download
2025-04-01T04:10:08.619359
2016-11-16T03:11:00
189571066
{ "authors": [ "boisy", "mitchmindtree" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13322", "repo": "Affectiva/affdexme-osx", "url": "https://github.com/Affectiva/affdexme-osx/issues/11" }
gharchive/issue
How to modify example to detect ethnicity? Hey there, after a quick look I couldn't find any documentation on how to set up ethnicity detection - would you mind providing any tips on how to modify the example to do this? Or maybe provide a link to some docs that demonstrate this? Thanks! Hi Mitch, We're not exposing ethnicity just yet in our SDK. It's still undergoing validation, but we expect to have it in a future release. @boisy legend, thanks a lot for this :+1: I ended up getting this working and the results were generally (roughly) correct 😸
2025-04-01T04:10:08.626991
2022-11-14T04:03:21
1447343156
{ "authors": [ "AgentRev", "manav0619" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13323", "repo": "AgentRev/WindowsAppsUnfukker", "url": "https://github.com/AgentRev/WindowsAppsUnfukker/issues/13" }
gharchive/issue
PsExec could not start powershell Hello, I am running W11 22H2. I have been trying to fix my WindowsApps folder that I messed up permissions for and came across your script. However, it's not working for some reason on my PC. I keep getting the "PsExec could not start powershell…" error. I was wondering if you could assist me a bit. Any help will be truly appreciated. Windows PowerShell Copyright (C) Microsoft Corporation. All rights reserved. Install the latest PowerShell for new features and improvements! https://aka.ms/PSWindows PS C:\Users\manav> cd J:\Users\manav\Downloads\PSTools PS J:\Users\manav\Downloads\PSTools> .\psexec.exe -s -i powershell -ExecutionPolicy Bypass -File "J:\Users\manav\Downloads\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1" PsExec v2.4 - Execute processes remotely Copyright (C) 2001-2022 Mark Russinovich Sysinternals - www.sysinternals.com PsExec could not start powershell on DESKTOP-99N6EMR: The system cannot find the file specified. It means Powershell cannot access the path to the script. In my experience, the usual culprit is network drives. I would say, try to move the script to a different drive. @AgentRev Hey, thanks for responding. I don't actually have any network drives. Just two SSDs and two HDDs, all local. I have tried moving the script to different locations on every drive, but it's the exact same error every time. If you'd like to see any log file or system information that might help you identify the issue, I'd be happy to provide it. I really have no idea what's going wrong. At that point, it's beyond my reach. You could use Process Monitor from PsTools to investigate where is PsExec trying to look and why it's not resolving it. Other people have also recommended PAExec instead, maybe give that a try. It wasn't PowerShell being unable to access the path to the script; it was rather PsExec being unable to access PowerShell. I just added the full path to powershell.exe in the command, and it ended up working. So, instead of .\psexec.exe -s -i powershell -ExecutionPolicy Bypass -File "C:\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1 I used .\psexec.exe -s -i "C:\Windows\System32\WindowsPowerShell\v1.0\powershell.exe" -ExecutionPolicy Bypass -File "C:\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1 (I guess my environment variables are messed up as well? Because I had to add the System32 folder to "Path" under User Variables for takeown command to work when I was manually trying to fix the permissions earlier.)
2025-04-01T04:10:08.628607
2023-08-03T21:04:05
1835733238
{ "authors": [ "albnunes", "mmabrouk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13324", "repo": "Agenta-AI/agenta", "url": "https://github.com/Agenta-AI/agenta/pull/224" }
gharchive/pull-request
Save test set update In this pull request, I removed the requested functionality mentioned in issue #220, and everything seems to be working. I hope to contribute even more effectively to this project in the future. Closes #220 Thanks for the PR @albnunes. The issue has been closed in #265 @all-contributors please add @albnunes for bug fix
2025-04-01T04:10:08.702846
2021-08-05T06:16:25
961484352
{ "authors": [ "paulmorar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13325", "repo": "Aguafrommars/azure-devops-extensions", "url": "https://github.com/Aguafrommars/azure-devops-extensions/pull/35" }
gharchive/pull-request
Fix typo in commit types Currently, there are quite a few typos in the commit type entries, and this PR addresses that. @aguacongas can you take a look over this one?
2025-04-01T04:10:08.707540
2024-03-29T08:36:18
2214928290
{ "authors": [ "AgustinSRG", "derekcmini" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13326", "repo": "AgustinSRG/crystals-dilithium-js", "url": "https://github.com/AgustinSRG/crystals-dilithium-js/issues/1" }
gharchive/issue
After the web page loads dilithium.js, an error occurs when executing DilithiumKeyPair.generate(2) Hi, I use in my html code, and Part of the js code in my frontend: const level = DilithiumAlgorithm.DilithiumLevel.get(2); const keyPair = DilithiumAlgorithm.DilithiumKeyPair.generate(level); Post-processing when web page is reloaded ReferenceError: Buffer is not defined at s (dilithium.js:2:63339) at l.generate (dilithium.js:2:42255) Am I doing anything wrong? Hello. Thanks for reporting this issue. It seems Webpack does not include a Buffer polyfill into the javascript library bundle by default. I changed the webpack configuration to include it. Now it should be working without any extra dependencies in the browser (https://github.com/AgustinSRG/crystals-dilithium-js/releases/tag/v1.1.1). I also added a basic test page (https://agustinsrg.github.io/crystals-dilithium-js/test.html) to test the browser bundle. I'm marking the issue as completed. In case you find any other issues with this library, feel free to open more issues in the future.
2025-04-01T04:10:08.721542
2015-08-28T08:34:33
103671445
{ "authors": [ "sbuffo" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13327", "repo": "AirVantage/av-sched", "url": "https://github.com/AirVantage/av-sched/issues/25" }
gharchive/issue
Improve configuration properties For prefomance reason : change the pools size configuration, add them as deploy parameters. PR : https://github.com/AirVantage/av-sched/pull/26
2025-04-01T04:10:08.723052
2021-09-15T01:27:10
996567892
{ "authors": [ "GH3336699", "Satoshiisgod" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13328", "repo": "Aircoin-official/AirCash", "url": "https://github.com/Aircoin-official/AirCash/issues/73" }
gharchive/issue
Apply for Air cash test Hello, I want to apply for the air coin test. I used to be the general manager of IT projects. I started to understand the blockchain in 2016 and later became a merchant of other exchanges. I still understand this aspect quite well. I am very interested in Air Cash, and I hereby apply, hoping to become a member of Air Cash and contribute to Air Cash. Anyone who applied for the AirCash testing please join the discord channel for test volunteers. https://discord.gg/eRSY7NtxZA
2025-04-01T04:10:08.728032
2021-04-18T00:22:29
860542121
{ "authors": [ "Aircoookie", "Chris210418", "LeonardoD91" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13329", "repo": "Aircoookie/Espalexa", "url": "https://github.com/Aircoookie/Espalexa/issues/175" }
gharchive/issue
Ghost Devices Alexa keeps discovering a few devices that i only used once for testing. Deleting them in the Alexa App doesnt work half the time and they prevent the discovery of the last device in my final Program. At least thats what i think happens. The weird thing is they are discovered even if my esp32 is turned off. Maybe thats more of an Alexa problem but i dont know how to fix it. I really enjoy the stuff that works. Thanks for making this. Help would be appreciated. Hi, I can definitely reproduce this issue. Originally I believed this issue only affected my account but it appears to be a general problem when using Espalexa. Devices always get "discovered" after deleting, even if said device has not been online for a year. Deleting all devices/groups/schedules from the alexa app did not help either. I have not yet been able to conclusively determine if the ghost devices are saved in the Echo hardware or in the Amazon account and am sadly unable to share any method that permanently deletes devices. One slightly random question - do you have a genuine Philips Hue bridge in your network and/or installed the associated Alexa skill? Hello, thanks for answering. I have been able to set up the last device in my Program. It was an error in my code. Yes i have a hue bridge in my home and the hue skill installed. But having it connected and or the skill installed seems to have no impact on the non existing yet discovered devices. A few test have shown that the "Ghost Devices" get rediscovered only if the right Alexa Device is connected to the network. I think its the one that was closest to the esp when i first testet the program. All of the rest is working brilliantly quick and reliable. Thanks again for making this and thanks for helping a lot of not so code talented people like me! :P I have the same issue, how do i stop this ghost devices from popping, i tried to erase flash the esp form esptool in terminal. I flashed many other project on the same esp chip but i always have this devices appearing, maybe alexa remind the mac address of the device and keep associate it to the old device names, i tried also to factory resety the alexa echo dot where those ghost devices connected forst time, but nothing changed. Myabe i have to reset the all alexa devices in my house? I appreciate any help
2025-04-01T04:10:08.792090
2023-04-25T08:45:44
1682716012
{ "authors": [ "nclsHart", "notFloran" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13331", "repo": "Akollade/mac-playbook", "url": "https://github.com/Akollade/mac-playbook/pull/184" }
gharchive/pull-request
:bug: Fix issue between homebrew/sudo and ansible become Fix #181 Related to #179 : the fix I made broke the password pass between ansible and sudo. I've got an issue which seems to be related to brew run as sudo (during dev setup tags=nginx). fatal: [<IP_ADDRESS>]: FAILED! => changed=true cmd: - brew - services - restart - nginx delta: '0:00:00.830175' end: '2023-04-25 14:06:08.832309' msg: non-zero return code rc: 1 start: '2023-04-25 14:06:08.002134' stderr: 'Error: Need to download https://formulae.brew.sh/api/formula.jws.json but cannot as root! Run `brew update` without `sudo` first then try again.' stderr_lines: <omitted> stdout: '' stdout_lines: <omitted> I tried to run brew update and run again the setup command but without success. I've got an issue which seems to be related to brew run as sudo (during dev setup tags=nginx). fatal: [<IP_ADDRESS>]: FAILED! => changed=true cmd: - brew - services - restart - nginx delta: '0:00:00.830175' end: '2023-04-25 14:06:08.832309' msg: non-zero return code rc: 1 start: '2023-04-25 14:06:08.002134' stderr: 'Error: Need to download https://formulae.brew.sh/api/formula.jws.json but cannot as root! Run `brew update` without `sudo` first then try again.' stderr_lines: <omitted> stdout: '' stdout_lines: <omitted> I tried to run brew update and run again the setup command but without success. Can you run dev dotiles ? Can you run dev dotiles ? Done. But no changes :/ It works... I just forget to reload my zsh config.
2025-04-01T04:10:08.810921
2020-02-06T10:05:23
560904531
{ "authors": [ "alexander-mai", "farfromrefug" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13332", "repo": "Akylas/nativescript-material-components", "url": "https://github.com/Akylas/nativescript-material-components/pull/110" }
gharchive/pull-request
feat: added features for bottomsheet dismissOnDraggingDownSheet added example on demo @alexander-mai thanks ! will try to test it today I could have created examples for Angular and Vue, but I can't get the demo apps up and running right away. @farfromrefug What do you think, when this brach can be merged? I made some improvements on top of this change and I would like the create a new pull request for it. @alexander-mai sorry i actually forgot to merge this :s Thanks a lot for your work! Will make a release soon
2025-04-01T04:10:08.965649
2020-10-23T18:54:55
728449189
{ "authors": [ "blackfalcon", "ecommerceUXdesign" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13333", "repo": "AlaskaAirlines/Icons", "url": "https://github.com/AlaskaAirlines/Icons/pull/46" }
gharchive/pull-request
Add new icons Alaska Airlines Pull Request Add new icons boadring.svg elite.svg Type of change: Please delete options that are not relevant. [ ] New capability [ ] Revision of an existing capability [ ] Infrastructure change (automation, etc.) [x] Other (please elaborate) Checklist: [x] My update follows the CONTRIBUTING guidelines of this project [x] I have performed a self-review of my own update By submitting this Pull Request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Pull Requests will be evaluated by their quality of update and whether it is consistent with the goals and values of this project. Any submission is to be considered a conversation between the submitter and the maintainers of this project and may require changes to your submission. Thank you for your submission! -- Orion Design System Team :tada: This PR is included in version 4.0.2 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T04:10:08.971923
2020-10-12T18:18:21
719570443
{ "authors": [ "blackfalcon", "ecommerceUXdesign" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13334", "repo": "AlaskaAirlines/auro-icon", "url": "https://github.com/AlaskaAirlines/auro-icon/pull/10" }
gharchive/pull-request
BREAKING CHANGE: support for Alaska logos by prop Alaska Airlines Pull Request BREAKING CHANGE: This update requires a breaking change as the API for applying custom color and size has been changed. Fixes: https://github.com/AlaskaAirlines/auro-icon/issues/7 Summary: This update address the need submitted by #7, as well as address additional concerns about custom color and size management. See http://auro-21.surge.sh/ for demo Type of change: Please delete options that are not relevant. [x] New capability [x] Revision of an existing capability [ ] Infrastructure change (automation, etc.) [ ] Other (please elaborate) Checklist: [x] My update follows the CONTRIBUTING guidelines of this project [x] I have performed a self-review of my own update By submitting this Pull Request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Pull Requests will be evaluated by their quality of update and whether it is consistent with the goals and values of this project. Any submission is to be considered a conversation between the submitter and the maintainers of this project and may require changes to your submission. Thank you for your submission! -- Auro Design System Team :tada: This PR is included in version 2.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T04:10:08.998889
2023-04-27T09:20:24
1686461012
{ "authors": [ "AleMorales", "jianchaoci" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13335", "repo": "AleMorales/VPL", "url": "https://github.com/AleMorales/VPL/issues/4" }
gharchive/issue
VPL can be precompiled? Hello, Thanks for your great work! I am trying to precompile the VPL package but got this error. I checked the folder of "VPL/SS2dp/src/Core", and I realized the "Algorithms.jl" is missed, but it has a file called "algorithms.jl". I am not really sure if it is a bug, and I am not really sure how to fix it, as I am really new to Julia. Can you try to fix it? I will really appreciate it. =========================================== Precompiling project... ✗ VPL 0 dependencies successfully precompiled in 38 seconds. 267 already precompiled. The following 1 direct dependency failed to precompile: VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] Failed to precompile VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] to /home/jianchao/.julia/compiled/v1.8/VPL/jl_EqEJDY. ERROR: LoadError: SystemError: opening file "/home/jianchao/.julia/packages/VPL/SS2dp/src/Core/Algorithms.jl": No such file or directory Hello, Thanks for your great work! I am trying to precompile the VPL package but got this error. I checked the folder of "VPL/SS2dp/src/Core", and I realized the "Algorithms.jl" is missed, but it has a file called "algorithms.jl". I am not really sure if it is a bug, and I am not really sure how to fix it, as I am really new to Julia. Can you try to fix it? I will really appreciate it. =========================================== Precompiling project... ✗ VPL 0 dependencies successfully precompiled in 38 seconds. 267 already precompiled. The following 1 direct dependency failed to precompile: VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] Failed to precompile VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] to /home/jianchao/.julia/compiled/v1.8/VPL/jl_EqEJDY. ERROR: LoadError: SystemError: opening file "/home/jianchao/.julia/packages/VPL/SS2dp/src/Core/Algorithms.jl": No such file or directory I think I just solved this problem, I am sure it is the problem of defining the folder/file name. some of them are wrong in using the small or capital letter. So I assume this issue is solved? From your description it looked that something went wrong with the Julia Pkg manager (sometimes it happens). Please confirm so that I can close the issue :)
2025-04-01T04:10:09.011355
2016-04-15T01:39:56
148530134
{ "authors": [ "AlecAivazis" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13336", "repo": "AlecAivazis/nautilus", "url": "https://github.com/AlecAivazis/nautilus/pull/81" }
gharchive/pull-request
Documentation updates This PR fixes a few of the documentation updates - resolving #76, #75, and #59. @NickDubelman you should look at this and check it against your list. There are probably still changes that I missed (this is just some of the more complicated things like the module index).
2025-04-01T04:10:09.028402
2023-04-29T02:58:15
1689332955
{ "authors": [ "58bits", "AlessioGr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13337", "repo": "AlessioGr/payload-plugin-lexical", "url": "https://github.com/AlessioGr/payload-plugin-lexical/issues/25" }
gharchive/issue
APIError: The collection with slug media can't be found Was looking at the demo app in this repo, and when trying to save a post with an image / upload in the post in the LexicalMinimal collection I receive the following error: [02:55:34] ERROR (payload): APIError: The collection with slug media can't be found. at new ExtendableError (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/errors/APIError.ts:26:11) at new APIError (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/errors/APIError.ts:43:5) at findByIDLocal (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/collections/operations/local/findByID.ts:48:11) at Payload.findByID (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/payload.ts:255:12) at loadUploadData (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:48:24) at traverseLexicalField (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:80:30) at traverseLexicalField (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:108:30) at populateLexicalRelationships (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:35:32) at processTicksAndRejections (node:internal/process/task_queues:95:5) at async /home/tony/Projects/payload/payload-plugin-lexical-fork/demo/node_modules/payload/src/fields/hooks/afterRead/promise.ts:173:31 Hmmmm - you're totally right, I could reproduce that. I'm pretty sure this issue only happens in the demo, not when you actually install the plugin. Prob some issue with the way I set up the dependencies / project structure. Will have a look! Should be fixed now!
2025-04-01T04:10:09.041920
2021-01-29T12:54:53
796859630
{ "authors": [ "Alex-D", "PiemP", "agriffard", "momentum-tunis", "ned-kelly" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13338", "repo": "Alex-D/Trumbowyg", "url": "https://github.com/Alex-D/Trumbowyg/pull/1192" }
gharchive/pull-request
plugin resizimg: updated plugin to fit jquery-resizable 0.35 Fixes #1191 updated the plugin to use the resizableSafe function instead of the resizable to avoid conflict with jquery UI. removed file resizable-resolveconflicts.js. updated the documentation: specified the jquery-resizable version's used. @Alex-D Would you have some time to review this PR? @Alex-D Can you please have a look at this PR? Thank you. @Alex-D Can you please find some time to review this PR? It would help us if it was merged and a new version was available. Thank you in advance. No news on this PR? Thank you :) Thank you very much @Alex-D 👍 @agriffard now we are waiting for orchard core to integrate the next release of Trumbowyg Hi @Alex-D - any chance we can see this one added into a release?
2025-04-01T04:10:09.044859
2021-06-12T12:25:49
919545465
{ "authors": [ "AlanFloyd", "alzubitariq" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13339", "repo": "Alex-Dobrynin/Xamarin.Controls.ImageCropper", "url": "https://github.com/Alex-Dobrynin/Xamarin.Controls.ImageCropper/issues/3" }
gharchive/issue
Sometimes stops working on Android I have seen that after a couple of times, the controls stops working on Android. Using the debugger, the last statement executed is the "await ImageCropper.Current.Crop(new CropSettings()", but there is no reply to IsFaulted, IsCanceled or IsCompleted. Here the code: Device.BeginInvokeOnMainThread(async () => { await ImageCropper.Current.Crop(new CropSettings() { AspectRatioX = 1, AspectRatioY = 1, CropShape = CropSettings.CropShapeType.Rectangle }, PhotoPath).ContinueWith(t => { if (t.IsFaulted) { var ex = t.Exception; //alert user } else if (t.IsCanceled) { var ex = t.Exception; //do nothing } else if (t.IsCompleted) { var result = t.Result; Device.BeginInvokeOnMainThread(() => { logo_image.Source = result; is_saved = false; photo_path = result; lbl_initials.IsVisible = false; logo_image.IsVisible = true; }); } }); }); Any idea what could be wrong in my code? Dietmar Hi AlanFloyd, If you are using MediaPicker, i think the workaround is to put some delay between picking a picture and the ImageCropper like this var file = await MediaPicker.PickPhotoAsync(); await Task.Delay(2000); await ImageCropper.Current.Crop(new CropSettings() etc ..... I think the issue who is calling the OnActivityResult firstly with resultCode
2025-04-01T04:10:09.056409
2024-01-16T13:46:41
2084001390
{ "authors": [ "gz83", "rollsicecream", "trimechee" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13340", "repo": "Alex313031/thorium", "url": "https://github.com/Alex313031/thorium/issues/510" }
gharchive/issue
strange github bug Hello, I make new issues with my other new account, but i don't see my issues when i use this account !!! @github ?!!! why ? @Alex313031 @gz83 just to check, have you seen my issues made by my other account? https://github.com/Alex313031/Mercury/issues/101 https://github.com/Alex313031/thorium/issues/506 https://github.com/Alex313031/thorium/issues/507 It seems that these issues do not exist @gz83 it's a serious and regrettable bug of @github, it's wasting my time and fortunately I noticed it, my issue exists but only me I can see my 3 issues with my other account :( I'm going to have to copy and paste this that I wrote and open new issues with another account :( Maybe your account got shadowbanned or something like that? I use a web browser with a lot of modified flags and added command line and special dns....github sent me an email to my censored account telling me github cannot determine your geographic location and suspects potential hacking... .yet I don't use VPN....disgusted and angry I went to the settings of my new account to delete it and but the delete account button is grayed out and github tells me your account has been reported so I have to contact github support and I have to enable 2-factor authentication, so I gave up....I still have this account to be able to contact the developers and try to post exclusive ideas to improve the software....anyway Thank you @rollsicecream for caring about my serious problem :) Also, I had this issue before. It was because I used an alias e-mail which is "suspicious" for GitHub..... Ohh very important information ! Thank you so much Mr @rollsicecream for your great advise, i will try to change my alias !
2025-04-01T04:10:09.074471
2021-09-15T22:10:59
997561860
{ "authors": [ "AlexDarigan", "olestourko" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13341", "repo": "AlexDarigan/WAT", "url": "https://github.com/AlexDarigan/WAT/issues/295" }
gharchive/issue
Modifying a test class or method causes the test runner to fail Godot 3.3.3 stable C# (Mono) WAT Version 6 When I run tests through the "Test" tabs after modifying a test class' name, a test method name, or ad a new test method, the test runner fails. I'm writing C# tests and using "Build" in Godot after my changes. This only appears to be happening if I had run tests before making changes - maybe WAT is caching class/method definitions? It appears that the new class/method names aren't picked up in the test runner's dropdowns. There isn't anything in the Output tab and no dropdowns under the failure in the Test tab. If I reload the project the new tests get picked up and it works again. Toggling the plugin in projects settings works too. Might be a cache problem alright. Are you using an external editor? Maybe the filesystem dock signals that WAT uses aren't being sent. I got around it by skipping the file system check here: https://github.com/AlexDarigan/WAT/blob/3a6e746523b658991e83289f56bd34afc196ee67/addons/WAT/ui/gui.gd#L64 Maybe there's a way of detecting changes to C#-based scripts? Just picked up Godot so very fresh to this all Are you using an external editor? Maybe the filesystem dock signals that WAT uses aren't being sent. Yes, VSCode on Ubuntu 20.04 If you save the file, a "resource_saved" signal will be emitted which will trigger an update of the system. This might be a VSCode issue. Worth trying to see if the changes get picked up when you change and save from the editor. Yup, WAT detects change made through the builtin editor Oh I have an idea. The filedock signals will probably trigger from an external editor but this resource_saved one won't because it is a plugin signal and that's the one we care about. I think having the system auto-update might be the best for now until I can figure this out. Maybe a launch.json for WAT specifically. I'm very new to Godot so I'm not sure what all of that means :sweat_smile: If its useful at all, the builtin editor is picking up changes I make externally in VSCode. I actually did try setting up a launch.json for WAT but the current version of Godot is ignore the args passed to it, so it wouldn't take the WAT cli args: https://github.com/godotengine/godot-csharp-vscode/issues/19 In the meantime just keeping my hack to skip the _filesystem.change check is enough for me, personally. If this ends up being a difficult thing to figure out, perhaps a setting on the extension to optionally skip the check for external editor users could be a compromise? A reminder to myself: Could experiment different files saved signals from other sources with a check that breaks if filesystem was updated. SriptEditor can receive signals about changes to active scripts, but I can't find a way to receive signals about inactive ones. Here's a simple draft PR for a setting to disable caching: https://github.com/AlexDarigan/WAT/pull/296 Check if this works c6bb81ab803e333a11352ee207ba904ed72b0030 Seems this issue has been resolved.
2025-04-01T04:10:09.077473
2022-12-28T21:20:56
1513177211
{ "authors": [ "AlexGladkov", "masterofdaemon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13342", "repo": "AlexGladkov/Odyssey", "url": "https://github.com/AlexGladkov/Odyssey/issues/71" }
gharchive/issue
how can i use in with hilt? and this methods are undefined: setupNavigation("splash") { splashScreen() } You can try 1.3.0-beta04 version and add odyssey-android in dependencies or wait stable 1.3.0 release You can read documentation about this here - https://github.com/AlexGladkov/Odyssey/blob/rc/release-1.3.0/documentation/HILT.md And it works only in 1.3+ versions
2025-04-01T04:10:09.120807
2023-08-20T18:06:55
1858217644
{ "authors": [ "TAforever", "waddyhussain" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13344", "repo": "Alexays/Waybar", "url": "https://github.com/Alexays/Waybar/issues/2425" }
gharchive/issue
hyprland/language add on-click config options Add support for on-click events (on-scroll events would be helpful too) to support customisations such as switching between different keyboard layouts Hi, this functionality has already been added. Here is an example of my configuration "hyprland/language": { "format-en": "us ", "format-ru": "ru ", "keyboard-name": "gaming-keyboard", "on-click": "hyprctl switchxkblayout gaming-keyboard next" }, Thanks, marking as closed
2025-04-01T04:10:09.122378
2019-05-17T09:17:18
445344826
{ "authors": [ "Alexays", "lockywolf", "unresolvedsymbol" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13345", "repo": "Alexays/Waybar", "url": "https://github.com/Alexays/Waybar/issues/324" }
gharchive/issue
Battery module rewrite Using udev Will closes #229 #237 This should be an added module/mode because not everyone has upower At the moment, if a battery is not present, the only thing the module does is complain in the log: [warning] module battery: Disabling module "battery", No batteries. May I suggest that that module present itself in a deliberately "broken" state, rather than silently failing, in order to urge the user to resolve the issue?
2025-04-01T04:10:09.123263
2022-01-14T18:37:11
1103996398
{ "authors": [ "Alexays", "nullobsi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13346", "repo": "Alexays/Waybar", "url": "https://github.com/Alexays/Waybar/pull/1394" }
gharchive/pull-request
fix: Use locale when formatting clock Fixes #1393 When formatting the clock, pass the locale to the fmt function. Thanks!
2025-04-01T04:10:09.178393
2023-08-30T14:11:19
1873769024
{ "authors": [ "AlexeyBoiko", "yogananda-muthaiah" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13347", "repo": "AlexeyBoiko/DgrmJS", "url": "https://github.com/AlexeyBoiko/DgrmJS/issues/35" }
gharchive/issue
Can Arrows be made to animated flow as new feature? Hi Alex, Can you think of adding animated flow to arrows.. so it gives a better flow visualization ?? Just a thought Hi @yogananda-muthaiah Please give an example. Closed due inactivity
2025-04-01T04:10:09.183127
2022-12-12T19:15:46
1492573993
{ "authors": [ "AlexisJW", "PinkMoon25" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13348", "repo": "AlexisJW/set-up-myFirst-mobile", "url": "https://github.com/AlexisJW/set-up-myFirst-mobile/issues/9" }
gharchive/issue
Portfolio: application deploy Deploy my portfolio using GitHub Pages. Check the online version of my portfolio and make sure that the page works properly. Update the README of my repository to include a link to the online version. The link to the online version of my application Hi @AlexisJW , Your project is complete! There is nothing else to say other than... it's time to merge it :shipit: Congratulations! 🎉 Highlights linter checks :heavy_check_mark: professional readme :heavy_check_mark: Added live demo link to reade :heavy_check_mark: deployed with github and live demo working :heavy_check_mark: Optional suggestions Every comment with the [OPTIONAL] prefix won't stop the approval of this PR. However, I strongly recommend you to take them into account as they can make your code better. Some of them were simply missed by the previous reviewer and addressing them will really improve your application. Cheers and Happy coding!👏👏👏 Feel free to leave any questions or comments in the PR thread if something is not 100% clear. Please, remember to tag me in your question so I can receive the notification. As described in the Code reviews limits policy you have a limited number of reviews per project (check the exact number in your Dashboard). If you think that the code review was not fair, you can request a second opinion using this form.
2025-04-01T04:10:09.203640
2024-11-20T19:44:15
2676997516
{ "authors": [ "dvenprasad", "jaclyn-taroni" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13349", "repo": "AlexsLemonade/scpca-docs", "url": "https://github.com/AlexsLemonade/scpca-docs/pull/373" }
gharchive/pull-request
Update bulk names for download illustrations Closes #369 Updated the bulk file names Bumped up the quality of the images too These are purposely 2x the width that they get presented as. Okay resized them. They were all consistently 1200px width - so I resized it to be that.
2025-04-01T04:10:09.205683
2022-07-01T18:58:37
1291739067
{ "authors": [ "jashapiro" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13350", "repo": "AlexsLemonade/scpca-docs", "url": "https://github.com/AlexsLemonade/scpca-docs/pull/87" }
gharchive/pull-request
Add information about multiplex samples to download section Here I am adding a section about the multiplex downloads to the download documentation. I tried to be fairly brief, but I think this reflects the plan we agreed upon. Let me know if there is more detail that you think is appropriate in this section, or additional links that might make sense. I am leaving it as a draft for now, as I think we still need illustrations, as per https://github.com/AlexsLemonade/scpca-docs/issues/75#issuecomment-1162082576 This PR replaces #82, so that what I write can be reviewed by @allyhawkins. Closes #75 I have also added/updated image files so they are all the same resolution. I wonder if they are a bit big on rendering, so I could add a width parameter (changing to html) if we think that is a good idea.
2025-04-01T04:10:09.209854
2020-11-20T17:13:10
747658186
{ "authors": [ "AlexxIT", "Violo1975", "nathanmay" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13351", "repo": "AlexxIT/SonoffLAN", "url": "https://github.com/AlexxIT/SonoffLAN/issues/287" }
gharchive/issue
Nothing happens :( Hi Everybody, I'm a newby... I have follow all the step using HACS and my Yaml file is: # SonOff sonoff: username<EMAIL_ADDRESS> password: xxx force_update: True scan_interval: '00:01:00' # (optional) default 5 minutes mode: auto reload: always # update device list every time HA starts but nothing happen, someone can help me.....many tks for info I've to add two sensors: 4CH Pro R2 2CH with RF 433 MHz System info: Versione HS: 0.118.1 Version OS: 4.19.127-v7 Grazie a tutti :) up.... please :( Nothing happens... What you expect to happen? Hi, It doesn't give me errors but I can't see my devices Could be linked to #294 Devices don't supported yet.
2025-04-01T04:10:09.217054
2021-04-19T11:10:35
861194303
{ "authors": [ "AlexxIT", "weitheng" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13352", "repo": "AlexxIT/WebRTC", "url": "https://github.com/AlexxIT/WebRTC/issues/22" }
gharchive/issue
Stuck on "Init Connection" Safari Desktop Updated to latest master version which does not require specifying ports, streams are stuck on init connection when using Safari Desktop. Tried specifying ports from 50000-59999, but did not work. Working well with other browsers (Chrome, Firefox) on the same laptop even when not specifying ports Try clear cache Cleared cache and tried on incognito, both didn't work. Went onto JS console, got this message Blocked *URL* from asking for credentials because it is a cross-origin request. Looks like url has been blocked This is not about component. Because it doesn't do any requests. I think this is browser cache problem. This is not about component. Because it doesn't do any requests. Could the request be from my RTSP stream? Because the RTSP url contains the username and password to access the stream. I have tried clearing the cache and gone onto incognito mode again, but with no luck. :( I also tried disabling the Cross-Origin Restrictions on Safari, but I am immediately redirected to the IP address of the camera upon refreshing the page, prompting me to enter the username and password (even though they are already included in the RTSP url, and works fine on my other devices). But after entering the correct credentials, I am redirected back to my HA domain, with the following message Failed to load resource: the server responded with a status of 401 (Unauthorized) on the JS console. You can use camera entity instead of url in latest version. For example, if the camera is configured as Generic. You can use camera entity instead of url in latest version. For example, if the camera is configured as Generic. Tried that, still getting the same error on JS console. :( Really unsure what's causing the issue, I kept refreshing the page and sometimes I managed to get it working temporarily, but another refresh will bring it down again. I managed to have a try on my other MacBook, and it works flawlessly on Safari. You could be right that it's the browser problem specifically the cache or website date. But I did clear both of them. This error not from component. You can remove card from page and check. This error not from component. You can remove card from page and check. I hope you didn't set up the card poster? Nope, no card poster. I created a new page and try adding the entity one by one, it seems that it works fine when there's only 1 or 2 WebRTC entities on the page. Once there's more than 2, it struggles to initialise connection. I will keep on trying to see what's the issue, it should be a problem with my browser. Thanks for your help! I have tested on Safari Version 14.0 (156<IP_ADDRESS>.9, 15610). What version do you have? I am using Version 14.0.3 Somehow the streams are working again now, I will continue to monitor the issue. Thanks! Reopen if there's a problem
2025-04-01T04:10:09.243472
2022-11-14T16:00:06
1448326570
{ "authors": [ "Alhajideen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13353", "repo": "Alhajideen/Space--Travelers", "url": "https://github.com/Alhajideen/Space--Travelers/issues/18" }
gharchive/issue
Team3 [1pt] Create basic structure for Dragons - Setup Create a route and a view component. Use <NavLink /> for the page navigation links and style active class to indicate which section/page the user is currently on (underline active navigation link). Create a directory for all Redux state slice files. N/A
2025-04-01T04:10:09.245713
2023-11-07T16:06:23
1981712252
{ "authors": [ "Ali-Muhsin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13354", "repo": "Ali-Muhsin/Whispers", "url": "https://github.com/Ali-Muhsin/Whispers/issues/2" }
gharchive/issue
Read and write into jsons We check the item name, and then go through the inventory slots, checking each one and seeing if there's a slot free, or one that can be used(it has an item of the same type and not filled to capacity). We delete the item, and change the item name in the slot from NIL(if it was empty) and then change the item count. For this, I need to work with JSON files. There are premade functions in the global.gd autoload script, but it isn't dynamic yet. Done.
2025-04-01T04:10:09.256339
2015-12-09T12:58:50
121232202
{ "authors": [ "AliSoftware", "Fosile" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13355", "repo": "AliSoftware/OHAttributedStringAdditions", "url": "https://github.com/AliSoftware/OHAttributedStringAdditions/issues/7" }
gharchive/issue
After changing the Font its not working anymore Thanks for this great piece of Code! I have a little Problem when i try to change the Font and the Text Size, because after that my Text (which should be bold) is only a normal Text. This is my Code: NSMutableAttributedString* htmlString = [NSMutableAttributedString attributedStringWithHTML:@"<b>Lorem Ipsum</b>"]; [htmlString setFont:[UIFont fontWithName:@"Arial" size:14.0f]]; self.textView = [UITextView new]; self.textView.scrollEnabled = NO; [self addSubview:self.textView]; self.textView.attributedText = htmlString; [self.textView sizeToFit]; Well that's quite logical and the intended behavior: you set some style (using HTML) then override it using setFont afterwards. The information you probably missed is that in any text engine, the bold and italic attributes are actually not an attribute of the text, like it's color for example, but is part of the font. So even in Word for example, if you write some text in Arial, then set it Bold, in fact it replaces the font with "Arial-Bold". Arial is the Font Family, and "Arial-Regular", "Arial-Bold" and "Arial-Italics" are variants of that font family. So to implement the behavior you intended, you have multiple solutions: Avoid HTML Use the bold font directly. You can use +[UIFont fontWithFamily:@"Arial" size:14.0f bold:YES italic:NO as your font (see here). Or use setFontBold: on your NSMutableAttributedString after setting its font In both these solutions, there will then be no need to use the <b> HTML tag (or even initialize the string using HTML, which is slower generally than using a plain string and add attributes yourself) to make it bold. Keep HTML If you really need to initialize your string via HTML, you could: set the font via HTML too (<font> tag) or use enumerateFontsInRange:includeUndefined:usingBlock: that it will iterate over each run of your text that have different fonts (in case your HTML mix different fonts and you want to change them all but still keep the bold/italics/regular attribute of each) and alter the existing font to replace with the correct variant of the "Arial" font family For example you can use the -[UIFont symbolicTraits] to get the traits (bold, italics, etc) of the current font being iterated, then use [UIFont fontWithFamily:@"Arial" size:14 traits:currentTraits] to apply the Arial font with the same bold/italics traits instead. Thank you very much for your fast response, your explanation was fantastic!! I need to keep the HTML because usally i get my Text from a Server (which includes HTML). I changed my Code to this now, but it seems i can´t access the traits, because my Text still shows up without any any bold: NSMutableAttributedString* htmlString = [NSMutableAttributedString attributedStringWithHTML:@"<b>Lorem Ipsum</b>"]; [htmlString enumerateFontsInRange:NSMakeRange(0,htmlString.length) includeUndefined:YES usingBlock:^(UIFont *font, NSRange range, BOOL *stop) { UIFontDescriptorSymbolicTraits traits = [font symbolicTraits]; UIFont *newFont = [UIFont fontWithFamily:@"Arial" size:14.0f traits:traits]; [htmlString setFont:newFont range:range]; }]; self.textView = [UITextView new]; self.textView.scrollEnabled = NO; [self addSubview:self.textView]; self.textView.attributedText = htmlString; [self.textView sizeToFit]; Did you try to debug your code? Does the block in your enumeration actually properly get called, and if so how many iterations? What is the value of the traits you retrieve in each iteration ? If those traits you gather from the existing string do contain the bold attribute as expected, do the newFont find Arial-Bold or fail to find it and return another non-bold font variant instead? etc. So it seems the Problem is symbolicTraits and the Text from my Server, because after I do: UIFontDescriptorSymbolicTraits traits = [font symbolicTraits]; i only get some strange numbers inside my traits variable like "268435458" When I use your String from the Example Project, everything works fine and Debugger gives me traits = UIFontDescriptorClassOldStyleSerifs This is the Text from my Server: htmlText = @"<b>Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat, sed diam voluptua.</b><br /><br />At vero eos et accusam et justo duo dolores et ea rebum. Stet clita kasd gubergren, no sea takimata sanctus est Lorem ipsum dolor sit amet. Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat, sed diam voluptua. <br /><br /><b>At vero eos et accusam et justo duo dolores et ea rebum.</b><br /> Stet clita kasd gubergren, no sea takimata sanctus est Lorem ipsum dolor sit amet." Mmmh strange, do you know what is the original font generated by your HTML? Trick: you could wrap your HTML returned by the server into a <font face="Arial">…</font> before parsing it, it should also do the trick in a simpler way. Thank you very much, with the little trick its working!
2025-04-01T04:10:09.258807
2024-09-20T12:16:39
2538663005
{ "authors": [ "AliYoussef96", "antagomir" ], "license": "Artistic-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13356", "repo": "AliYoussef96/LimROTS", "url": "https://github.com/AliYoussef96/LimROTS/issues/6" }
gharchive/issue
Homepage You can set package homepage from the settings. It is automatically generated with pkgdown. The result will be like: https://microbiome.github.io/mia/ @ake123 can advice how it is done if you can't find instructions easily. The package's homepage is now live and accessible at: https://aliyoussef96.github.io/LimROTS/.
2025-04-01T04:10:09.294798
2021-03-10T05:12:53
827124628
{ "authors": [ "Alischrec" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13357", "repo": "Alischrec/aws-thought", "url": "https://github.com/Alischrec/aws-thought/issues/1" }
gharchive/issue
Create an AWS Account User Stories As a developer, I want to be able to view the management console in AWS. As a developer, I want to manage my IAM role. As a developer, I want to set up a billing alert. Completed this task
2025-04-01T04:10:09.352159
2017-08-30T01:48:15
253854102
{ "authors": [ "prajapati-parth", "uynap" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13358", "repo": "AllenFang/react-bootstrap-table", "url": "https://github.com/AllenFang/react-bootstrap-table/issues/1577" }
gharchive/issue
Is it possible to display multiple datafields in one column? For example, the front end got JSON {"firstName": "David", "lastName": "Jones"}. I would like to display it like: "Full Name" with the combination of firstName and lastName. Is it possible to do? Or I have to change the backend API ? @uynap Yes. Please consider the following example const data = [ { empId: 1, name: { firstName: 'David', lastName: 'Jones' } }, { empId: 2, name: { firstName: 'xyz', lastName: 'abc' } } ] function nameFormatter(cell, row) { return `${cell.firstName} ${cell.lastName}` } class MyTable extends React.component { render() { return ( <BootstrapTable data={ data }> <TableHeaderColumn dataField='empId'>Employee Id</TableHeaderColumn> <TableHeaderColumn dataField='name' dataFormat={ nameFormatter }>Full name</TableHeaderColumn> </BootstrapTable> ) } } Refer this page Section: Column Format with HTML String Thank you for the detail example. prajapati-parth . I was considering formatting the data at Redux's middleware level. Then it's one time for all. @uynap All right. I wasn't aware that you were using Redux. Well that is at least better than changing the backend API. :+1:
2025-04-01T04:10:09.357491
2022-01-06T14:11:27
1095357244
{ "authors": [ "elega", "jiacheliu3" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13359", "repo": "Alluxio/alluxio", "url": "https://github.com/Alluxio/alluxio/issues/14803" }
gharchive/issue
Make worker block lock pool size observable Is your feature request related to a problem? Please describe. Alluxio 2.7 The worker has a pool of block locks. Each lock item is locking one block (serve one to many clients). https://github.com/Alluxio/alluxio/blob/309fd69f45730910b3c0a7e127316d529c3544ef/core/server/worker/src/main/java/alluxio/worker/block/BlockLockManager.java#L55 The pool size should be observable by metric and logging, so we know when this pool is full. Describe the solution you'd like A clear and concise description of what you want to happen. Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered. Urgency MEDIUM Additional context Add any other context or screenshots about the feature request here. /assign @elega
2025-04-01T04:10:09.359406
2021-02-24T03:13:19
815042886
{ "authors": [ "alluxio-ci", "apc999", "maobaolong", "waterlx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13360", "repo": "Alluxio/alluxio", "url": "https://github.com/Alluxio/alluxio/pull/12924" }
gharchive/pull-request
Remove redundant blank in PropertyKey A format change, no logic involved. Can one of the admins verify this patch? @waterlx LGTM, please sign the contribution license agreement @waterlx Thank for your contribution, @apc999 Could you please take a look? Thank you @maobaolong and @apc999 for taking care of this change. alluxio-bot, merge this please
2025-04-01T04:10:09.360530
2023-06-29T21:54:48
1781563804
{ "authors": [ "Zijian-Zhu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13361", "repo": "Alluxio/alluxio", "url": "https://github.com/Alluxio/alluxio/pull/17714" }
gharchive/pull-request
[DOCFIX] Fix Language dropdown bug in new layout in os and add pr check Fix Language dropdown in new docs layout shows English twice alluxio-bot, merge this please
2025-04-01T04:10:09.362841
2017-03-28T15:45:27
217604981
{ "authors": [ "AmplabJenkins", "Reidddddd", "apc999" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13362", "repo": "Alluxio/alluxio", "url": "https://github.com/Alluxio/alluxio/pull/5004" }
gharchive/pull-request
[ALLUXIO-2365] Parameterize the returned type Inode in TempInodePathForDescendant.getInode https://alluxio.atlassian.net/browse/ALLUXIO-2365 Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/14099/ Test PASSed. Merged build finished. Test PASSed. LGTM. thanks!
2025-04-01T04:10:09.364467
2017-09-29T00:20:34
261498239
{ "authors": [ "AmplabJenkins", "calvinjia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13363", "repo": "Alluxio/alluxio", "url": "https://github.com/Alluxio/alluxio/pull/6206" }
gharchive/pull-request
[SMALLFIX] Clean up doc descriptions Start with capital letter and end with period. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/17070/ Test PASSed.
2025-04-01T04:10:09.373224
2020-10-23T19:00:19
728452200
{ "authors": [ "Almighty-Alpaca", "InfRandomness" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13364", "repo": "Almighty-Alpaca/JetBrains-Discord-Integration", "url": "https://github.com/Almighty-Alpaca/JetBrains-Discord-Integration/issues/131" }
gharchive/issue
New project settings having an issue I created a new project today with the plugin installed in my IDE and when I went inside my IDE to set the project settings I saw that for some reason, there's an "ask" field, and whenever I switch to another value, the field ask disappear. That felt weird to see and I don't know if that's a bug or not so I reported it just in case This actually is intentional: When you didn't click the pop-up yet the value is set to "Ask". But it doesn't make sense to manually select that as visibility because it would just mean that the next time you open the project the pop-up would show again. yeah, I see, thanks !
2025-04-01T04:10:09.387674
2020-10-07T12:19:40
716478630
{ "authors": [ "colourful-land", "hboon", "vladyslav-iosdev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13365", "repo": "AlphaWallet/alpha-wallet-ios", "url": "https://github.com/AlphaWallet/alpha-wallet-ios/issues/2198" }
gharchive/issue
AWS Lambda expect Bitcoin-style signature but was given Vitalik-style signature AWS Lambda accepts DER encoded signature like the one used in Bitcoin. Here is the code @vladyslav-iosdev used to create a signature: public func sign(hash: Data, privateKey: Data) throws -> Data { precondition(hash.count == 32, "Expect hash size to be 32") precondition(privateKey.count == 32, "Expect private key size to be 32") var signature = secp256k1_ecdsa_recoverable_signature() try privateKey.withUnsafeBytes { (key: UnsafePointer<UInt8>) in if secp256k1_ec_seckey_verify(context, key) != 1 { throw Secp256k1Error.invalidPrivateKey } let result = hash.withUnsafeBytes { hash in secp256k1_ecdsa_sign_recoverable(context, &signature, hash, key, nil, nil) } if result == 0 { throw Secp256k1Error.invalidPrivateKey } } var output = Data(count: 65) var recid = 0 as Int32 _ = output.withUnsafeMutableBytes { (output: UnsafeMutablePointer<UInt8>) in secp256k1_ecdsa_recoverable_signature_serialize_compact(context, output, &recid, &signature) } // add back recid to get 65 bytes sig output[64] = UInt8(recid) return output } Which apparently uses Vitalik style Ethereum signature. Vitalik could have the courtesy to extend the DER siganture format to do what he want, given DER was originally intended to be extensible and was extended all the time by different crypto standards, and if done right can result in a signature that is compatible with older (e.g. Bitcoin) codebase even without recompiling! But no, Vitalik likes to do things his way. There is a bit of hope that @vladyslav-iosdev said the code might actually call https://github.com/bitcoin-core/secp256k1/blob/master/include/secp256k1.h Which is a Bitcoin codebase - hence it must be able to store Bitcoin signature. (for one way to tell, a Bitcoin-style, i.e. Standard Comaptible, a signature is about 71 bytes, can be shorter occassionally, and has a variable length - which begets the question to Vitalik - 𝑎𝑟𝑒 𝑦𝑜𝑢 𝑠𝑢𝑟𝑒 that you want people to rewrite all libraries when you decided to extend the keysize when 256-bit is no longer enough? Or will you consider extensible standard, which was created after the pain of extending key size back in the 80s in the first place?) How can we get a Bitcoin-style Signature? I can change the Lambda to be compatible with Vitalik-style but that requires some research work into how that works. Alternatively, consider letting the mobile passing the Signature as (𝑥,𝑦) where each is BigInteger of 256-bits long, and I should be able to reconstruct needed signature in bytes at the server end. @vladyslav-iosdev I’ll look into this but if you have time before me, can you try look into the pod which contains HDWallet? Since it has multi-coin support, we might be able to use it or copy the code from there. The latter may be because the interface only supports seeds and not private keys directly. i found solution for this public func sign(_ data: Data, privateKey: Data) throws -> Data { let signature = UnsafeMutablePointer<secp256k1_ecdsa_signature>.allocate(capacity: 1) defer { signature.deallocate(capacity: 1) } let status = data.withUnsafeBytes { (ptr: UnsafePointer<UInt8>) in try? privateKey.withUnsafeBytes { (key: UnsafePointer<UInt8>) in secp256k1_ecdsa_sign(context, signature, ptr, key, nil, nil) } } guard status == 1 else { throw CryptoError.signFailed } let normalizedsig = UnsafeMutablePointer<secp256k1_ecdsa_signature>.allocate(capacity: 1) defer { normalizedsig.deallocate(capacity: 1) } secp256k1_ecdsa_signature_normalize(context, normalizedsig, signature) var length: size_t = 128 var der = Data(count: length) guard der.withUnsafeMutableBytes({ return secp256k1_ecdsa_signature_serialize_der(context, $0, &length, normalizedsig) }) == 1 else { throw CryptoError.noEnoughSpace } der.count = length return der }
2025-04-01T04:10:09.406719
2021-01-16T05:13:17
787351200
{ "authors": [ "Altimis", "ratriik", "the-data-mogul" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13367", "repo": "Altimis/Scweet", "url": "https://github.com/Altimis/Scweet/issues/31" }
gharchive/issue
Request to add the Location feature based on Latitude / Longitude or Zipcode Hi, Is it possible to replicate the location function (from Twint)? Or is that not possible because of the upgrades made by Twitter? Thank you! @the-data-mogul I'll see about that. @the-data-mogul I'll see about that. @the-data-mogul can you specify the exact file where the location function exists in twint project ? @the-data-mogul can you specify the exact file where the location function exists in twint project ? Thankyou very much for this repository. I've been searching how to scrape tweet in certain date then I found this, it really helps a lot! Btw, is there any ways to filter tweet based on location using Scweet? Sorry, I'm kinda new. Thankyou.
2025-04-01T04:10:09.427257
2023-08-10T12:14:01
1845059933
{ "authors": [ "vscaiceanu-1a" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13368", "repo": "AmadeusITGroup/otter", "url": "https://github.com/AmadeusITGroup/otter/issues/657" }
gharchive/issue
[Feature]: o3r/testing schematics should handle playwright setup Context Currently, o3r/core is adding the Playwright framework. Proposal Similarly to https://github.com/AmadeusITGroup/otter/issues/602 (unit tests), Playwright framework (e2) should be setup by the o3r/testing schematics. Fixed in https://github.com/AmadeusITGroup/otter/pull/663
2025-04-01T04:10:09.430042
2023-09-04T08:30:28
1879778209
{ "authors": [ "kpanot", "matthieu-crouzet" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13369", "repo": "AmadeusITGroup/otter", "url": "https://github.com/AmadeusITGroup/otter/issues/718" }
gharchive/issue
[Feature]: VSCode intellisense for configuration/styling metadata Context As a developper I would like to have help from my IDE to add metadata to the different modules of Otter. Proposal [ ] Intellisense for Configuration metadata [ ] Intellisense for Styling metadata [ ] EsLint plugin rule to validate configuration metadata [ ] Support of @o3rWidget https://github.com/AmadeusITGroup/otter/pull/893 Intellisense for @o3rWidget and @o3rWidgetParam https://github.com/AmadeusITGroup/otter/pull/896
2025-04-01T04:10:09.431801
2024-11-19T11:21:44
2671813966
{ "authors": [ "kpanot" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13370", "repo": "AmadeusITGroup/otter", "url": "https://github.com/AmadeusITGroup/otter/pull/2486" }
gharchive/pull-request
feat: move schematic to package dependencies to simplify ng-add Proposed change move @o3r/schematic package to packages dependencies to simplify ng-add and supporting it outside of Otter project. Related issues - No issue associated - Temporary closed due to the large amount of dependencies bring by it
2025-04-01T04:10:09.474534
2018-01-09T05:59:49
286978971
{ "authors": [ "Aminadav", "Passion08" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13371", "repo": "Aminadav/screenshot-extension", "url": "https://github.com/Aminadav/screenshot-extension/issues/135" }
gharchive/issue
Visible screen shot is not working Visible screenshot is not capturing any image inst ead it is displaying ba blank page Please test the last version, it should be fixed.