added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:10:07.889317
| 2021-07-28T15:48:21
|
954999308
|
{
"authors": [
"daverodgman",
"mpg",
"yuhaoth"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13269",
"repo": "ARMmbed/mbedtls",
"url": "https://github.com/ARMmbed/mbedtls/issues/4820"
}
|
gharchive/issue
|
TLS 1.3: simultaneous support for TLS 1.2 and 1.3
Enable support for building the library with support for both 1.2 and 1.3 available.
prototype issue: https://github.com/hannestschofenig/mbedtls/issues/297
From the above comment linking to PR in the prorotype's repo, and according to @yuhaoth's memories, this was about doing the work in the prototype, so it doesn't seem relevant for upstream, much less for the upstream MVP.
Hence, moving out of the MVP epic.
|
2025-04-01T04:10:07.893655
| 2022-02-25T10:19:39
|
1150278245
|
{
"authors": [
"gilles-peskine-arm",
"stevew817"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13270",
"repo": "ARMmbed/mbedtls",
"url": "https://github.com/ARMmbed/mbedtls/pull/5579"
}
|
gharchive/pull-request
|
Erase secrets in allocated memory before freeing said memory
Description
The PSA Crypto storage module uses dynamically-allocated buffers to buffer sensitive key material during the process of loading said key material from NVM, or storing it to NVM. To avoid leaking the key material through the heap, zero out the buffer before freeing it.
Status
READY
Requires Backporting
Unsure, maintainer can make a decision
Migrations
NO
Additional comments
Todos
[x] Tests
[x] Documentation
[x] Changelog updated
[ ] Backported
Steps to test or reproduce
@daverodgman does this need a changelog entry?
Yes, please add a changelog entry.
grep '\* Zeroize' ChangeLog
If this needs backporting, can anybody tell me which branches this would be expected on?
There is currently only one long-time support branch: 2.28. So please backport to mbedtls-2.28.
Backported.
|
2025-04-01T04:10:07.897014
| 2017-01-10T13:39:00
|
199823498
|
{
"authors": [
"NWilson",
"andresag01",
"daverodgman"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13271",
"repo": "ARMmbed/mbedtls",
"url": "https://github.com/ARMmbed/mbedtls/pull/745"
}
|
gharchive/pull-request
|
DRAFT: Threefish implementation
This PR is an early prototype implementation of the Threefish tweakable block cipher. The implementation is based on the specification at https://www.schneier.com/academic/paperfiles/skein1.3.pdf.
This PR includes:
Block cipher implementation and header file (threefish.c and threefish.h)
Changes to the make and CMake build files
A test suite test_suite_threefish
Documentation
Outstanding work:
Additional tests (both: selt test and the test suite)
Integration with mbed TLS cipher abstraction: Need to accommodate the mbedtls_threefish_settweak() function with the existing API
Hey, looks cool! Nicer than AES (in my opinion). Do you expect Threefish to be used over ChaCha20 - or is it losing the popularity contest as the successor to AES?
Closing as we don't appear to have demand for this algorithm.
|
2025-04-01T04:10:07.900813
| 2018-06-05T13:05:47
|
329447557
|
{
"authors": [
"adbridge",
"artokin",
"ciarmcom"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13272",
"repo": "ARMmbed/nanostack-border-router",
"url": "https://github.com/ARMmbed/nanostack-border-router/issues/114"
}
|
gharchive/issue
|
Thread commission fails when Border Router is built with using IAR or ARMC6
Thread commissioning fails when Border Router is build with IAR or ARMC6 compiler. Commissioning works fine when builds are made by using ARM or GCC_ARM.
Commissioning device (mobile phone) does not enter to screen where QR-code can be scanned. Therefore I'm not able to add new devices to the network.
Tests were made against mbed-os-5.9.0-oob branch.
@artokin ARMC6 is not officially supported yet. Does it work with ARMC5 ?
ARM Internal Ref: IOTTHD-2595
This is working fine witn ARMC5 and GCC__ARM. It looks like the problem exists only with IAR/ARMC6.
I was using IAR ANSI C/C++ Compiler V<IP_ADDRESS>/LNX for ARM (from CI) and IAR Workbench 8.11.2
PR https://github.com/ARMmbed/nanostack-border-router/pull/116 will adjust mbedtls MBEDTLS_SSL_MAX_CONTENT_LEN size to 4096 (from 16k). Then IAR BR commissioning works as memory allocation succeeds.
PR merged to mbed-os-5.9.0-oob branch.
Issue with ARMC6 commissioning still remains but as ARMC6 is not officially supported no further actions will be made.
Closing this issue.
|
2025-04-01T04:10:07.903761
| 2017-08-21T15:11:46
|
251688718
|
{
"authors": [
"deepikabhavnani",
"soramame21"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13273",
"repo": "ARMmbed/sd-driver",
"url": "https://github.com/ARMmbed/sd-driver/pull/56"
}
|
gharchive/pull-request
|
Add SD card reader support over DISCO_L476VG platform for running Cloud Client.
@MarceloSalazar Add SPI pins setting for support SD card reader/writer over DISCO_L476VG, which is required and tested by using ARMmbed/mbed-cloud-client-example-sources-internal.
@soramame21 - Please verify the SD card specific test, you can have more details of that in "Build the mbed OS Test Cases" section of Readme.
Required tests to be successful:
sd-driver-tests-block_device-basic
sd-driver-tests-filesystem-basic
sd-driver-tests-filesystem-fopen
@deepakvenugopal, @MarceloSalazar Thanks for your comments. I had run and passed the test. please check the attached test result.
sd-driver-test-result-Ren0822.txt
|
2025-04-01T04:10:07.919985
| 2023-02-22T21:21:31
|
1595849147
|
{
"authors": [
"asjohnston-asf",
"jtherrmann"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13274",
"repo": "ASFHyP3/gis-services",
"url": "https://github.com/ASFHyP3/gis-services/pull/8"
}
|
gharchive/pull-request
|
Refactoring for make-hand-service.py
Makes the following changes:
Logging with print statements.
Use tempfiles where appropriate.
Add cli options.
I'm currently running the script with:
python make_hand_service.py /home/arcgis/jtherrmann/ --rasters-filter='.*Copernicus_DSM_COG_10_N08_00_W01*'
But it fails; full log output:
CreateFileGDB
CreateMosaicDataset
AddFields
AddRastersToMosaicDataset
CalculateFields
BuildFootprints
Traceback (most recent call last):
File "make_hand_service.py", line 90, in <module>
update_boundary='UPDATE_BOUNDARY',
File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/management.py", line 14697, in BuildFootprints
raise e
File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/management.py", line 14694, in BuildFootprints
retval = convertArcObjectToPythonObject(gp.BuildFootprints_management(*gp_fixargs((in_mosaic_dataset, where_clause, reset_footprint, min_data_value, max_data_value, approx_num_vertices, shrink_distance, maintain_edges, skip_derived_images, update_boundary, request_size, min_region_size, simplification_method, edge_tolerance, max_sliver_size, min_thinness_ratio), True)))
File "/home/arcgis/miniconda3/envs/arcpy/lib/python3.7/site-packages/arcpy/geoprocessing/_base.py", line 512, in <lambda>
return lambda *args: val(*gp_fixargs(args, True))
arcgisscripting.ExecuteError: ERROR 999999: Something unexpected caused the tool to fail. Contact Esri Technical Support (http://esriurl.com/support) to Report a Bug, and refer to the error help for potential solutions or workarounds.
No mosaic dataset item selected. This operation cannot be performed.
Failed to execute (BuildFootprints).
Start Time: Wednesday, February 22, 2023 9:16:36 PM
ERROR 999999: Something unexpected caused the tool to fail. Contact Esri Technical Support (http://esriurl.com/support) to Report a Bug, and refer to the error help for potential solutions or workarounds.
No mosaic dataset item selected. This operation cannot be performed.
Failed to execute (BuildFootprints).
Failed at Wednesday, February 22, 2023 9:16:38 PM (Elapsed Time: 2.59 seconds)
I wonder if perhaps the regex filter is not matching any rasters? This will be my next line of investigation.
Drop the .s from the filter, I suspect. Per https://pro.arcgis.com/en/pro-app/latest/tool-reference/data-management/add-rasters-to-mosaic-dataset.htm by default you're doing simple wildcard matching, e.g. '*Copernicus_DSM_COG_10_N08_00_W01*'. It's only if you put the word REGEX` in the filter that you get full perl regular expressions.
Drop the .s from the filter, I suspect. Per https://pro.arcgis.com/en/pro-app/latest/tool-reference/data-management/add-rasters-to-mosaic-dataset.htm by default you're doing simple wildcard matching, e.g. '*Copernicus_DSM_COG_10_N08_00_W01*. It's only if you put the word REGEX in the filter that you get full perl regular expressions.
Ah, I see, I added REGEX because I thought you just left that part out of the filter you gave me. So I thought you had just forgotten the ..
I ran it as python make_hand_service.py --dataset-name GLO30_HAND_ASJ --rasters-filter '*Copernicus_DSM_COG_10_N08_00_W01*' /home/arcgis/asjohnston/gis-services/image_services/glo_30_hand/ and everything worked how I expected. The .sd file published with no issues.
|
2025-04-01T04:10:07.927300
| 2019-08-28T19:00:53
|
486554890
|
{
"authors": [
"chrsengel",
"heldyboy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13275",
"repo": "ASRG/asrg.io",
"url": "https://github.com/ASRG/asrg.io/issues/24"
}
|
gharchive/issue
|
Calendar Function
Need to find a calendar program. Could be the same as what the auto-ISAC uses. Then we should collect all the conferences from the calendar sources and put them into a single database. This database should be used to show the conference calendar.
Additional Ideas
Webcrawlers to find new sources
Chris, can you see if you can start this? The best way from my point of view is to take an existing calendar and copy / paste.... then we can add our own content.
Sure. I'll look for possible solutions.
Hey, can we discuss this one as well?
We could embed a public google calendar into the site. You only need to create one and make it public. Here's a guide.
Otherwise I don't know of any reliable solution for static pages. If we transition to a CMS like wordpress or build up our own infrastructure there will be more options.
|
2025-04-01T04:10:07.937543
| 2024-08-19T19:44:15
|
2474074462
|
{
"authors": [
"ASzc",
"adambkaplan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13276",
"repo": "ASzc/fbc-utils",
"url": "https://github.com/ASzc/fbc-utils/issues/4"
}
|
gharchive/issue
|
Allow OCI Manifests as Valid MIME Type
OCI Manifest lists are the equivalent/replacement of Docker manifests. Konflux produces operator bundles with this MIME type.
MIME: application/vnd.oci.image.manifest.v1+json
Done, thanks!
|
2025-04-01T04:10:07.940873
| 2023-02-05T11:55:42
|
1571400089
|
{
"authors": [
"SuelaCark",
"asynchroza"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13277",
"repo": "AUBGTheHUB/spa-website-2022",
"url": "https://github.com/AUBGTheHUB/spa-website-2022/issues/212"
}
|
gharchive/issue
|
Change "Welcome to The Hub" to something else in landing section.
Reference design in Adobe
Forgot to say that this issue is covered within another ticket #341
|
2025-04-01T04:10:07.974863
| 2020-10-13T01:56:22
|
719790363
|
{
"authors": [
"codecov-io",
"siangernlow"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13278",
"repo": "AY2021S1-CS2103T-T13-1/tp",
"url": "https://github.com/AY2021S1-CS2103T-T13-1/tp/pull/147"
}
|
gharchive/pull-request
|
Alert user when SHN has been violated
When a visit is being added, a warning should be shown if the person has made an illegal visit.
A visit is illegal if:
The visiting person is infected or in quarantine or both.
The location that the person visited is not the address stated in the person.
Subsequently, the visit will still be added, but the relevant warning will be generated as well.
Close #143
Codecov Report
Merging #147 into master will increase coverage by 0.04%.
The diff coverage is 85.71%.
@@ Coverage Diff @@
## master #147 +/- ##
============================================
+ Coverage 73.98% 74.03% +0.04%
- Complexity 789 794 +5
============================================
Files 113 113
Lines 2464 2476 +12
Branches 304 308 +4
============================================
+ Hits 1823 1833 +10
- Misses 551 552 +1
- Partials 90 91 +1
Impacted Files
Coverage Δ
Complexity Δ
.../address/logic/commands/visit/AddVisitCommand.java
88.57% <85.71%> (-2.74%)
13.00 <6.00> (+5.00)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update aab6c10...e8471c0. Read the comment docs.
|
2025-04-01T04:10:07.981357
| 2021-11-08T00:48:49
|
1046902396
|
{
"authors": [
"codecov-commenter",
"ruoyann"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13279",
"repo": "AY2122S1-CS2103T-T09-1/tp",
"url": "https://github.com/AY2122S1-CS2103T-T09-1/tp/pull/252"
}
|
gharchive/pull-request
|
Reordering of clear before help feature
Thought it makes more sense to list out all the commands before help and exit. Changed the summary to match the flow as well.
Codecov Report
Merging #252 (511d9c8) into master (7f8dd2c) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #252 +/- ##
=========================================
Coverage 68.19% 68.19%
Complexity 745 745
=========================================
Files 94 94
Lines 2581 2581
Branches 346 346
=========================================
Hits 1760 1760
Misses 763 763
Partials 58 58
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 7f8dd2c...511d9c8. Read the comment docs.
|
2025-04-01T04:10:07.985493
| 2021-10-30T05:34:57
|
1040057848
|
{
"authors": [
"angnobel",
"nus-pe-script"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13280",
"repo": "AY2122S1-CS2103T-W13-1/tp",
"url": "https://github.com/AY2122S1-CS2103T-W13-1/tp/issues/199"
}
|
gharchive/issue
|
[PE-D] Consistent sequencing of features
Minor point again but would be good if the sequence of features are consistent. Eg since it starts with the job position features in the table of contents, start with job position features in command summary.
Labels: severity.Low type.DocumentationBug
original: yucheng11122017/ped#17
Split into 3 tables
|
2025-04-01T04:10:07.988389
| 2021-10-30T05:17:56
|
1040054805
|
{
"authors": [
"huizhuansam",
"nus-pe-script"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13281",
"repo": "AY2122S1-CS2103T-W13-2/tp",
"url": "https://github.com/AY2122S1-CS2103T-W13-2/tp/issues/155"
}
|
gharchive/issue
|
[PE-D] Find Command "--phone" flag issues
Input : find --phone 99272758
Output: Invalid command format! find: Finds all persons whose fields contain any of the specified keywords (case-insensitive) and displays them as a list with index numbers. Parameters: FLAG KEYWORDS [MORE_FLAGS] Example: find -n alex -a serangoon
User Guide Suggests that this is an acceptable command.
Labels: type.FunctionalityBug severity.Medium
original: Jai2501/ped#3
Related to #150
Update: this issue is actually related to the lack of support of long-form prefixes by the Find command. @Zhou-Jiahao-1998 please look into this.
|
2025-04-01T04:10:08.023350
| 2022-10-26T11:28:43
|
1423867914
|
{
"authors": [
"codecov-commenter",
"teoyuqi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13282",
"repo": "AY2223S1-CS2103T-T10-1/tp",
"url": "https://github.com/AY2223S1-CS2103T-T10-1/tp/pull/107"
}
|
gharchive/pull-request
|
Update DG
Reduces steps in add-task
Add rationale for design decisions of Task
Update section on Logic, Ui, Model and Storage to be in line with our current implementation:
Change links from SEEDU repo to our repo
Update diagrams
Remove unused diagrams
Codecov Report
Base: 77.84% // Head: 77.55% // Decreases project coverage by -0.29% :warning:
Coverage data is based on head (5a62927) compared to base (e41d3c0).
Patch has no changes to coverable lines.
Additional details and impacted files
@@ Coverage Diff @@
## master #107 +/- ##
============================================
- Coverage 77.84% 77.55% -0.30%
- Complexity 741 745 +4
============================================
Files 111 111
Lines 2221 2245 +24
Branches 264 265 +1
============================================
+ Hits 1729 1741 +12
- Misses 398 408 +10
- Partials 94 96 +2
Impacted Files
Coverage Δ
...eedu/address/logic/commands/DeleteLinkCommand.java
85.71% <0.00%> (-5.72%)
:arrow_down:
...c/main/java/seedu/address/model/module/Module.java
84.84% <0.00%> (-5.32%)
:arrow_down:
...a/seedu/address/model/module/UniqueModuleList.java
80.32% <0.00%> (-4.29%)
:arrow_down:
src/main/java/seedu/address/ui/TaskCard.java
0.00% <0.00%> (ø)
src/main/java/seedu/address/ui/ModuleCard.java
0.00% <0.00%> (ø)
src/main/java/seedu/address/ui/TaskListPanel.java
0.00% <0.00%> (ø)
...c/main/java/seedu/address/model/person/Person.java
100.00% <0.00%> (ø)
src/main/java/seedu/address/model/AddressBook.java
93.33% <0.00%> (+0.15%)
:arrow_up:
...a/seedu/address/model/person/UniquePersonList.java
90.19% <0.00%> (+0.83%)
:arrow_up:
src/main/java/seedu/address/model/person/Name.java
90.90% <0.00%> (+0.90%)
:arrow_up:
... and 1 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
|
2025-04-01T04:10:08.025567
| 2022-10-29T01:58:44
|
1428018166
|
{
"authors": [
"Eclipse-Dominator",
"nus-se-script"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13283",
"repo": "AY2223S1-CS2103T-T11-1/tp",
"url": "https://github.com/AY2223S1-CS2103T-T11-1/tp/issues/102"
}
|
gharchive/issue
|
[PE-D][Tester A] App has no error message when add command pairs with invalid email format
No error message when inputing email address without @ command. It remains at the output of previous input
Labels: severity.Medium type.FeatureFlaw
original: PokezardVGC/ped#8
This issue will be addressed by #153
|
2025-04-01T04:10:08.026758
| 2022-09-27T13:58:33
|
1387818686
|
{
"authors": [
"chao890",
"kevinchangjk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13284",
"repo": "AY2223S1-CS2103T-T15-1/tp",
"url": "https://github.com/AY2223S1-CS2103T-T15-1/tp/pull/39"
}
|
gharchive/pull-request
|
Update delete function for User Guide
Resolves #31
Added and modified the existing delete function in the User Guide.
Good.
|
2025-04-01T04:10:08.027636
| 2022-09-19T10:57:07
|
1377770162
|
{
"authors": [
"jialatteo",
"wongyewjon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13285",
"repo": "AY2223S1-CS2103T-W09-2/tp",
"url": "https://github.com/AY2223S1-CS2103T-W09-2/tp/pull/3"
}
|
gharchive/pull-request
|
Add parameter hints
Parameter hints are displayed in ResultDisplay when a command word is typed
Resolve #67
Feature would have to update new commands in the future
|
2025-04-01T04:10:08.039567
| 2022-10-01T20:58:48
|
1393549397
|
{
"authors": [
"alvinjiang1",
"codecov-commenter",
"jq1836"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13286",
"repo": "AY2223S1-CS2103T-W12-3/tp",
"url": "https://github.com/AY2223S1-CS2103T-W12-3/tp/pull/55"
}
|
gharchive/pull-request
|
Update Use Cases and DG Format
Things done:
Update Use cases to follow the format as stated in Issue #44.
Update DESIGN section of DG.
Before:
After:
Update headers: "(OUTDATED)" appended to the end when that section is outdated.
Update Glossary to include CSV.
Codecov Report
Merging #55 (b296751) into master (3fb88d6) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #55 +/- ##
=========================================
Coverage 72.15% 72.15%
Complexity 399 399
=========================================
Files 70 70
Lines 1232 1232
Branches 125 125
=========================================
Hits 889 889
Misses 311 311
Partials 32 32
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
Documentation
[X] User Guide
[X] Developer Guide
[X] Javadocs
Sections updated under UG
NIL
Sections updated under DG
DESIGN
Appendix: Requirements -> Use cases
Classes with Javadocs update
NIL
Here are some guidelines I followed while updating the DG, it would be good to follow these in future commits to DG, to ensure consistency:
I intend to follow closely the HighWater DG for most aspects of our DG as I feel that it is very well-written, and if our DG is similar in quality to that, there should be no issues to our documentation.
I have also referred to the AB3 DG a lot regarding when to use callouts and markups.
Use of present tense to signify work in progress (e.g. developers are to add additional classes...)
Make use of Cause - Effect Structure (e.g. CLI gives the user an easy way to input commands... Therefore ...)
Liberally make use of Markups (e.g. when describing relevant classes)
When coming across jargon, do remember to add it in the Glossary section in alphabetical order.
As a follow up of point 4, when referencing words from the Glossary, do add a clickable link that takes the reader back to the glossary section. I have decided that for this particular purpose, the format will just be a bolded, clickable word. Our DG is online so the word will show up in blue.
For example, "RC4" is a word in the glossary, and should be written as [**RC4**](#glossary)
IMPORTANT!!!
When writing the description in your PRs in the future, do add the section I have written at the top. I will check the documentation based on this section of the PR description.
Checking the tasklist signifies that that portion of the documentation is ready to be checked (even if it may not have been changed at all). The "Sections updated under..." portion will tell me which part of the UG/DG/java files to look at/compare.
As of now, please DO NOT change the headers with the (OUTDATED) word. I intend to check through those sections again before we submit v1.2, and so I will remove it myself :))
Oh one more thing that would be of great help is to deploy the docs on your repo and include a link to the DG. Would help a lot with seeing how the actual documentation will look like.
|
2025-04-01T04:10:08.049111
| 2023-10-08T09:04:56
|
1931719365
|
{
"authors": [
"codecov-commenter",
"derekjxtan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13287",
"repo": "AY2324S1-CS2103T-T08-1/tp",
"url": "https://github.com/AY2324S1-CS2103T-T08-1/tp/pull/68"
}
|
gharchive/pull-request
|
Derek/add generic uniqueitemlist
Added a generic UniqueItemList class, which mirrors the functionality of uniquePersonList. To be used for patients, doctors and appointments. Meant to replace UniquePersonList.
Added interface Listable for all classes that are to be managed using a UniqueItemList.
Codecov Report
Merging #68 (31bc6c5) into master (b772575) will increase coverage by 0.30%.
The diff coverage is 84.90%.
:exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality.
@@ Coverage Diff @@
## master #68 +/- ##
============================================
+ Coverage 75.85% 76.15% +0.30%
- Complexity 440 459 +19
============================================
Files 73 76 +3
Lines 1379 1430 +51
Branches 133 140 +7
============================================
+ Hits 1046 1089 +43
- Misses 303 307 +4
- Partials 30 34 +4
Files
Coverage Δ
...odel/person/exceptions/DuplicateItemException.java
100.00% <100.00%> (ø)
...model/person/exceptions/ItemNotFoundException.java
100.00% <100.00%> (ø)
...va/seedu/address/model/person/patient/Patient.java
96.55% <100.00%> (+0.25%)
:arrow_up:
...u/address/model/person/patient/UniqueItemList.java
82.60% <82.60%> (ø)
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
|
2025-04-01T04:10:08.054087
| 2024-10-23T09:14:26
|
2607866274
|
{
"authors": [
"InfinityTwo",
"codecov-commenter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13288",
"repo": "AY2425S1-CS2103T-T10-2/tp",
"url": "https://github.com/AY2425S1-CS2103T-T10-2/tp/pull/144"
}
|
gharchive/pull-request
|
Modify ModelClassDiagram and Diagram Colours
Describe your changes
Remove tags from ModelClassDiagram
Add Department, Role and ContractEndDate to ModelClassDiagram
Modify the Diagram colours a bit (for extra lines of code)
Issue ticket number and link
Builds on #138
Checklist before requesting a review
[x] I have performed a self-review of my code.
[x] If it is a core feature, I have added thorough tests.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
|
2025-04-01T04:10:08.056004
| 2016-12-14T20:42:00
|
195640359
|
{
"authors": [
"Jfairfield1"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13289",
"repo": "AZMAG/map-Employment",
"url": "https://github.com/AZMAG/map-Employment/issues/33"
}
|
gharchive/issue
|
Legend doesn't always load properly
From @Jfairfield1 on December 14, 2016 20:39
Sometimes the legend will not properly load into the page.
My best guess is that a race is going on between the cbr request and the boundaries request.
Copied from original issue: AZMAG/map-DemographicState#42
This issue was moved to AZMAG/map-DemographicRegional#40
|
2025-04-01T04:10:08.072669
| 2023-04-06T14:50:37
|
1657574625
|
{
"authors": [
"lauritapio",
"mhinkkan",
"murgui"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13290",
"repo": "Aalto-Electric-Drives/motulator",
"url": "https://github.com/Aalto-Electric-Drives/motulator/issues/85"
}
|
gharchive/issue
|
mt.BaseValues() examples and documentation don't match expected value in 0.1.2
Hi,
I'm using v0.1.2, I tried running 'plot_vector_ctrl_pmsm_2kw.py' and it reports the error
Traceback (most recent call last):
base = mt.BaseValues(
TypeError: BaseValues.__init__() got an unexpected keyword argument 'n_p'
I checked the reference at https://aalto-electric-drives.github.io/motulator/autoapi/motulator/index.html#motulator.BaseValues and indeed the example used the suggested arguments.
I removed the argument and then the error was that it was missing argument 'p', so it looks like n_p has been renamed to p, so either the examples should be changed to p or the function should be changed to n_p.
Regards
Thanks for pointing this out! The polepair number was indeed recently named to n_p, but it seems that there are some places still where this has not been changed. We'll get this fixed shortly!
Thank you for the valuable feedback!
The reason for the problem might be that we have updated the repository since last release quite a lot, including the change in the variable name (p -> n_p). To fix the problem, new release 0.1.3 was published a moment ago.
Furthermore, the documentation regarding installation was updated:
https://aalto-electric-drives.github.io/motulator/installation.html
You might be interested in Option 2, i.e. cloning the repository. It allows you to modify the existing system models and the example controllers. The example instructions are for VS Code, but quite similar steps could be used with other IDEs.
Please inform us if the problem still appears after installing the latest release.
|
2025-04-01T04:10:08.092661
| 2021-12-17T09:01:41
|
1083031677
|
{
"authors": [
"AaronErhardt",
"euclio"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13291",
"repo": "AaronErhardt/relm4",
"url": "https://github.com/AaronErhardt/relm4/issues/57"
}
|
gharchive/issue
|
Inconsistent behavior between FactoryVec and FactoryVecDeque
I'm trying to create a widget that adds and removes an arbitrary number of children at the position before some hard-coded widgets. I tried modifying the factory example like so:
gtk::ApplicationWindow {
set_default_width: 300,
set_default_height: 200,
set_child = Some(>k::Box) {
set_orientation: gtk::Orientation::Vertical,
set_margin_all: 5,
set_spacing: 5,
factory!(model.counters),
append = >k::Button {
set_label: "Add",
connect_clicked(sender) => move |_| {
send!(sender, AppMsg::Add);
}
},
append = >k::Button {
set_label: "Remove",
connect_clicked(sender) => move |_| {
send!(sender, AppMsg::Remove);
}
},
append = >k::Button {
set_label: "I'M LAST",
},
}
}
If I use FactoryVec for my storage, the new widgets are added and removed at the last position (after the "I'M LAST" label). However, if I use FactoryVecDeque for my storage, the widgets are added at the first position (before the "Add" button). The latter behavior happens to be what I want, but I wanted to open this issue to see if this is something I can rely on, or if there's a better way to accomplish what I want.
The reason why this happens is because there are different traits used behind the scenes. FactoryView is used by both but FactoryListView is only required by FactoryVecDeque to enable more flexibility. That's because a Vec can only push and pop at the end whereas the VecDeque needs to take care about positioning. Since not all widgets support arbitrary positions both traits exist to at least allow FactoryVec to be used.
This also causes FactoryVec to use append to insert widgets and FactoryVecDeque uses insert_after unless it's the first widget. Of course it's better to use a separate container only for your factory but I'm pretty sure that this behaviour won't change anytime soon.
Of course it's better to use a separate container only for your factory
Agreed, but in this case I need the children to all be children of the same widget (a PanelPaned from libpanel).
I'd suggest then to document the current behaviour. Any change to this (which is not very likely to happen anyway) would then be avoided or at least considered a breaking change.
The behavior is now documented: https://aaronerhardt.github.io/docs/relm4/relm4/factory/collections/index.html
|
2025-04-01T04:10:08.115400
| 2019-02-26T09:47:45
|
414519947
|
{
"authors": [
"nre-ableton",
"rco-ableton"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13292",
"repo": "AbletonDevTools/groovylint",
"url": "https://github.com/AbletonDevTools/groovylint/pull/55"
}
|
gharchive/pull-request
|
Add logging support to run_codenarc.py
This PR adds logging support to run_codenarc.py, and also adds a custom log formatter so that the CodeNarc output matches the output of this script. With these changes, a sample --verbose run (with errors) looks like this:
DEBUG Executing CodeNarc command: java -Dorg.slf4j.simpleLogger.showThreadName=false -Dorg.slf4j.simpleLogger.defaultLogLevel=debug -classpath /Users/nre/Code/AbletonDevTools/groovylint/jars:/usr/local/opt/groovysdk/libexec/lib/*:/Users/nre/Code/AbletonDevTools/groovylint/jars/CodeNarc-1.2.1.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/GMetrics-1.0.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/slf4j-1.7.25/slf4j-api-1.7.25.jar:/Users/nre/Code/AbletonDevTools/groovylint/jars/slf4j-1.7.25/slf4j-simple-1.7.25.jar org.codenarc.CodeNarc -rulesetfiles=ruleset.groovy -report=xml:codenarc-report.xml -includes=./Jenkinsfile,**/*.groovy
INFO org.codenarc.ruleregistry.PropertiesFileRuleRegistry - Loaded properties file in 43ms; 362 rules
INFO org.codenarc.ruleset.GroovyDslRuleSet - Loading ruleset from [ruleset.groovy]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/basic.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/braces.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/convention.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/design.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/exceptions.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/formatting.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/generic.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/groovyism.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/imports.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/naming.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/security.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/size.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/unnecessary.xml]
INFO org.codenarc.ruleset.XmlFileRuleSet - Loading ruleset from [rulesets/unused.xml]
INFO org.codenarc.ruleset.PropertiesFileRuleSetConfigurer - RuleSet configuration properties file [codenarc.properties] not found.
DEBUG org.codenarc.CodeNarcRunner - results=DirectoryResults(null) [DirectoryResults() [DirectoryResults(vars) [], DirectoryResults(jars) [], FileResults(Jenkinsfile) [Violation[rule=VariableTypeRequiredRule[name=VariableTypeRequired, priority=3], lineNumber=59, sourceLine=def foo = 'bar', message=The type is not specified for variable "foo"], Violation[rule=UnusedVariableRule[name=UnusedVariable, priority=2], lineNumber=59, sourceLine=def foo = 'bar', message=The variable [foo] in class None is not used]]]]
INFO org.codenarc.report.AbstractReportWriter - No custom message bundle found for [codenarc-messages]. Using default messages.
INFO org.codenarc.report.AbstractReportWriter - Report file [codenarc-report.xml] created.
INFO CodeNarc completed: (p1=0; p2=1; p3=1) 7291ms
DEBUG CodeNarc returned with code 0
DEBUG Reading report file codenarc-report.xml
DEBUG Removing report file codenarc-report.xml
DEBUG Parsing report XML
INFO Scanned 4 files
DEBUG Parsing violations in package: .
DEBUG Parsing violations in file: ./Jenkinsfile
ERROR ./Jenkinsfile:59: VariableTypeRequired: The type is not specified for variable "foo"
ERROR ./Jenkinsfile:59: UnusedVariable: The variable [foo] in class None is not used
ERROR Found 2 violation(s)
ping @AbletonDevTools/gotham-city
Hmm, it seems that pylint does not like the fact that --verbose appears in two separate files. 🙄 How should I fix this? Can I just # noqa it?
Sorry for the force-push, but I made some pretty big changes on this PR and the fixups would have gotten out of control quickly.
Namely, I decided to abandon the idea of using a custom Formatter class and now instead introduced a function to examine each line of CodeNarc's output and re-log it appropriately. I realized that previously we were not respecting CodeNarc's logging levels and were just logging everything it produced as logging.DEBUG, which is not very nice. I think that it's better to look at the actual log level (which is pretty easy to parse) and use the same when calling logging.log().
@rco-ableton @mst-ableton do either of you have an opinion on this PR?
@nre-ableton please rebase and I'll review!
@rco-ableton Done!
|
2025-04-01T04:10:08.124695
| 2019-07-19T19:48:07
|
470511030
|
{
"authors": [
"AnshumanDwivedi",
"geethab123",
"realraviteja",
"satish765001",
"yruslan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13293",
"repo": "AbsaOSS/cobrix",
"url": "https://github.com/AbsaOSS/cobrix/issues/140"
}
|
gharchive/issue
|
multi copybook or not
I have a copybook which has multiple 01 levels. Can this copy book is parsed as variable length.
can I parse like below
cobolDataframe = spark
.read
.format("cobol")
.option("copybook", v_copybook)
.option("schema_retention_policy", "collapse_root") //removes the root record headerc
.option("drop_group_fillers", "false")
.option("generate_record_id", false) // this adds the file id and record id
.option("is_record_sequence", "true") // reader to use 4 byte record headers to extract records from a mainframe file
.option("is_rdw_big_endian", "true")
//.option("is_rdw_part_of_record_length", false)
.option("rdw_adjustment", -4)
.load(v_data)
Please confirm.
After parsing I am not getting correct parsed data.
Yes it can. If your level 1 entries represent segments they should also redefine each other.
01 RECORD1.
...
01 RECORD2 REDEFINES RECORD1.
...
They are not redefining. Still can be treated as variable length
Seems Cobrix is not able to process multiple copybooks like syncsort is doing. @CobrixTeam : could you please check and suggest in this regards ?
Sure, we can help. Could you provide an example set of copybooks and a small example data file?
@yruslan we are also having similar issue:
Quick summary About the issue:
In my project i have a requirement to read multiple copybooks as One main Copybook using Cobrix.
Example:
TRANSACTION.CPY
01 CUSTOMERS
05 PURCHASES
FIRST_NAME PIC X (04)
LASTNAME_NAME PIC X (04)
01 ORGANIZATION
05 DEPARTMENTS
ORG_NAME PIC X (04)
VENDOR_NAME PIC X (09)
01 MEDICARE
05 BILLS
TREATMENT_TYPE PIC X (04)
LOCATION_NAME PIC X (09)
01 MEMBERSHIP
05 PARTNERS
PARTNER_TYPE PIC X (04)
PARTNER_NAME PIC X (09)
I have similar kind of layout copybook and i'm using cobrix to parse the data.
In my research what i understood is i cant read any copy which will have multiple main headers .
Cause Cobrix can handle only one Main header copybook no multiple.
Please provide your thoughts and resolutions for the above issue.
i have similar requirement where copybook contains three 01 levels . One header , one trailer and the another for data with occurs depending on . Header and trailer has length 30 and data levels varies as its occurs depending on. Any suggestion would be greatly appreciated .
|
2025-04-01T04:10:08.130384
| 2018-10-25T15:46:54
|
374014011
|
{
"authors": [
"Jack12816",
"mungler"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13294",
"repo": "Absolventa/emarsys-rb",
"url": "https://github.com/Absolventa/emarsys-rb/pull/61"
}
|
gharchive/pull-request
|
Improved the Contact.query method
- What is it good for
This PR ships an enhanced Contact.query method which allows to query all contacts without requiring any filter. Furthermore it transforms the key_id and key_value when passed.
- What I did
I added the transformation of the passed key_id and the possibility to pass in custom filters. The YARD documentation was updated to reflect this with more examples and for the new usages specs were added.
- A picture of a cute animal (not mandatory but encouraged)
Looks good 👍 and kudos on the cute animal 😍 😁
FYI @Jack12816 published new version 0.3.13 to rubygems.org with these changes 👍
|
2025-04-01T04:10:08.133358
| 2024-08-14T03:10:13
|
2464798660
|
{
"authors": [
"jstone-lucasfilm",
"ld-kerley"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13295",
"repo": "AcademySoftwareFoundation/MaterialX",
"url": "https://github.com/AcademySoftwareFoundation/MaterialX/pull/1970"
}
|
gharchive/pull-request
|
Refine MaterialX render module build configuration
Add more refined control over building MaterialX render modules. Specifically USD only needs the core render module in some build configurations (iOS).
Also add a guard to ensure that MaterialXGenGlsl is built if MaterialXGenMsl is built, because of the dependency in the data library. USD uses MaterialX::GlslShaderGenerator::TARGET to locate those resources.
@ld-kerley I like the idea behind this proposal, though the split between MATERIALX_BUILD_RENDER and MATERIALX_BUILD_RENDER_CORE seems counterintuitive to me, as there's no corresponding module named MaterialXRenderCore.
As an alternative, what if we added a MATERIALX_BUILD_RENDER_PLATFORMS flag, defaulting to ON, with all platform-specific render modules gated by this flag?
This would allow USD developers to opt out of platform-specific render modules such as MaterialXRenderGlsl and MaterialXRenderOsl, without losing access to the MaterialXRender module that they still need.
Thanks for the updates, @ld-kerley, and this looks really promising! I'll take a closer look in upcoming days, and let's see if we can get this change ready to include in MaterialX v1.39.2.
|
2025-04-01T04:10:08.136263
| 2021-04-07T22:28:10
|
852875935
|
{
"authors": [
"lgritz",
"pellerington"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13296",
"repo": "AcademySoftwareFoundation/OpenShadingLanguage",
"url": "https://github.com/AcademySoftwareFoundation/OpenShadingLanguage/pull/1355"
}
|
gharchive/pull-request
|
Added farmhash for lower OpenImageIO versions
Older versions of OpenImageIO don't have the farmhash which is now nessisary for optix 7 strings.
To fix this, I added the file into into OpenShadingLangauge wrapped around version guards. Some existing functions had to be included again with OIIO_HOSTDEVICE added (Uint128Low64, Uint128High64, Uint128, CopyUint128, Hash128to64) so I wrapped them in the "inlined" namespace to prevent confilcts.
I made these changes so that I could build OSL with my version of OIIO, I am opening this pull request incase there is any interest in supporting it in the official branch. Potentially the VFX platform might have moved far enough by the time of the OSL release.
Hi, sorry for the delay in responding to this (the last couple weeks have been really busy for me).
I've backported these farmhash fixes to the OIIO release branch and it's included in OIIO <IP_ADDRESS>. The in-development OSL (and in particular the OptiX support) is technically considered "experimental/optional", it's a pre-release, so I think it's not unreasonable to expect it to be built against the latest OIIO release. I think I would prefer to require using a sufficiently recent OIIO release than to have to duplicate this code in the OSL codebase as well. But if people are really inconvenienced by that, I can certainly reconsider.
|
2025-04-01T04:10:08.139044
| 2024-09-02T22:44:05
|
2501651867
|
{
"authors": [
"cary-ilm"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13297",
"repo": "AcademySoftwareFoundation/OpenTimelineIO",
"url": "https://github.com/AcademySoftwareFoundation/OpenTimelineIO/issues/1791"
}
|
gharchive/issue
|
Consider signing release artifacts
The OpenSSF Best Practices Badge suggests signing release artifacts. Consider using OpenEXR's release-sign.yml workflow as a template. It's triggered on release creation and does these steps:
Runs get archive to generate a <release>.tar.gz artifact
Signs the <release>.tar.gz via sigstore
Uploads the resulting sigstore signature file along with the tarball.
Oops, this duplicates #1782, but with a bit more explicit suggestions!
|
2025-04-01T04:10:08.154579
| 2024-11-12T14:07:32
|
2652299381
|
{
"authors": [
"evelynnesher",
"sgbaird"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13298",
"repo": "AccelerationConsortium/ac-training-lab",
"url": "https://github.com/AccelerationConsortium/ac-training-lab/issues/113"
}
|
gharchive/issue
|
Autotrickler v4 - custom RS232 control with A&D scale
As a follow up to #20, the goal is to get a similar interface and IolT setup to that of the US Solid scale. Some related material at https://ac-microcourses.readthedocs.io/en/latest/courses/robotics/3.2-serial-communication.html. In particular:
✅ Read HR Analytical Balance Literature
This document provides an overview of the analytical balance we'll be interfacing with. Understanding the capabilities and specifications of the instrument is crucial for effective integration.
✅ Skim HR Analytical Balance Manual. Read Sections 15 and 17 closely.
Sections 15 and 17 detail the RS232C interface and the commands used to communicate with the balance. This information is essential for programming our microcontroller to interact with the balance correctly.
The connection at the back of the scale is similar. We have two scales (the 1 mg and the 0.1 mg resolution). The 0.1 mg resolution scale is at NRC. Let's focus on the 1 mg at UoT.
Iset up the RS232 connection with the scale. As shown below, I needed to unplug the auto trickler as it was plugged into the RS-232C serial interface.
Currently, it prints “ST” when the weight measurement is stable and “US” when the weight is actively changing and unstable.
That was fast! Great. There should be a second pico 2ch rs232 module available (I think in the same room?), so we can have one for the US solid scale and one for the A&D scale in the office.
Got HiveMQ set up and publishing with no issues, so I’m moving on to the app now. Do you want this one to have the autotrickler feature too? I’ll have to figure out a way to make it work since they both use the same port.
Nice! You'll use the same port, but adjust the MQTT topic structure to differentiate between the two devices. I think a separate app might be better, but I'm open to the idea of integrating it into the existing app.
I think you misunderstood, I meant the physical port on the scale, where both the Raspberry Pi (for the RS232 connection) and the AutoTrickler are connected. But as you can see, only one can connect at a time.
Ah, my bad 😅 The autotrickler can be left disconnected for now.
I set up the app in a private space on my personal account to test it out first. It is currently identical to the previous one but ready to be implemented into the AC Hugging Face Spaces.
App is public and scale is embedded into GatherTown.
Done!
I added a connection status bar that is identical to the one for the continuous output scale. I set the connection status refresh rate to 10 seconds, matching the other scale.
The graph now skips unstable data inputs, keeping it cleaner and only showing weights of properly placed items—not the weight of your hand as you place them.
Thanks! It's not urgent, but I'd rather that the unstable values be included to be more realistic to if someone had it in front of them. Likewise, there may be other uses for what would be considered unstable data.
|
2025-04-01T04:10:08.156509
| 2016-05-05T16:50:28
|
153273577
|
{
"authors": [
"dsingh07",
"nickdgriffin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13299",
"repo": "Accenture/adop-docker-compose",
"url": "https://github.com/Accenture/adop-docker-compose/pull/65"
}
|
gharchive/pull-request
|
Adding minor bug fixes
Bug fixes:
Certificate generation will now exit with a helpful message if bash is not run as admin (openssl command to create a cert will fail since it needs to source a .srl file)
Fix to allow "source ./conf/env.provider.sh" to be run if CONF_PROVIDER_DIR is not set, where we just use the current directory as a reference
LGTM.
|
2025-04-01T04:10:08.158444
| 2022-10-04T11:25:42
|
1396101078
|
{
"authors": [
"DannyCastroVentura",
"JoernBerkefeld"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13300",
"repo": "Accenture/sfmc-devtools-copado",
"url": "https://github.com/Accenture/sfmc-devtools-copado/pull/112"
}
|
gharchive/pull-request
|
#111: set skipInteraction properly to avoid mcdev's interactive wizard mode
PR details
What is the purpose of this pull request? (put an "X" next to an item)
[x] Bug fix
What changes did you make? (Give an overview)
closes #111
Is there anything you'd like reviewers to focus on?
the lines where ['npm install ${installer} --foreground-scripts', CONFIG.mcdev_exec + ' --version'] gets changed are just line break changes caused by a recent update to eslint/prettier
Ok, I thought I was changing the issue I've created, but instead was changing this one. Sory, mind bug
|
2025-04-01T04:10:08.167888
| 2024-02-19T15:55:14
|
2142689272
|
{
"authors": [
"DataTriny",
"mwcampbell"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13301",
"repo": "AccessKit/accesskit",
"url": "https://github.com/AccessKit/accesskit/pull/352"
}
|
gharchive/pull-request
|
fix: Factor out core AT-SPI translation layer
This will allow us to re-use the logic for translating between AccessKit and AT-SPI, in both the current Unix adapter and a new client-side library (for ATs like Orca) that is being developed as part of the new GNOME accessibility stack.
I don't expect to make any major changes to the accesskit_atspi_common API at this point, though I can't rule out needing to make a few more tweaks similar to my last few pushes.
Now I'm ready to show what I'm doing with this refactor. Note that the new GNOME accessibility stack is code-named Newton.
See this repository with a newton_atspi_compat crate and a Python binding on top:
https://gitlab.gnome.org/mwcampbell/newton_atspi_compat
and this proof-of-concept Orca modification that uses the Python module:
https://gitlab.gnome.org/mwcampbell/orca/tree/newton
It's still kind of a mess. I might be doing too much directly in the Python binding. And I do have to write a lot of conditionals in Orca, mostly because Orca is hard-coded to call functions on Atspi.Accessible, so I can't just take advantage of duck typing. And I need to document the steps to set up a complete working system, including my fork of Mutter (the compositor). But the Newton protocols are now working end to end, without passing through the AT-SPI D-Bus interface as I did in an earlier demo. And, back to this PR, my hope is that as we continue to develop the AccessKit AT-SPI backend, most of that work can be reused for the Newton project.
As soon as I add support for the action and component interfaces to the new "simplified" API I just pushed, I will be done adding things to this PR.
I don't plan to add or change any more in this PR. Once it passes review, I'm ready to merge it.
I wonder, though, if we should classify this as a refactor instead of a fix, even though the only way to do that is to use refactor!, which will trigger a breaking version increment on accesskit_unix. Really though, the main user of accesskit_unix so far is accesskit_winit, and the automatic dependency bump shouldn't cause a breaking version increment on accesskit_winit.
Yes, but then accesskit_unix would have been broken on the main branch in the meantime, leading to failing CI. I'm not sure that release-please really gives us a good option here.
I do have a breaking change in mind for accesskit_unix. To prepare the API to support both Newton and legacy AT-SPI in a single adapter, I believe we'll have to modify Adapter::new to take a RawWindowHandle, since the Newton backend requires a Wayland display and surface.
I don't see why this would have broken accesskit_unix? We'd just have a lot of duplicate code for a brief period of time.
Oh, you're right, of course. Then I can still pull the accesskit_unix changes out into a separate PR.
Doing it in two steps would also make it a bit simpler for me to rebase my unix-text branch.
Then I'll back out the accesskit_unix changes in this PR, and open the second one when this PR is merged. Before I do that, do you have any feedback about the design as a whole (that would require me to change both sides at once)?
I think you already addressed my concerns. The last remaining one would be from an outsider point of view: we know the dependency count of accesskit_unix is an issue for some downstream projects, yet here you are adding another one. I haven't looked into this, but I wouldn't be surprized if build times would slightly decrease because of that though.
It bothers me that Bevy isn't enabling the Unix adapter by default. But I can't let perceptions related to dependency count and binary size dictate all design decisions, especially when it comes to adding just one small crate. Anyway, once Newton is stabilized, the accesskit_unix adapter could let users disable AT-SPI and only enable Newton via features, thus eliminating the dependency on zbus.
OK, the changes to accesskit_unix are now in the unix-use-atspi-common branch, which I'll squash and rebase once this is merged.
Oh, one final comment but it's not a blocker: I wonder if the platforms directory really is the best place for this crate. I now have the same feeling for accesskit_winit actually.
I agree it's not ideal, but nothing better immediately comes to mind. We can always move it later without breaking anything user-visible.
Are we ready to merge this first PR?
|
2025-04-01T04:10:08.214859
| 2024-04-30T16:21:48
|
2271962172
|
{
"authors": [
"dhensle",
"i-am-sijia"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13302",
"repo": "ActivitySim/activitysim-prototype-mtc",
"url": "https://github.com/ActivitySim/activitysim-prototype-mtc/issues/12"
}
|
gharchive/issue
|
Full Scale Performance: Sharrow On
This is the issue to report on memory usage and runtime performance when using sharrow...
First ran sharrow compile with the following settings:
households_sample_size: 100
sharrow: test
Run completed in 76 minutes.
log_sh_compile.zip
Then ran in production mode
households_sample_size: 0 100 percent sample
sharrow: require
Run completed in 7.7 hours with a memory peak at about 163 GB in trip destination.
logs_sh_full.zip
Followed by multiprocessing
households_sample_size: 0 100 percent sample
sharrow: require
multiprocessing: True
num_processors: 24
Run completed in 110 minutes (1.8 hours).
log_sh_full_mp.zip
Ran with 100% households and sharrow on, single process.
Run completed in 1090.3 minutes (18.2 hours). This is much longer than the previous time posted above of 7.7 hours.
Current run was performed using PR #867 commit c9d4205.
log.zip
Timing statements comparing the old run above to this current run show large differences mainly in the destination models:
Will try again with the main branch of ActivitySim instead of PR 867 to see if that makes a difference.
Ran using an older environment that uses the current version of ActivitySim (main@bd48d3db), but has sharrow v2.8.2 instead of the previous run's main@8d63a66 (> v2.9.1). Numba was also older using 0.56.4 compared to 0.59.1.
The run results were pretty much exactly the same -- run time was 1080.3 minutes.
log.zip
One difference between these current set of runs and the 7.7 hour run above is the server. The 7.7 hour run was done on SANDAG's 1TB RAM, 40 Core machine. These were done on RSG's 500 GB RAM, 24 core machine.
Sharrow, single process, MTC extended model ran in 10.7 hours on WSP's 512 GB RAM, AMD server. Using everything the latest as of June 26. Memory peak 145 GB in trip destination.
ActivitySim: pr/867@c9d4205
Sharrow: v2.10.0
MTC: extended@a3da8bd
activitysim.log
timing_log.csv
Running the same tests as above and on the same machine, but using multiprocessing instead of multi-threading:
Comments:
16 and 24 core runs are incomplete due to https://github.com/ActivitySim/activitysim/issues/876
Saw roughly linear decreases in runtime for computationally intensive models going from 4 to 12 cores, but after that the gains decreased.
20 cores took longer than 12 cores. This is due to some models being slower (school escorting, school location, joint tour scheduling, etc), and increased time spend apportioning and coalescing all of the cores. However, this runtime difference was pretty minimal.
The runtime in the final activitysim.log file is slightly longer than the total in the timing_log.csv file across all runs. The difference increases with the number of cores.
|
2025-04-01T04:10:08.249868
| 2017-12-18T16:13:21
|
282934161
|
{
"authors": [
"CheezItMan",
"mcgmar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13303",
"repo": "Ada-C8/ada-trader",
"url": "https://github.com/Ada-C8/ada-trader/pull/38"
}
|
gharchive/pull-request
|
Carets - Maria - ada-trader
Ada Trader
Congratulations!
Comprehension Questions
Question
Answer
How do Backbone Views help structure your code?
The views help with separation of concerns and acted as a facilitator between data, like functions in the models, and between the DOM.
Did you use jQuery directly in your Views? How, and why?
I did via the $el which is a jQuery object. I did so so that the views would render.
What was an example of an event you triggered? Why did you need to trigger it?
An event I triggered was listing the quotes. That list change was triggered when a user clicked the buttons.
In what was is unit testing in JavaScript similar to unit testing in Ruby? In what ways is it different?
The tests are similar in that they had validations and you tested for each function that was built.
Ada Trader
What We're Looking For
Feature
Feedback
Core Requirements
Git hygiene
Not very many commits, good commit mesages
Comprehension questions
Check, however you didn't actually manually trigger any events.
Organization
Models and collections are defined in separate files
Check
Code that relies on the DOM is located in or called by $(document).ready
Check
Functionality
Quote prices change when clicking Buy and Sell
Check
The Trade History updates when buying and selling a quote
Check
A user can create an open order using the Order Entry Form
NOT WORKING
An open order removes itself from the open orders and updates the Trade History when fulfilled
MISSING
General
Has separate views for different parts of the app
Check
Uses events (listening/handling and triggering) to manage different behavior in views
You have the QuoteListView listening for changes in the collection and models and re-rendering. However You don't have the OrderListView working or listening to the form, or any component listening for changes in the Quote instances.
Practices good standards for views (ES6 syntax, doesn't directly use jQuery, returns this in render)
Check, but you're not utilizing the form for the OrderListView.
Error handling for the Order Entry Form
MISSING
Testing
Has unit tests for models
MISSING
Overall
You've got a lot incomplete here. You set up a view for the Quotes and QuoteListView perfectly and append trades to the list well. You didn't complete the OrderView or OrderListView or complete any testing. You can see some notes I put in your code as to what could be done to get it working. The biggest error is the fact that both QuoteListView and OrderListView are both tied to the same main html element which means they could theoretically interfere with each other. See my comments in your code.
|
2025-04-01T04:10:08.279264
| 2016-02-10T03:19:38
|
132597357
|
{
"authors": [
"SirePi",
"brogine",
"ilexp"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13304",
"repo": "AdamsLair/duality",
"url": "https://github.com/AdamsLair/duality/issues/276"
}
|
gharchive/issue
|
Rectangle shape for rigid collision detection
I'm currently testing the framework, which i consider very simple and easy to learn as about now.
I'm trying to make a rpg-styled game but i think polygon is not very accurate in this case.
I think Rectangle must be added when editing rigid bodies.
That'd solve a few of my problems.
Thanks Adam for your work.
Well, a Rectangle is in fact a Polygon..
if you check the Rigidbody in the object inspector panel, you will see that you can edit the vertices individually, so it's easier to position them to form a perfect rectangle.
Basically what SirePi said: You can express a rectangle using the existing polygon shape.
Still, it would of course be more convenient if there was an actual "Rect" control, so you would be able to quickly create rectangular polygons. I'm aware that the current RigidBody Editor is not ideal in some cases (see issue #54), but this is currently a rather low-priority task, since all the typical use cases can be accomplished - just sometimes not in the most convenient way.
Other than that, I'm currently working on a Tilemaps plugin, which you can find more about in issue #249 or in the development report forum thread. It won't be released (or officially supported) anytime soon, but in case you're curious, the latest state is available in the tilemap_wip branch on GitHub. From a more practical perspective, there are a lot of tilemap-based projects on the forum, so I imagine you could get more advice there, if you need help.
Yes, but if you have a scene with 50 polygons, you have to draw everyone of them by hand and then edit them individually, there's some time there.
If you try to check collisions between hand-drawed polygons, you don't get what you are expecting.
I just think this as a minor but useful feature.
Yes, but if you have a scene with 50 polygons, you have to draw everyone of them by hand and then edit them individually, there's some time there.
Hmm, I think I already answered this, see here:
Still, it would of course be more convenient if there was an actual "Rect" control, so you would be able to quickly create rectangular polygons. I'm aware that the current RigidBody Editor is not ideal in some cases (see issue #54), but this is currently a rather low-priority task, since all the typical use cases can be accomplished - just sometimes not in the most convenient way.
So yes, you definitely have a point that I acknowledge and it would be more convenient to have a distinct Rect tool, but we already have an issue for this - even though it's currently a low-priority task. :)
For now, you could maybe copy-paste your RigidBodies (Perform a right-click-drag from the source to the target GameObject), use Prefabs, write a custom editor plugin with a script for rect shape generation or write a custom Component that automatically generates a RigidBody for tile-like objects - whichever suits your use case best.
|
2025-04-01T04:10:08.283001
| 2024-04-14T01:12:27
|
2241848170
|
{
"authors": [
"illuhad",
"nilsfriess",
"ouankou"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13305",
"repo": "AdaptiveCpp/AdaptiveCpp",
"url": "https://github.com/AdaptiveCpp/AdaptiveCpp/pull/1436"
}
|
gharchive/pull-request
|
Replace deprecated APIs in the src and tests
While building ACPP from source and running the built-in tests, I noticed a few warnings caused by deprecated APIs.
bool llvm::StringRef::startswith(llvm::StringRef) const (in the source)
warning: ‘bool llvm::StringRef::startswith(llvm::StringRef) const’ is deprecated: Use starts_with instead [-Wdeprecated-declarations]
Atomic accessors in SYCL 2020 (in the tests)
warning: 'operator[]' is deprecated: Atomic accessors are deprecated as of SYCL 2020 [-Wdeprecated-declarations]
'get_count' in SYCL 2020 (in the tests)
warning: 'get_count' is deprecated: renamed to 'size' in SYCL 2020 Specification [-Wdeprecated-declarations]
I'm still new to SYCL, so please let me know if the proposed change contains any errors. Thanks!
Re the deprecated SYCL functions: Since these are only deprecated and not (yet) removed from the SYCL spec, I think we should still include them in the tests. Ideally we should test both the deprecated function and the "replacement". Maybe we can consider disable deprecation warnings for the tests to reduce the noise
Re the deprecated SYCL functions: Since these are only deprecated and not (yet) removed from the SYCL spec, I think we should still include them in the tests. Ideally we should test both the deprecated function and the "replacement".
Yes. They are officially still part of SYCL, and SYCL CTS also tests for deprecated APIs.
Thanks for the clarifications! I'll close the PR since those deprecated APIs must still be kept and tested.
|
2025-04-01T04:10:08.445221
| 2021-11-11T20:16:18
|
1051326512
|
{
"authors": [
"ameshkov",
"dnmTX",
"krystian3w",
"sfionov",
"slavaleleka"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13306",
"repo": "AdguardTeam/CoreLibs",
"url": "https://github.com/AdguardTeam/CoreLibs/issues/1550"
}
|
gharchive/issue
|
Add regexp support for $domain modifier
now there is no chance to create single cosmetic rule for hiding the same element on such domains:
example1.com, example2.com, example3.org
it would be nice if $domain modifier supports regular expressions, just like $path does
something like this:
[$domain=/example[0-9]\.(com|org)/]##.ad
Why limit just to the cosmetic rules? If we do it here, we should extend it to basic rules as well.
@sfionov can we include this in v1.10 for instance?
Yes, assigned to 1.10
https://github.com/AdguardTeam/tsurlfilter/issues/41
https://github.com/AdguardTeam/tsurlfilter/issues/58
https://github.com/AdguardTeam/AdguardKnowledgeBase/pull/1109
@AdguardTeam maybe unrelated question but what about redirects? For example links from voe.sx are allways redirected to another domain and they change those domains very frequently. Here is a example:
Original Link: https://voe.sx/ffgecklhux79 redirects(as of now) to: https://tubelessceliolymph.com/ffgecklhux79
So basically rules that applied for voe.sx simply don't work. And domains like tubelessceliolymph.com are disposable and replaced every week or so and user(s) like me have to reaply same rules to the new domain.
Any workaournd(s) for this?
uBo implemented regex, if someone trust something like:
/bad-js.$domain=/[a-z-]{18}\.[\w]{2,20}/
/[a-z-]{18}\.[\w]{2,20}/##.bad-class
but fail for , in cosmetic/scriptlet filters and | in network filters if gorhill rejected repair:
https://github.com/uBlockOrigin/uBlock-issues/discussions/2234#discussioncomment-5403472
https://github.com/AdguardTeam/AGLint/issues/161
|
2025-04-01T04:10:08.446357
| 2018-10-17T12:56:32
|
371059906
|
{
"authors": [
"sfionov",
"szolin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13307",
"repo": "AdguardTeam/CoreLibs",
"url": "https://github.com/AdguardTeam/CoreLibs/issues/545"
}
|
gharchive/issue
|
Don't send onBrowserApi callback on wrong token
Also, it will be better if, in case of invalid token, message will say about this and will have DEBUG log level, not ERROR.
ERROR AGWebSocketHandler *7: json request parse: (6) bad value last value:d83cfbe7-bd3b-4d27-9364-a949c7462ba4
/core/pull-requests/747
|
2025-04-01T04:10:08.449726
| 2019-06-26T21:56:47
|
461199688
|
{
"authors": [
"ameshkov"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13308",
"repo": "AdguardTeam/CoreLibs",
"url": "https://github.com/AdguardTeam/CoreLibs/issues/986"
}
|
gharchive/issue
|
Wrong web page encoding or Automatic download page files
@Chandler-Lu commented on Wed Jun 26 2019
When I visit a page without HTTPS, just like http://bbs.pcbeta.com/forum-win10-1.html, the web page may display garbled characters or Automatic download page file (forum-win10-1.html). Then I need to refresh some times, the page can display normally.
Expected behavior
display normally
Actual behavior
garbled characters, lose css style, automatic download page file
Your environment
Environment name and version: (e.g. Chrome 59): safari 12.1.1 (146<IP_ADDRESS>.1)
Any specific potentially conflicting software installed: (e.g. antiviruses, firewalls, traffic counters, cleaners): ClashX
Cannot reproduce this anymore, seems to be fixed
|
2025-04-01T04:10:08.502932
| 2024-05-19T16:42:41
|
2304691110
|
{
"authors": [
"Admonstrator"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13309",
"repo": "Admonstrator/glinet-uptime",
"url": "https://github.com/Admonstrator/glinet-uptime/issues/65"
}
|
gharchive/issue
|
⚠️ GLDDNS Server (ns2.glddns.com) has degraded performance
In b982428, GLDDNS Server (ns2.glddns.com) (ns2.glddns.com) experienced degraded performance:
HTTP code: 200
Response time: 495 ms
Resolved: GLDDNS Server (ns2.glddns.com) performance has improved in b5674ea after 6 minutes.
|
2025-04-01T04:10:08.537482
| 2020-05-01T16:53:01
|
610852042
|
{
"authors": [
"joerghoh"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13310",
"repo": "Adobe-Consulting-Services/acs-aem-commons",
"url": "https://github.com/Adobe-Consulting-Services/acs-aem-commons/issues/2282"
}
|
gharchive/issue
|
Do not start services on AEMaaCS instances if they are not working there
In the documentation a number of features are listed already as "do not use them when you are deploying on AEM as a CloudService".
Right now this is not technically enforced, but with the RequireAem feature (see https://github.com/Adobe-Consulting-Services/acs-aem-commons/blob/b391b2455bb1aac018183ae4b76794379ebb8a59/bundle/src/main/java/com/adobe/acs/commons/util/RequireAem.java) it can be enforced quite easily.
Right now these features are tagged as "do not run within AEMaaCS":
bulk-workflow-manager
ensure-oak-index
healthchecks/smtp
healthchecks/status-emailer
healthchecks
json-event-logger
versioned-clientlibs
workflow-processes/assets-rendition-matter
workflow-processes/assets-watermark-process
workflow-processes/brand-portal
file-fetcher
mcp-tools/asset-ingestion/asset-ingestor
mcp-tools/asset-ingestion
mcp-tools/asset-ingestion/s3-asset-ingestor
mcp-tools/asset-ingestion/url-asset-ingestor
mcp-tools/refresh-folder-thumbnails
redirect-map-manager
remote-assets
(I have extracted this list from the official documentation)
see #2285
#2285 has been merged, closing this one.
|
2025-04-01T04:10:08.544968
| 2020-03-25T14:07:20
|
587722286
|
{
"authors": [
"adamcin",
"davidjgonzalez",
"justinedelson",
"kwin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13311",
"repo": "Adobe-Consulting-Services/acs-aem-commons",
"url": "https://github.com/Adobe-Consulting-Services/acs-aem-commons/pull/2233"
}
|
gharchive/pull-request
|
Feature/2232 - Removes META-INF/vault files in SCM; Minor sonarqube fixes
Removes the old properties.txt and friends (cnd, etc.) from SCM.
Fixed a few CM related quality gate checks to allow it to pass/deploy to CM.
@davidjgonzalez this is a good idea. have you done a comparison between the generated META-INF/vault/* files between the last release and with these changes? If not, I'll try to script that and post the results here.
@justinedelson adding the generated files for ui.apps and ui.content
here's the ui.apps properties.txt for a quick look
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd">
<properties>
<comment>ACS AEM Commons UI.Apps Package</comment>
<entry key="acHandling">merge</entry>
<entry key="created">2020-03-25T11:31:00.424-0400</entry>
<entry key="groupId">com.adobe.acs</entry>
<entry key="description">ACS AEM Commons content package containing code (/apps).</entry>
<entry key="version">4.4.3-SNAPSHOT</entry>
<entry key="requiresRoot">false</entry>
<entry key="dependencies">day/cq60/product:cq-content:6.3.64</entry>
<entry key="path">/etc/packages/adobe/consulting/acs-aem-commons-ui.apps.zip</entry>
<entry key="createdBy">davidg</entry>
<entry key="name">acs-aem-commons-ui.apps</entry>
<entry key="artifactId">acs-aem-commons-ui.apps</entry>
<entry key="group">adobe/consulting</entry>
</properties>
ui.apps.meta-inf-vault.zip
ui.content.meta-inf-vault.zip
side by side diff: https://gist.github.com/justinedelson/8392242adf39ef7e746bf273e4a5df98
base don the diff, looks like a benign change but keeps the code repo a little cleaner.
Am planning on cutting a release for @wimsymons since we've been putting him off for too long, and didnt want to get us caught w/ that possible properties.txt issue.
I think this breaks oakpal. I now get errors like
[INFO] --- oakpal-maven-plugin:1.5.1:scan (oakpal-scan) @ acs-aem-commons-ui.apps ---
[INFO] Found a new index node [reference]. Reindexing is requested
[INFO] Reindexing will be performed for following indexes: [/oak:index/uuid, /oak:index/reference, /oak:index/nodetype]
[INFO] Indexing report
- /oak:index/uuid*(0)
- /oak:index/reference*(0)
- /oak:index/nodetype*(1258)
[INFO] Reindexing completed
[INFO] Reindexing will be performed for following indexes: [/oak:index/principalName, /oak:index/authorizableId, /oak:index/acPrincipalName, /oak:index/repMembers]
[INFO] Indexing report
- /oak:index/principalName*(2)
- /oak:index/authorizableId*(2)
- /oak:index/acPrincipalName*(0)
- /oak:index/repMembers*(0)
[INFO] Reindexing completed
[ERROR] Error during processing of /apps/acs-commons/touchui-widgets/icon-picker: javax.jcr.nodetype.NoSuchNodeTypeException: Node type cq:ClientLibraryFolder does not exist
[ERROR] E /apps/acs-commons/touchui-widgets/icon-picker (javax.jcr.nodetype.NoSuchNodeTypeException: Node type cq:ClientLibraryFolder does not exist)
That is an issue though in oakpal which should be fixed there, as we shouldn't include any irrelevant metadata in packages (like the cnd for nodetypes which are always existing already).
@kwin Sorry for just now seeing your question, as it looks like this issue has already been resolved. There are a couple ways to provide the platform CNDs to oakpal other than maintaining redundant CNDs in the package artifacts. One way is to use the cndNames parameter to reference specific CND resources on the test classpath. Another is to enable resolution by detecting Sling-Nodetypes manifest headers in test-scope dependencies. This enabled by setting slingNodeTypes to true. This option might be more convenient in the future now that Adobe's uber-jars (starting with 6.5.0) include the platform nodetypes. Finally, it is also possible to serialize the nodetypes into the jcrNodetypes json property of plans and checklists, but this method makes it harder to manually maintain those files if we are talking about the full set of aem nodetypes, rather than just nodetypes directly referenced by progress check implementations or forcedRoot primary/mixin types.
|
2025-04-01T04:10:08.551647
| 2023-02-01T20:06:03
|
1566747100
|
{
"authors": [
"synox",
"theManikJindal"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13312",
"repo": "AdobeDocs/adobe-dev-console",
"url": "https://github.com/AdobeDocs/adobe-dev-console/issues/65"
}
|
gharchive/issue
|
The docs don't mention where to find the api_key
Issue in /src/pages/guides/authentication/APIKeyIntegration.md
When the API has been successfully connected, you will be able to access the newly generated API Key.
But there is no field with the label API key on the console page. Maybe you mean CLIENTID?
Yes API key and client id mean the same thing
|
2025-04-01T04:10:08.553570
| 2024-05-08T01:02:33
|
2284469293
|
{
"authors": [
"Alicesnk",
"jtwood426"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13313",
"repo": "AdobeDocs/campaign.en",
"url": "https://github.com/AdobeDocs/campaign.en/issues/7"
}
|
gharchive/issue
|
Text Presentation Error
Issue in ./help/v8/connect/ac-ajo.md
The bottom of the page shows this raw text:
[Learn more in Journey Optimizer documentation]https://experienceleague.adobe.com/en/docs/journey-optimizer/using/orchestrate-journeys/about-journey-building/using-adobe-campaign-v7-v8){target=“_blank”}.
Thanks for raising this - we'll update shortly
Internal reference > DOCAC-10478
|
2025-04-01T04:10:08.559786
| 2021-09-07T15:37:07
|
990113140
|
{
"authors": [
"3phase",
"DaveInMatrix"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13314",
"repo": "AdobeDocs/uxp-xd",
"url": "https://github.com/AdobeDocs/uxp-xd/issues/109"
}
|
gharchive/issue
|
Manifest link leads to old manifest version
Issue in /src/pages/develop/plugin-development/index.md
Hi,
I'd like to ask if (or just point out) there's any specific reason why Plugin Development > Plugin Structure > Manifest leads to Manifest v3 link (which would even explicitly point to v3 in the URL), while the latest is v4? This was little obfuscating at least to me to me.
Regards,
Petko
This is deliberate as XD currently only works with the v3 manifest. We are in the process of adding support for v4 features.
|
2025-04-01T04:10:08.562021
| 2019-12-31T17:14:38
|
544233533
|
{
"authors": [
"ericdrobinson",
"pklaschka"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13315",
"repo": "AdobeXD/typings",
"url": "https://github.com/AdobeXD/typings/issues/72"
}
|
gharchive/issue
|
Suggestion: Call out custom type name overrides in ReadMe
PR #70 changed the Selection type declaration to be XDSelection. It may be helpful for users to have this explicitly called out (with explanation for why) in the ReadMe.
Suggestion: Add a new section to the ReadMe file (called something like "Notes" or "Errata") that calls out customizations like this and includes a blurb for their raison d'être.
@ericdrobinson The funny thing is: I wanted to implement this tomorrow, anyway. You're too fast at creating issues :stuck_out_tongue_winking_eye:.
I'll also (before creating the release) rename the LinearGradient (I've waited for too long for the docs to update this :neutral_face:), cf. #41, and will then include these as "exceptions" in the section where we've specified that the typings follow the docs.
|
2025-04-01T04:10:08.563489
| 2018-07-03T14:59:19
|
337937555
|
{
"authors": [
"sxa555"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13316",
"repo": "AdoptOpenJDK/openjdk-build",
"url": "https://github.com/AdoptOpenJDK/openjdk-build/issues/394"
}
|
gharchive/issue
|
Add arm32 into the nightly openjdk10 pipelines
Now that https://github.com/AdoptOpenJDK/openjdk-build/issues/332 is no longer a problem, we can add these back into the nightly pipelines :-)
Added in https://github.com/AdoptOpenJDK/openjdk-build/commit/03128d78f590d09c9fa2bb8ac8146cff9df3d60c
|
2025-04-01T04:10:08.567401
| 2018-04-11T18:24:40
|
313436379
|
{
"authors": [
"karianna",
"vsebe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13317",
"repo": "AdoptOpenJDK/openjdk-infrastructure",
"url": "https://github.com/AdoptOpenJDK/openjdk-infrastructure/pull/340"
}
|
gharchive/pull-request
|
Add xz compression to CentOS build tools
Signed-off-by: Violeta Sebe<EMAIL_ADDRESS>
What's this required for?
It is required to unpack git source: git-2.15.0.tar.xz (in Extract git source task.
|
2025-04-01T04:10:08.575649
| 2021-04-17T08:11:03
|
860355019
|
{
"authors": [
"Gaminguide1000",
"Silfryi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13318",
"repo": "Advanced-Rocketry/AdvancedRocketry",
"url": "https://github.com/Advanced-Rocketry/AdvancedRocketry/issues/2062"
}
|
gharchive/issue
|
Advanced Rocketry planet data save bug
Version of Advanced Rocketry
AdvancedRocketry-1.12.2-2.0.0-242-universal
Have you verified this is an issue in the latest unstable build
[x] Y
Version of LibVulpes
LibVulpes-1.12.2-0.4.2-83-universal
Version of Minecraft
1.12.2
Does this occur without other mods installed
[x] Y
If Y, what is the MINIMUM set of mods required.
Idk, but i think only AdvRocketry&LibVulpes
Crash report and log, or VisualVM snapshot (if applicable)
http://pastebin.com is a good place to put them
crash reports that are put in the issue itself are impossible to read
If it's an issue involving lag, please submit a VisualVM sampler snapshot. Mods like lag goggles cannot provide all of the information needed to debug lag in any useful capacity.
For more information on how to create a snapshot, please see "How to Run CPU Sampling" here. Then export the snapshot as an .nps. You may have to zip it since github is weird, then please upload it here.
Description of the problem
First of all: I have a custom PlanetDefs.xml (not mine). So, always after a while, the server console says: "[Server thread / FATAL] [advancedrocketry]: An error has occurred saving planet data,
this can happen if another mod causes the game to crash during game load. If the game has fully loaded, then this is a serious error, Advanced Rocketry data has not been saved. "There are no conflicts between different mods, nor does the game / server crash when I start it. should that be a "serious error".
Yes, because it means your planetDefs has a high chance of becoming so corrupt you cannot load your world.
|
2025-04-01T04:10:08.580412
| 2023-09-13T11:30:45
|
1894330245
|
{
"authors": [
"StefanGreve"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13319",
"repo": "Advanced-Systems/anonpy",
"url": "https://github.com/Advanced-Systems/anonpy/issues/33"
}
|
gharchive/issue
|
Improve Publishing Workflow for Development
Discussed in https://github.com/Advanced-Systems/anonpy/discussions/4
Originally posted by StefanGreve August 9, 2023
Preamble
Editing the workflow file responsible for the publishing process in a live environment can be a little bit disheartening if there are errors, to put it mildly.
Luckily for us, PyPI provides the facilities to test builds on https://test.pypi.org/. Documentation about that can be found here:
https://packaging.python.org/en/latest/guides/using-testpypi/
Requirements
Essentially, we need a workflow file that can be manually triggered with the workflow_dispatch event on GitHub actions which requires that the committer updates the __version__ variable in src/anonfile/anonfile.py to major.minor.patch-buildnumber. This makes it much easier to test the release workflow during development, and avoids scenarios in which we have to publish a patch because something went wrong with python-publish.yml.
The secrets for test.pypi.org need to be configured by the project administrator (@StefanGreve) through the repository settings panel. I would propose calling them TEST_PYPI_USERNAME and TEST_PYPI_PASSWORD, respectively.
Planning Status
Scheduled for milestone 1.0.0.
Closed as duplicate, see also #30
|
2025-04-01T04:10:08.612119
| 2016-07-24T09:50:49
|
167226012
|
{
"authors": [
"Grover-c13",
"Spl3en",
"rubenvereecken"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13320",
"repo": "AeonLucid/POGOProtos",
"url": "https://github.com/AeonLucid/POGOProtos/pull/52"
}
|
gharchive/pull-request
| |
2025-04-01T04:10:08.614814
| 2024-08-16T22:12:43
|
2471108036
|
{
"authors": [
"593Coding",
"Aeonss"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13321",
"repo": "Aeonss/BubbleBlaster",
"url": "https://github.com/Aeonss/BubbleBlaster/issues/35"
}
|
gharchive/issue
|
I've this error
I ran pip install -r requirements.txt
and python bubbleblaster.py
Try the newest 2.0.1 version first, see if that works. Run the "INSTALL_DEPENDENCIES.bat" first and then the "bubbleblaster.bat"
I included a exe version... try if it doesn't work
I included a exe version... try if it doesn't work
I found the problem, you need to put this dll on system32 in windows11
https://www.dllme.com/dll/files/libomp140_x86_64/00637fe34a6043031c9ae4c6cf0a891d/download
|
2025-04-01T04:10:08.619359
| 2016-11-16T03:11:00
|
189571066
|
{
"authors": [
"boisy",
"mitchmindtree"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13322",
"repo": "Affectiva/affdexme-osx",
"url": "https://github.com/Affectiva/affdexme-osx/issues/11"
}
|
gharchive/issue
|
How to modify example to detect ethnicity?
Hey there, after a quick look I couldn't find any documentation on how to set up ethnicity detection - would you mind providing any tips on how to modify the example to do this? Or maybe provide a link to some docs that demonstrate this? Thanks!
Hi Mitch,
We're not exposing ethnicity just yet in our SDK. It's still undergoing validation, but we expect to have it in a future release.
@boisy legend, thanks a lot for this :+1: I ended up getting this working and the results were generally (roughly) correct 😸
|
2025-04-01T04:10:08.626991
| 2022-11-14T04:03:21
|
1447343156
|
{
"authors": [
"AgentRev",
"manav0619"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13323",
"repo": "AgentRev/WindowsAppsUnfukker",
"url": "https://github.com/AgentRev/WindowsAppsUnfukker/issues/13"
}
|
gharchive/issue
|
PsExec could not start powershell
Hello,
I am running W11 22H2. I have been trying to fix my WindowsApps folder that I messed up permissions for and came across your script. However, it's not working for some reason on my PC. I keep getting the "PsExec could not start powershell…" error. I was wondering if you could assist me a bit. Any help will be truly appreciated.
Windows PowerShell
Copyright (C) Microsoft Corporation. All rights reserved.
Install the latest PowerShell for new features and improvements! https://aka.ms/PSWindows
PS C:\Users\manav> cd J:\Users\manav\Downloads\PSTools
PS J:\Users\manav\Downloads\PSTools> .\psexec.exe -s -i powershell -ExecutionPolicy Bypass -File "J:\Users\manav\Downloads\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1"
PsExec v2.4 - Execute processes remotely
Copyright (C) 2001-2022 Mark Russinovich
Sysinternals - www.sysinternals.com
PsExec could not start powershell on DESKTOP-99N6EMR:
The system cannot find the file specified.
It means Powershell cannot access the path to the script. In my experience, the usual culprit is network drives. I would say, try to move the script to a different drive.
@AgentRev Hey, thanks for responding. I don't actually have any network drives. Just two SSDs and two HDDs, all local. I have tried moving the script to different locations on every drive, but it's the exact same error every time.
If you'd like to see any log file or system information that might help you identify the issue, I'd be happy to provide it. I really have no idea what's going wrong.
At that point, it's beyond my reach. You could use Process Monitor from PsTools to investigate where is PsExec trying to look and why it's not resolving it. Other people have also recommended PAExec instead, maybe give that a try.
It wasn't PowerShell being unable to access the path to the script; it was rather PsExec being unable to access PowerShell.
I just added the full path to powershell.exe in the command, and it ended up working.
So, instead of
.\psexec.exe -s -i powershell -ExecutionPolicy Bypass -File "C:\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1
I used
.\psexec.exe -s -i "C:\Windows\System32\WindowsPowerShell\v1.0\powershell.exe" -ExecutionPolicy Bypass -File "C:\WindowsAppsUnfukker-main\WindowsAppsUnfukker.ps1
(I guess my environment variables are messed up as well? Because I had to add the System32 folder to "Path" under User Variables for takeown command to work when I was manually trying to fix the permissions earlier.)
|
2025-04-01T04:10:08.628607
| 2023-08-03T21:04:05
|
1835733238
|
{
"authors": [
"albnunes",
"mmabrouk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13324",
"repo": "Agenta-AI/agenta",
"url": "https://github.com/Agenta-AI/agenta/pull/224"
}
|
gharchive/pull-request
|
Save test set update
In this pull request, I removed the requested functionality mentioned in issue #220, and everything seems to be working.
I hope to contribute even more effectively to this project in the future.
Closes #220
Thanks for the PR @albnunes. The issue has been closed in #265
@all-contributors please add @albnunes for bug fix
|
2025-04-01T04:10:08.702846
| 2021-08-05T06:16:25
|
961484352
|
{
"authors": [
"paulmorar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13325",
"repo": "Aguafrommars/azure-devops-extensions",
"url": "https://github.com/Aguafrommars/azure-devops-extensions/pull/35"
}
|
gharchive/pull-request
|
Fix typo in commit types
Currently, there are quite a few typos in the commit type entries, and this PR addresses that.
@aguacongas can you take a look over this one?
|
2025-04-01T04:10:08.707540
| 2024-03-29T08:36:18
|
2214928290
|
{
"authors": [
"AgustinSRG",
"derekcmini"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13326",
"repo": "AgustinSRG/crystals-dilithium-js",
"url": "https://github.com/AgustinSRG/crystals-dilithium-js/issues/1"
}
|
gharchive/issue
|
After the web page loads dilithium.js, an error occurs when executing DilithiumKeyPair.generate(2)
Hi,
I use in my html code, and
Part of the js code in my frontend:
const level = DilithiumAlgorithm.DilithiumLevel.get(2);
const keyPair = DilithiumAlgorithm.DilithiumKeyPair.generate(level);
Post-processing when web page is reloaded
ReferenceError: Buffer is not defined
at s (dilithium.js:2:63339)
at l.generate (dilithium.js:2:42255)
Am I doing anything wrong?
Hello. Thanks for reporting this issue.
It seems Webpack does not include a Buffer polyfill into the javascript library bundle by default. I changed the webpack configuration to include it. Now it should be working without any extra dependencies in the browser (https://github.com/AgustinSRG/crystals-dilithium-js/releases/tag/v1.1.1).
I also added a basic test page (https://agustinsrg.github.io/crystals-dilithium-js/test.html) to test the browser bundle.
I'm marking the issue as completed. In case you find any other issues with this library, feel free to open more issues in the future.
|
2025-04-01T04:10:08.721542
| 2015-08-28T08:34:33
|
103671445
|
{
"authors": [
"sbuffo"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13327",
"repo": "AirVantage/av-sched",
"url": "https://github.com/AirVantage/av-sched/issues/25"
}
|
gharchive/issue
|
Improve configuration properties
For prefomance reason : change the pools size configuration, add them as deploy parameters.
PR : https://github.com/AirVantage/av-sched/pull/26
|
2025-04-01T04:10:08.723052
| 2021-09-15T01:27:10
|
996567892
|
{
"authors": [
"GH3336699",
"Satoshiisgod"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13328",
"repo": "Aircoin-official/AirCash",
"url": "https://github.com/Aircoin-official/AirCash/issues/73"
}
|
gharchive/issue
|
Apply for Air cash test
Hello, I want to apply for the air coin test.
I used to be the general manager of IT projects. I started to understand the blockchain in 2016 and later became a merchant of other exchanges. I still understand this aspect quite well.
I am very interested in Air Cash, and I hereby apply, hoping to become a member of Air Cash and contribute to Air Cash.
Anyone who applied for the AirCash testing please join the discord channel for test volunteers.
https://discord.gg/eRSY7NtxZA
|
2025-04-01T04:10:08.728032
| 2021-04-18T00:22:29
|
860542121
|
{
"authors": [
"Aircoookie",
"Chris210418",
"LeonardoD91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13329",
"repo": "Aircoookie/Espalexa",
"url": "https://github.com/Aircoookie/Espalexa/issues/175"
}
|
gharchive/issue
|
Ghost Devices
Alexa keeps discovering a few devices that i only used once for testing.
Deleting them in the Alexa App doesnt work half the time and they prevent the discovery of the last device in my final Program.
At least thats what i think happens.
The weird thing is they are discovered even if my esp32 is turned off.
Maybe thats more of an Alexa problem but i dont know how to fix it.
I really enjoy the stuff that works.
Thanks for making this.
Help would be appreciated.
Hi, I can definitely reproduce this issue.
Originally I believed this issue only affected my account but it appears to be a general problem when using Espalexa.
Devices always get "discovered" after deleting, even if said device has not been online for a year.
Deleting all devices/groups/schedules from the alexa app did not help either.
I have not yet been able to conclusively determine if the ghost devices are saved in the Echo hardware or in the Amazon account and am sadly unable to share any method that permanently deletes devices.
One slightly random question - do you have a genuine Philips Hue bridge in your network and/or installed the associated Alexa skill?
Hello,
thanks for answering.
I have been able to set up the last device in my Program. It was an error in my code.
Yes i have a hue bridge in my home and the hue skill installed. But having it connected and or the skill installed seems to have no impact on the non existing yet discovered devices.
A few test have shown that the "Ghost Devices" get rediscovered only if the right Alexa Device is connected to the network.
I think its the one that was closest to the esp when i first testet the program.
All of the rest is working brilliantly quick and reliable.
Thanks again for making this and thanks for helping a lot of not so code talented people like me! :P
I have the same issue, how do i stop this ghost devices from popping, i tried to erase flash the esp form esptool in terminal. I flashed many other project on the same esp chip but i always have this devices appearing, maybe alexa remind the mac address of the device and keep associate it to the old device names, i tried also to factory resety the alexa echo dot where those ghost devices connected forst time, but nothing changed. Myabe i have to reset the all alexa devices in my house?
I appreciate any help
|
2025-04-01T04:10:08.792090
| 2023-04-25T08:45:44
|
1682716012
|
{
"authors": [
"nclsHart",
"notFloran"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13331",
"repo": "Akollade/mac-playbook",
"url": "https://github.com/Akollade/mac-playbook/pull/184"
}
|
gharchive/pull-request
|
:bug: Fix issue between homebrew/sudo and ansible become
Fix #181
Related to #179 : the fix I made broke the password pass between ansible and sudo.
I've got an issue which seems to be related to brew run as sudo (during dev setup tags=nginx).
fatal: [<IP_ADDRESS>]: FAILED! => changed=true
cmd:
- brew
- services
- restart
- nginx
delta: '0:00:00.830175'
end: '2023-04-25 14:06:08.832309'
msg: non-zero return code
rc: 1
start: '2023-04-25 14:06:08.002134'
stderr: 'Error: Need to download https://formulae.brew.sh/api/formula.jws.json but cannot as root! Run `brew update` without `sudo` first then try again.'
stderr_lines: <omitted>
stdout: ''
stdout_lines: <omitted>
I tried to run brew update and run again the setup command but without success.
I've got an issue which seems to be related to brew run as sudo (during dev setup tags=nginx).
fatal: [<IP_ADDRESS>]: FAILED! => changed=true
cmd:
- brew
- services
- restart
- nginx
delta: '0:00:00.830175'
end: '2023-04-25 14:06:08.832309'
msg: non-zero return code
rc: 1
start: '2023-04-25 14:06:08.002134'
stderr: 'Error: Need to download https://formulae.brew.sh/api/formula.jws.json but cannot as root! Run `brew update` without `sudo` first then try again.'
stderr_lines: <omitted>
stdout: ''
stdout_lines: <omitted>
I tried to run brew update and run again the setup command but without success.
Can you run dev dotiles ?
Can you run dev dotiles ?
Done. But no changes :/
It works... I just forget to reload my zsh config.
|
2025-04-01T04:10:08.810921
| 2020-02-06T10:05:23
|
560904531
|
{
"authors": [
"alexander-mai",
"farfromrefug"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13332",
"repo": "Akylas/nativescript-material-components",
"url": "https://github.com/Akylas/nativescript-material-components/pull/110"
}
|
gharchive/pull-request
|
feat: added features for bottomsheet
dismissOnDraggingDownSheet
added example on demo
@alexander-mai thanks ! will try to test it today
I could have created examples for Angular and Vue, but I can't get the demo apps up and running right away.
@farfromrefug What do you think, when this brach can be merged?
I made some improvements on top of this change and I would like the create a new pull request for it.
@alexander-mai sorry i actually forgot to merge this :s Thanks a lot for your work!
Will make a release soon
|
2025-04-01T04:10:08.965649
| 2020-10-23T18:54:55
|
728449189
|
{
"authors": [
"blackfalcon",
"ecommerceUXdesign"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13333",
"repo": "AlaskaAirlines/Icons",
"url": "https://github.com/AlaskaAirlines/Icons/pull/46"
}
|
gharchive/pull-request
|
Add new icons
Alaska Airlines Pull Request
Add new icons
boadring.svg
elite.svg
Type of change:
Please delete options that are not relevant.
[ ] New capability
[ ] Revision of an existing capability
[ ] Infrastructure change (automation, etc.)
[x] Other (please elaborate)
Checklist:
[x] My update follows the CONTRIBUTING guidelines of this project
[x] I have performed a self-review of my own update
By submitting this Pull Request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Pull Requests will be evaluated by their quality of update and whether it is consistent with the goals and values of this project. Any submission is to be considered a conversation between the submitter and the maintainers of this project and may require changes to your submission.
Thank you for your submission!
-- Orion Design System Team
:tada: This PR is included in version 4.0.2 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T04:10:08.971923
| 2020-10-12T18:18:21
|
719570443
|
{
"authors": [
"blackfalcon",
"ecommerceUXdesign"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13334",
"repo": "AlaskaAirlines/auro-icon",
"url": "https://github.com/AlaskaAirlines/auro-icon/pull/10"
}
|
gharchive/pull-request
|
BREAKING CHANGE: support for Alaska logos by prop
Alaska Airlines Pull Request
BREAKING CHANGE: This update requires a breaking change as the API for
applying custom color and size has been changed.
Fixes: https://github.com/AlaskaAirlines/auro-icon/issues/7
Summary:
This update address the need submitted by #7, as well as address additional concerns about custom color and size management.
See http://auro-21.surge.sh/ for demo
Type of change:
Please delete options that are not relevant.
[x] New capability
[x] Revision of an existing capability
[ ] Infrastructure change (automation, etc.)
[ ] Other (please elaborate)
Checklist:
[x] My update follows the CONTRIBUTING guidelines of this project
[x] I have performed a self-review of my own update
By submitting this Pull Request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Pull Requests will be evaluated by their quality of update and whether it is consistent with the goals and values of this project. Any submission is to be considered a conversation between the submitter and the maintainers of this project and may require changes to your submission.
Thank you for your submission!
-- Auro Design System Team
:tada: This PR is included in version 2.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T04:10:08.998889
| 2023-04-27T09:20:24
|
1686461012
|
{
"authors": [
"AleMorales",
"jianchaoci"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13335",
"repo": "AleMorales/VPL",
"url": "https://github.com/AleMorales/VPL/issues/4"
}
|
gharchive/issue
|
VPL can be precompiled?
Hello, Thanks for your great work! I am trying to precompile the VPL package but got this error.
I checked the folder of "VPL/SS2dp/src/Core", and I realized the "Algorithms.jl" is missed, but it has a file called "algorithms.jl". I am not really sure if it is a bug, and I am not really sure how to fix it, as I am really new to Julia. Can you try to fix it? I will really appreciate it.
===========================================
Precompiling project...
✗ VPL
0 dependencies successfully precompiled in 38 seconds. 267 already precompiled.
The following 1 direct dependency failed to precompile:
VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48]
Failed to precompile VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] to /home/jianchao/.julia/compiled/v1.8/VPL/jl_EqEJDY.
ERROR: LoadError: SystemError: opening file "/home/jianchao/.julia/packages/VPL/SS2dp/src/Core/Algorithms.jl": No such file or directory
Hello, Thanks for your great work! I am trying to precompile the VPL package but got this error. I checked the folder of "VPL/SS2dp/src/Core", and I realized the "Algorithms.jl" is missed, but it has a file called "algorithms.jl". I am not really sure if it is a bug, and I am not really sure how to fix it, as I am really new to Julia. Can you try to fix it? I will really appreciate it.
=========================================== Precompiling project... ✗ VPL 0 dependencies successfully precompiled in 38 seconds. 267 already precompiled.
The following 1 direct dependency failed to precompile:
VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48]
Failed to precompile VPL [bb250ff9-e5b1-4ad2-87ac-b577c1acaa48] to /home/jianchao/.julia/compiled/v1.8/VPL/jl_EqEJDY.
ERROR: LoadError: SystemError: opening file "/home/jianchao/.julia/packages/VPL/SS2dp/src/Core/Algorithms.jl": No such file or directory
I think I just solved this problem, I am sure it is the problem of defining the folder/file name. some of them are wrong in using the small or capital letter.
So I assume this issue is solved? From your description it looked that something went wrong with the Julia Pkg manager (sometimes it happens). Please confirm so that I can close the issue :)
|
2025-04-01T04:10:09.011355
| 2016-04-15T01:39:56
|
148530134
|
{
"authors": [
"AlecAivazis"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13336",
"repo": "AlecAivazis/nautilus",
"url": "https://github.com/AlecAivazis/nautilus/pull/81"
}
|
gharchive/pull-request
|
Documentation updates
This PR fixes a few of the documentation updates - resolving #76, #75, and #59.
@NickDubelman you should look at this and check it against your list. There are probably still changes that I missed (this is just some of the more complicated things like the module index).
|
2025-04-01T04:10:09.028402
| 2023-04-29T02:58:15
|
1689332955
|
{
"authors": [
"58bits",
"AlessioGr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13337",
"repo": "AlessioGr/payload-plugin-lexical",
"url": "https://github.com/AlessioGr/payload-plugin-lexical/issues/25"
}
|
gharchive/issue
|
APIError: The collection with slug media can't be found
Was looking at the demo app in this repo, and when trying to save a post with an image / upload in the post in the LexicalMinimal collection I receive the following error:
[02:55:34] ERROR (payload): APIError: The collection with slug media can't be found.
at new ExtendableError (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/errors/APIError.ts:26:11)
at new APIError (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/errors/APIError.ts:43:5)
at findByIDLocal (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/collections/operations/local/findByID.ts:48:11)
at Payload.findByID (/home/tony/Projects/payload/payload-plugin-lexical-fork/node_modules/payload/src/payload.ts:255:12)
at loadUploadData (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:48:24)
at traverseLexicalField (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:80:30)
at traverseLexicalField (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:108:30)
at populateLexicalRelationships (/home/tony/Projects/payload/payload-plugin-lexical-fork/src/fields/LexicalAfterReadHook.ts:35:32)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
at async /home/tony/Projects/payload/payload-plugin-lexical-fork/demo/node_modules/payload/src/fields/hooks/afterRead/promise.ts:173:31
Hmmmm - you're totally right, I could reproduce that. I'm pretty sure this issue only happens in the demo, not when you actually install the plugin.
Prob some issue with the way I set up the dependencies / project structure. Will have a look!
Should be fixed now!
|
2025-04-01T04:10:09.041920
| 2021-01-29T12:54:53
|
796859630
|
{
"authors": [
"Alex-D",
"PiemP",
"agriffard",
"momentum-tunis",
"ned-kelly"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13338",
"repo": "Alex-D/Trumbowyg",
"url": "https://github.com/Alex-D/Trumbowyg/pull/1192"
}
|
gharchive/pull-request
|
plugin resizimg: updated plugin to fit jquery-resizable 0.35
Fixes #1191
updated the plugin to use the resizableSafe function instead of the resizable to avoid conflict with jquery UI.
removed file resizable-resolveconflicts.js.
updated the documentation: specified the jquery-resizable version's used.
@Alex-D Would you have some time to review this PR?
@Alex-D Can you please have a look at this PR?
Thank you.
@Alex-D Can you please find some time to review this PR?
It would help us if it was merged and a new version was available.
Thank you in advance.
No news on this PR?
Thank you :)
Thank you very much @Alex-D 👍
@agriffard now we are waiting for orchard core to integrate the next release of Trumbowyg
Hi @Alex-D - any chance we can see this one added into a release?
|
2025-04-01T04:10:09.044859
| 2021-06-12T12:25:49
|
919545465
|
{
"authors": [
"AlanFloyd",
"alzubitariq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13339",
"repo": "Alex-Dobrynin/Xamarin.Controls.ImageCropper",
"url": "https://github.com/Alex-Dobrynin/Xamarin.Controls.ImageCropper/issues/3"
}
|
gharchive/issue
|
Sometimes stops working on Android
I have seen that after a couple of times, the controls stops working on Android. Using the debugger, the last statement executed is the "await ImageCropper.Current.Crop(new CropSettings()", but there is no reply to IsFaulted, IsCanceled or IsCompleted. Here the code:
Device.BeginInvokeOnMainThread(async () =>
{
await ImageCropper.Current.Crop(new CropSettings()
{
AspectRatioX = 1,
AspectRatioY = 1,
CropShape = CropSettings.CropShapeType.Rectangle
}, PhotoPath).ContinueWith(t =>
{
if (t.IsFaulted)
{
var ex = t.Exception;
//alert user
}
else if (t.IsCanceled)
{
var ex = t.Exception;
//do nothing
}
else if (t.IsCompleted)
{
var result = t.Result;
Device.BeginInvokeOnMainThread(() =>
{
logo_image.Source = result;
is_saved = false;
photo_path = result;
lbl_initials.IsVisible = false;
logo_image.IsVisible = true;
});
}
});
});
Any idea what could be wrong in my code?
Dietmar
Hi AlanFloyd,
If you are using MediaPicker, i think the workaround is to put some delay between picking a picture and the ImageCropper
like this
var file = await MediaPicker.PickPhotoAsync();
await Task.Delay(2000);
await ImageCropper.Current.Crop(new CropSettings() etc .....
I think the issue who is calling the OnActivityResult firstly with resultCode
|
2025-04-01T04:10:09.056409
| 2024-01-16T13:46:41
|
2084001390
|
{
"authors": [
"gz83",
"rollsicecream",
"trimechee"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13340",
"repo": "Alex313031/thorium",
"url": "https://github.com/Alex313031/thorium/issues/510"
}
|
gharchive/issue
|
strange github bug
Hello, I make new issues with my other new account, but i don't see my issues when i use this account !!! @github ?!!! why ? @Alex313031 @gz83 just to check, have you seen my issues made by my other account?
https://github.com/Alex313031/Mercury/issues/101
https://github.com/Alex313031/thorium/issues/506
https://github.com/Alex313031/thorium/issues/507
It seems that these issues do not exist
@gz83 it's a serious and regrettable bug of @github, it's wasting my time and fortunately I noticed it, my issue exists but only me I can see my 3 issues with my other account :( I'm going to have to copy and paste this that I wrote and open new issues with another account :(
Maybe your account got shadowbanned or something like that?
I use a web browser with a lot of modified flags and added command line and special dns....github sent me an email to my censored account telling me github cannot determine your geographic location and suspects potential hacking... .yet I don't use VPN....disgusted and angry I went to the settings of my new account to delete it and but the delete account button is grayed out and github tells me your account has been reported so I have to contact github support and I have to enable 2-factor authentication, so I gave up....I still have this account to be able to contact the developers and try to post exclusive ideas to improve the software....anyway Thank you @rollsicecream for caring about my serious problem :)
Also, I had this issue before. It was because I used an alias e-mail which is "suspicious" for GitHub.....
Ohh very important information ! Thank you so much Mr @rollsicecream for your great advise, i will try to change my alias !
|
2025-04-01T04:10:09.074471
| 2021-09-15T22:10:59
|
997561860
|
{
"authors": [
"AlexDarigan",
"olestourko"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13341",
"repo": "AlexDarigan/WAT",
"url": "https://github.com/AlexDarigan/WAT/issues/295"
}
|
gharchive/issue
|
Modifying a test class or method causes the test runner to fail
Godot 3.3.3 stable
C# (Mono)
WAT Version 6
When I run tests through the "Test" tabs after modifying a test class' name, a test method name, or ad a new test method, the test runner fails. I'm writing C# tests and using "Build" in Godot after my changes.
This only appears to be happening if I had run tests before making changes - maybe WAT is caching class/method definitions? It appears that the new class/method names aren't picked up in the test runner's dropdowns. There isn't anything in the Output tab and no dropdowns under the failure in the Test tab.
If I reload the project the new tests get picked up and it works again. Toggling the plugin in projects settings works too.
Might be a cache problem alright.
Are you using an external editor? Maybe the filesystem dock signals that WAT uses aren't being sent.
I got around it by skipping the file system check here: https://github.com/AlexDarigan/WAT/blob/3a6e746523b658991e83289f56bd34afc196ee67/addons/WAT/ui/gui.gd#L64
Maybe there's a way of detecting changes to C#-based scripts? Just picked up Godot so very fresh to this all
Are you using an external editor? Maybe the filesystem dock signals that WAT uses aren't being sent.
Yes, VSCode on Ubuntu 20.04
If you save the file, a "resource_saved" signal will be emitted which will trigger an update of the system.
This might be a VSCode issue. Worth trying to see if the changes get picked up when you change and save from the editor.
Yup, WAT detects change made through the builtin editor
Oh I have an idea. The filedock signals will probably trigger from an external editor but this resource_saved one won't because it is a plugin signal and that's the one we care about.
I think having the system auto-update might be the best for now until I can figure this out.
Maybe a launch.json for WAT specifically.
I'm very new to Godot so I'm not sure what all of that means :sweat_smile: If its useful at all, the builtin editor is picking up changes I make externally in VSCode.
I actually did try setting up a launch.json for WAT but the current version of Godot is ignore the args passed to it, so it wouldn't take the WAT cli args:
https://github.com/godotengine/godot-csharp-vscode/issues/19
In the meantime just keeping my hack to skip the _filesystem.change check is enough for me, personally. If this ends up being a difficult thing to figure out, perhaps a setting on the extension to optionally skip the check for external editor users could be a compromise?
A reminder to myself: Could experiment different files saved signals from other sources with a check that breaks if filesystem was updated.
SriptEditor can receive signals about changes to active scripts, but I can't find a way to receive signals about inactive ones. Here's a simple draft PR for a setting to disable caching: https://github.com/AlexDarigan/WAT/pull/296
Check if this works c6bb81ab803e333a11352ee207ba904ed72b0030
Seems this issue has been resolved.
|
2025-04-01T04:10:09.077473
| 2022-12-28T21:20:56
|
1513177211
|
{
"authors": [
"AlexGladkov",
"masterofdaemon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13342",
"repo": "AlexGladkov/Odyssey",
"url": "https://github.com/AlexGladkov/Odyssey/issues/71"
}
|
gharchive/issue
|
how can i use in with hilt?
and this methods are undefined: setupNavigation("splash") { splashScreen() }
You can try 1.3.0-beta04 version and add odyssey-android in dependencies or wait stable 1.3.0 release
You can read documentation about this here - https://github.com/AlexGladkov/Odyssey/blob/rc/release-1.3.0/documentation/HILT.md
And it works only in 1.3+ versions
|
2025-04-01T04:10:09.120807
| 2023-08-20T18:06:55
|
1858217644
|
{
"authors": [
"TAforever",
"waddyhussain"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13344",
"repo": "Alexays/Waybar",
"url": "https://github.com/Alexays/Waybar/issues/2425"
}
|
gharchive/issue
|
hyprland/language add on-click config options
Add support for on-click events (on-scroll events would be helpful too) to support customisations such as switching between different keyboard layouts
Hi, this functionality has already been added.
Here is an example of my configuration
"hyprland/language": {
"format-en": "us ",
"format-ru": "ru ",
"keyboard-name": "gaming-keyboard",
"on-click": "hyprctl switchxkblayout gaming-keyboard next"
},
Thanks, marking as closed
|
2025-04-01T04:10:09.122378
| 2019-05-17T09:17:18
|
445344826
|
{
"authors": [
"Alexays",
"lockywolf",
"unresolvedsymbol"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13345",
"repo": "Alexays/Waybar",
"url": "https://github.com/Alexays/Waybar/issues/324"
}
|
gharchive/issue
|
Battery module rewrite
Using udev
Will closes #229 #237
This should be an added module/mode because not everyone has upower
At the moment, if a battery is not present, the only thing the module does is complain in the log:
[warning] module battery: Disabling module "battery", No batteries.
May I suggest that that module present itself in a deliberately "broken" state, rather than silently failing, in order to urge the user to resolve the issue?
|
2025-04-01T04:10:09.123263
| 2022-01-14T18:37:11
|
1103996398
|
{
"authors": [
"Alexays",
"nullobsi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13346",
"repo": "Alexays/Waybar",
"url": "https://github.com/Alexays/Waybar/pull/1394"
}
|
gharchive/pull-request
|
fix: Use locale when formatting clock
Fixes #1393
When formatting the clock, pass the locale to the fmt function.
Thanks!
|
2025-04-01T04:10:09.178393
| 2023-08-30T14:11:19
|
1873769024
|
{
"authors": [
"AlexeyBoiko",
"yogananda-muthaiah"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13347",
"repo": "AlexeyBoiko/DgrmJS",
"url": "https://github.com/AlexeyBoiko/DgrmJS/issues/35"
}
|
gharchive/issue
|
Can Arrows be made to animated flow as new feature?
Hi Alex,
Can you think of adding animated flow to arrows.. so it gives a better flow visualization ?? Just a thought
Hi @yogananda-muthaiah
Please give an example.
Closed due inactivity
|
2025-04-01T04:10:09.183127
| 2022-12-12T19:15:46
|
1492573993
|
{
"authors": [
"AlexisJW",
"PinkMoon25"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13348",
"repo": "AlexisJW/set-up-myFirst-mobile",
"url": "https://github.com/AlexisJW/set-up-myFirst-mobile/issues/9"
}
|
gharchive/issue
|
Portfolio: application deploy
Deploy my portfolio using GitHub Pages.
Check the online version of my portfolio and make sure that the page works properly.
Update the README of my repository to include a link to the online version.
The link to the online version of my application
Hi @AlexisJW ,
Your project is complete! There is nothing else to say other than... it's time to merge it :shipit:
Congratulations! 🎉
Highlights
linter checks :heavy_check_mark:
professional readme :heavy_check_mark:
Added live demo link to reade :heavy_check_mark:
deployed with github and live demo working :heavy_check_mark:
Optional suggestions
Every comment with the [OPTIONAL] prefix won't stop the approval of this PR. However, I strongly recommend you to take them into account as they can make your code better. Some of them were simply missed by the previous reviewer and addressing them will really improve your application.
Cheers and Happy coding!👏👏👏
Feel free to leave any questions or comments in the PR thread if something is not 100% clear.
Please, remember to tag me in your question so I can receive the notification.
As described in the Code reviews limits policy you have a limited number of reviews per project (check the exact number in your Dashboard). If you think that the code review was not fair, you can request a second opinion using this form.
|
2025-04-01T04:10:09.203640
| 2024-11-20T19:44:15
|
2676997516
|
{
"authors": [
"dvenprasad",
"jaclyn-taroni"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13349",
"repo": "AlexsLemonade/scpca-docs",
"url": "https://github.com/AlexsLemonade/scpca-docs/pull/373"
}
|
gharchive/pull-request
|
Update bulk names for download illustrations
Closes #369
Updated the bulk file names
Bumped up the quality of the images too
These are purposely 2x the width that they get presented as.
Okay resized them. They were all consistently 1200px width - so I resized it to be that.
|
2025-04-01T04:10:09.205683
| 2022-07-01T18:58:37
|
1291739067
|
{
"authors": [
"jashapiro"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13350",
"repo": "AlexsLemonade/scpca-docs",
"url": "https://github.com/AlexsLemonade/scpca-docs/pull/87"
}
|
gharchive/pull-request
|
Add information about multiplex samples to download section
Here I am adding a section about the multiplex downloads to the download
documentation. I tried to be fairly brief, but I think this reflects the plan we agreed upon. Let me know if there is more detail that you think is appropriate in this section, or additional links that might make sense.
I am leaving it as a draft for now, as I think we still need illustrations, as per https://github.com/AlexsLemonade/scpca-docs/issues/75#issuecomment-1162082576
This PR replaces #82, so that what I write can be reviewed by @allyhawkins.
Closes #75
I have also added/updated image files so they are all the same resolution. I wonder if they are a bit big on rendering, so I could add a width parameter (changing to html) if we think that is a good idea.
|
2025-04-01T04:10:09.209854
| 2020-11-20T17:13:10
|
747658186
|
{
"authors": [
"AlexxIT",
"Violo1975",
"nathanmay"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13351",
"repo": "AlexxIT/SonoffLAN",
"url": "https://github.com/AlexxIT/SonoffLAN/issues/287"
}
|
gharchive/issue
|
Nothing happens :(
Hi Everybody, I'm a newby...
I have follow all the step using HACS and my Yaml file is:
# SonOff
sonoff:
username<EMAIL_ADDRESS> password: xxx
force_update: True
scan_interval: '00:01:00' # (optional) default 5 minutes
mode: auto
reload: always # update device list every time HA starts
but nothing happen, someone can help me.....many tks
for info I've to add two sensors:
4CH Pro R2
2CH with RF 433 MHz
System info:
Versione HS: 0.118.1
Version OS: 4.19.127-v7
Grazie a tutti :)
up.... please :(
Nothing happens... What you expect to happen?
Hi,
It doesn't give me errors but I can't see my devices
Could be linked to #294
Devices don't supported yet.
|
2025-04-01T04:10:09.217054
| 2021-04-19T11:10:35
|
861194303
|
{
"authors": [
"AlexxIT",
"weitheng"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13352",
"repo": "AlexxIT/WebRTC",
"url": "https://github.com/AlexxIT/WebRTC/issues/22"
}
|
gharchive/issue
|
Stuck on "Init Connection" Safari Desktop
Updated to latest master version which does not require specifying ports, streams are stuck on init connection when using Safari Desktop. Tried specifying ports from 50000-59999, but did not work. Working well with other browsers (Chrome, Firefox) on the same laptop even when not specifying ports
Try clear cache
Cleared cache and tried on incognito, both didn't work. Went onto JS console, got this message Blocked *URL* from asking for credentials because it is a cross-origin request.
Looks like url has been blocked
This is not about component. Because it doesn't do any requests.
I think this is browser cache problem.
This is not about component. Because it doesn't do any requests.
Could the request be from my RTSP stream? Because the RTSP url contains the username and password to access the stream.
I have tried clearing the cache and gone onto incognito mode again, but with no luck. :(
I also tried disabling the Cross-Origin Restrictions on Safari, but I am immediately redirected to the IP address of the camera upon refreshing the page, prompting me to enter the username and password (even though they are already included in the RTSP url, and works fine on my other devices). But after entering the correct credentials, I am redirected back to my HA domain, with the following message Failed to load resource: the server responded with a status of 401 (Unauthorized) on the JS console.
You can use camera entity instead of url in latest version. For example, if the camera is configured as Generic.
You can use camera entity instead of url in latest version. For example, if the camera is configured as Generic.
Tried that, still getting the same error on JS console. :(
Really unsure what's causing the issue, I kept refreshing the page and sometimes I managed to get it working temporarily, but another refresh will bring it down again.
I managed to have a try on my other MacBook, and it works flawlessly on Safari. You could be right that it's the browser problem specifically the cache or website date. But I did clear both of them.
This error not from component. You can remove card from page and check.
This error not from component. You can remove card from page and check.
I hope you didn't set up the card poster?
Nope, no card poster. I created a new page and try adding the entity one by one, it seems that it works fine when there's only 1 or 2 WebRTC entities on the page. Once there's more than 2, it struggles to initialise connection.
I will keep on trying to see what's the issue, it should be a problem with my browser. Thanks for your help!
I have tested on Safari Version 14.0 (156<IP_ADDRESS>.9, 15610).
What version do you have?
I am using Version 14.0.3
Somehow the streams are working again now, I will continue to monitor the issue. Thanks!
Reopen if there's a problem
|
2025-04-01T04:10:09.243472
| 2022-11-14T16:00:06
|
1448326570
|
{
"authors": [
"Alhajideen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13353",
"repo": "Alhajideen/Space--Travelers",
"url": "https://github.com/Alhajideen/Space--Travelers/issues/18"
}
|
gharchive/issue
|
Team3 [1pt] Create basic structure for Dragons - Setup
Create a route and a view component. Use <NavLink /> for the page navigation links and style active class to indicate which section/page the user is currently on (underline active navigation link).
Create a directory for all Redux state slice files.
N/A
|
2025-04-01T04:10:09.245713
| 2023-11-07T16:06:23
|
1981712252
|
{
"authors": [
"Ali-Muhsin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13354",
"repo": "Ali-Muhsin/Whispers",
"url": "https://github.com/Ali-Muhsin/Whispers/issues/2"
}
|
gharchive/issue
|
Read and write into jsons
We check the item name, and then go through the inventory slots, checking each one and seeing if there's a slot free, or one that can be used(it has an item of the same type and not filled to capacity). We delete the item, and change the item name in the slot from NIL(if it was empty) and then change the item count. For this, I need to work with JSON files. There are premade functions in the global.gd autoload script, but it isn't dynamic yet.
Done.
|
2025-04-01T04:10:09.256339
| 2015-12-09T12:58:50
|
121232202
|
{
"authors": [
"AliSoftware",
"Fosile"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13355",
"repo": "AliSoftware/OHAttributedStringAdditions",
"url": "https://github.com/AliSoftware/OHAttributedStringAdditions/issues/7"
}
|
gharchive/issue
|
After changing the Font its not working anymore
Thanks for this great piece of Code!
I have a little Problem when i try to change the Font and the Text Size, because after that my Text (which should be bold) is only a normal Text.
This is my Code:
NSMutableAttributedString* htmlString = [NSMutableAttributedString attributedStringWithHTML:@"<b>Lorem Ipsum</b>"];
[htmlString setFont:[UIFont fontWithName:@"Arial" size:14.0f]];
self.textView = [UITextView new];
self.textView.scrollEnabled = NO;
[self addSubview:self.textView];
self.textView.attributedText = htmlString;
[self.textView sizeToFit];
Well that's quite logical and the intended behavior: you set some style (using HTML) then override it using setFont afterwards.
The information you probably missed is that in any text engine, the bold and italic attributes are actually not an attribute of the text, like it's color for example, but is part of the font. So even in Word for example, if you write some text in Arial, then set it Bold, in fact it replaces the font with "Arial-Bold".
Arial is the Font Family, and "Arial-Regular", "Arial-Bold" and "Arial-Italics" are variants of that font family.
So to implement the behavior you intended, you have multiple solutions:
Avoid HTML
Use the bold font directly. You can use +[UIFont fontWithFamily:@"Arial" size:14.0f bold:YES italic:NO as your font (see here).
Or use setFontBold: on your NSMutableAttributedString after setting its font
In both these solutions, there will then be no need to use the <b> HTML tag (or even initialize the string using HTML, which is slower generally than using a plain string and add attributes yourself) to make it bold.
Keep HTML
If you really need to initialize your string via HTML, you could:
set the font via HTML too (<font> tag)
or use enumerateFontsInRange:includeUndefined:usingBlock: that it will iterate over each run of your text that have different fonts (in case your HTML mix different fonts and you want to change them all but still keep the bold/italics/regular attribute of each) and alter the existing font to replace with the correct variant of the "Arial" font family
For example you can use the -[UIFont symbolicTraits] to get the traits (bold, italics, etc) of the current font being iterated, then use [UIFont fontWithFamily:@"Arial" size:14 traits:currentTraits] to apply the Arial font with the same bold/italics traits instead.
Thank you very much for your fast response, your explanation was fantastic!! I need to keep the HTML because usally i get my Text from a Server (which includes HTML). I changed my Code to this now, but it seems i can´t access the traits, because my Text still shows up without any any bold:
NSMutableAttributedString* htmlString = [NSMutableAttributedString attributedStringWithHTML:@"<b>Lorem Ipsum</b>"];
[htmlString enumerateFontsInRange:NSMakeRange(0,htmlString.length)
includeUndefined:YES
usingBlock:^(UIFont *font, NSRange range, BOOL *stop)
{
UIFontDescriptorSymbolicTraits traits = [font symbolicTraits];
UIFont *newFont = [UIFont fontWithFamily:@"Arial" size:14.0f traits:traits];
[htmlString setFont:newFont range:range];
}];
self.textView = [UITextView new];
self.textView.scrollEnabled = NO;
[self addSubview:self.textView];
self.textView.attributedText = htmlString;
[self.textView sizeToFit];
Did you try to debug your code? Does the block in your enumeration actually properly get called, and if so how many iterations? What is the value of the traits you retrieve in each iteration ? If those traits you gather from the existing string do contain the bold attribute as expected, do the newFont find Arial-Bold or fail to find it and return another non-bold font variant instead? etc.
So it seems the Problem is symbolicTraits and the Text from my Server, because after I do:
UIFontDescriptorSymbolicTraits traits = [font symbolicTraits];
i only get some strange numbers inside my traits variable like "268435458"
When I use your String from the Example Project, everything works fine and Debugger gives me
traits = UIFontDescriptorClassOldStyleSerifs
This is the Text from my Server:
htmlText = @"<b>Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat, sed diam voluptua.</b><br /><br />At vero eos et accusam et justo duo dolores et ea rebum. Stet clita kasd gubergren, no sea takimata sanctus est Lorem ipsum dolor sit amet. Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat, sed diam voluptua. <br /><br /><b>At vero eos et accusam et justo duo dolores et ea rebum.</b><br /> Stet clita kasd gubergren, no sea takimata sanctus est Lorem ipsum dolor sit amet."
Mmmh strange, do you know what is the original font generated by your HTML?
Trick: you could wrap your HTML returned by the server into a <font face="Arial">…</font> before parsing it, it should also do the trick in a simpler way.
Thank you very much, with the little trick its working!
|
2025-04-01T04:10:09.258807
| 2024-09-20T12:16:39
|
2538663005
|
{
"authors": [
"AliYoussef96",
"antagomir"
],
"license": "Artistic-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13356",
"repo": "AliYoussef96/LimROTS",
"url": "https://github.com/AliYoussef96/LimROTS/issues/6"
}
|
gharchive/issue
|
Homepage
You can set package homepage from the settings. It is automatically generated with pkgdown.
The result will be like:
https://microbiome.github.io/mia/
@ake123 can advice how it is done if you can't find instructions easily.
The package's homepage is now live and accessible at: https://aliyoussef96.github.io/LimROTS/.
|
2025-04-01T04:10:09.294798
| 2021-03-10T05:12:53
|
827124628
|
{
"authors": [
"Alischrec"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13357",
"repo": "Alischrec/aws-thought",
"url": "https://github.com/Alischrec/aws-thought/issues/1"
}
|
gharchive/issue
|
Create an AWS Account
User Stories
As a developer, I want to be able to view the management console in AWS.
As a developer, I want to manage my IAM role.
As a developer, I want to set up a billing alert.
Completed this task
|
2025-04-01T04:10:09.352159
| 2017-08-30T01:48:15
|
253854102
|
{
"authors": [
"prajapati-parth",
"uynap"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13358",
"repo": "AllenFang/react-bootstrap-table",
"url": "https://github.com/AllenFang/react-bootstrap-table/issues/1577"
}
|
gharchive/issue
|
Is it possible to display multiple datafields in one column?
For example, the front end got JSON {"firstName": "David", "lastName": "Jones"}. I would like to display it like: "Full Name" with the combination of firstName and lastName.
Is it possible to do? Or I have to change the backend API ?
@uynap Yes. Please consider the following example
const data = [
{
empId: 1,
name: {
firstName: 'David',
lastName: 'Jones'
}
},
{
empId: 2,
name: {
firstName: 'xyz',
lastName: 'abc'
}
}
]
function nameFormatter(cell, row) {
return `${cell.firstName} ${cell.lastName}`
}
class MyTable extends React.component {
render() {
return (
<BootstrapTable data={ data }>
<TableHeaderColumn dataField='empId'>Employee Id</TableHeaderColumn>
<TableHeaderColumn dataField='name' dataFormat={ nameFormatter }>Full name</TableHeaderColumn>
</BootstrapTable>
)
}
}
Refer this page
Section: Column Format with HTML String
Thank you for the detail example. prajapati-parth .
I was considering formatting the data at Redux's middleware level. Then it's one time for all.
@uynap All right. I wasn't aware that you were using Redux. Well that is at least better than changing the backend API. :+1:
|
2025-04-01T04:10:09.357491
| 2022-01-06T14:11:27
|
1095357244
|
{
"authors": [
"elega",
"jiacheliu3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13359",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/issues/14803"
}
|
gharchive/issue
|
Make worker block lock pool size observable
Is your feature request related to a problem? Please describe.
Alluxio 2.7
The worker has a pool of block locks. Each lock item is locking one block (serve one to many clients).
https://github.com/Alluxio/alluxio/blob/309fd69f45730910b3c0a7e127316d529c3544ef/core/server/worker/src/main/java/alluxio/worker/block/BlockLockManager.java#L55
The pool size should be observable by metric and logging, so we know when this pool is full.
Describe the solution you'd like
A clear and concise description of what you want to happen.
Describe alternatives you've considered
A clear and concise description of any alternative solutions or features you've considered.
Urgency
MEDIUM
Additional context
Add any other context or screenshots about the feature request here.
/assign @elega
|
2025-04-01T04:10:09.359406
| 2021-02-24T03:13:19
|
815042886
|
{
"authors": [
"alluxio-ci",
"apc999",
"maobaolong",
"waterlx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13360",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/12924"
}
|
gharchive/pull-request
|
Remove redundant blank in PropertyKey
A format change, no logic involved.
Can one of the admins verify this patch?
@waterlx LGTM, please sign the contribution license agreement
@waterlx Thank for your contribution, @apc999 Could you please take a look?
Thank you @maobaolong and @apc999 for taking care of this change.
alluxio-bot, merge this please
|
2025-04-01T04:10:09.360530
| 2023-06-29T21:54:48
|
1781563804
|
{
"authors": [
"Zijian-Zhu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13361",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/17714"
}
|
gharchive/pull-request
|
[DOCFIX] Fix Language dropdown bug in new layout in os and add pr check
Fix Language dropdown in new docs layout shows English twice
alluxio-bot, merge this please
|
2025-04-01T04:10:09.362841
| 2017-03-28T15:45:27
|
217604981
|
{
"authors": [
"AmplabJenkins",
"Reidddddd",
"apc999"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13362",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/5004"
}
|
gharchive/pull-request
|
[ALLUXIO-2365] Parameterize the returned type Inode in TempInodePathForDescendant.getInode
https://alluxio.atlassian.net/browse/ALLUXIO-2365
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/14099/
Test PASSed.
Merged build finished. Test PASSed.
LGTM. thanks!
|
2025-04-01T04:10:09.364467
| 2017-09-29T00:20:34
|
261498239
|
{
"authors": [
"AmplabJenkins",
"calvinjia"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13363",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/6206"
}
|
gharchive/pull-request
|
[SMALLFIX] Clean up doc descriptions
Start with capital letter and end with period.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/17070/
Test PASSed.
|
2025-04-01T04:10:09.373224
| 2020-10-23T19:00:19
|
728452200
|
{
"authors": [
"Almighty-Alpaca",
"InfRandomness"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13364",
"repo": "Almighty-Alpaca/JetBrains-Discord-Integration",
"url": "https://github.com/Almighty-Alpaca/JetBrains-Discord-Integration/issues/131"
}
|
gharchive/issue
|
New project settings having an issue
I created a new project today with the plugin installed in my IDE and when I went inside my IDE to set the project settings I saw that
for some reason, there's an "ask" field, and whenever I switch to another value, the field ask disappear.
That felt weird to see and I don't know if that's a bug or not so I reported it just in case
This actually is intentional: When you didn't click the pop-up yet the value is set to "Ask". But it doesn't make sense to manually select that as visibility because it would just mean that the next time you open the project the pop-up would show again.
yeah, I see, thanks !
|
2025-04-01T04:10:09.387674
| 2020-10-07T12:19:40
|
716478630
|
{
"authors": [
"colourful-land",
"hboon",
"vladyslav-iosdev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13365",
"repo": "AlphaWallet/alpha-wallet-ios",
"url": "https://github.com/AlphaWallet/alpha-wallet-ios/issues/2198"
}
|
gharchive/issue
|
AWS Lambda expect Bitcoin-style signature but was given Vitalik-style signature
AWS Lambda accepts DER encoded signature like the one used in Bitcoin.
Here is the code @vladyslav-iosdev used to create a signature:
public func sign(hash: Data, privateKey: Data) throws -> Data {
precondition(hash.count == 32, "Expect hash size to be 32")
precondition(privateKey.count == 32, "Expect private key size to be 32")
var signature = secp256k1_ecdsa_recoverable_signature()
try privateKey.withUnsafeBytes { (key: UnsafePointer<UInt8>) in
if secp256k1_ec_seckey_verify(context, key) != 1 {
throw Secp256k1Error.invalidPrivateKey
}
let result = hash.withUnsafeBytes { hash in
secp256k1_ecdsa_sign_recoverable(context, &signature, hash, key, nil, nil)
}
if result == 0 {
throw Secp256k1Error.invalidPrivateKey
}
}
var output = Data(count: 65)
var recid = 0 as Int32
_ = output.withUnsafeMutableBytes { (output: UnsafeMutablePointer<UInt8>) in
secp256k1_ecdsa_recoverable_signature_serialize_compact(context, output, &recid, &signature)
}
// add back recid to get 65 bytes sig
output[64] = UInt8(recid)
return output
}
Which apparently uses Vitalik style Ethereum signature. Vitalik could have the courtesy to extend the DER siganture format to do what he want, given DER was originally intended to be extensible and was extended all the time by different crypto standards, and if done right can result in a signature that is compatible with older (e.g. Bitcoin) codebase even without recompiling! But no, Vitalik likes to do things his way.
There is a bit of hope that @vladyslav-iosdev said the code might actually call
https://github.com/bitcoin-core/secp256k1/blob/master/include/secp256k1.h
Which is a Bitcoin codebase - hence it must be able to store Bitcoin signature. (for one way to tell, a Bitcoin-style, i.e. Standard Comaptible, a signature is about 71 bytes, can be shorter occassionally, and has a variable length - which begets the question to Vitalik - 𝑎𝑟𝑒 𝑦𝑜𝑢 𝑠𝑢𝑟𝑒 that you want people to rewrite all libraries when you decided to extend the keysize when 256-bit is no longer enough? Or will you consider extensible standard, which was created after the pain of extending key size back in the 80s in the first place?)
How can we get a Bitcoin-style Signature? I can change the Lambda to be compatible with Vitalik-style but that requires some research work into how that works.
Alternatively, consider letting the mobile passing the Signature as (𝑥,𝑦) where each is BigInteger of 256-bits long, and I should be able to reconstruct needed signature in bytes at the server end.
@vladyslav-iosdev I’ll look into this but if you have time before me, can you try look into the pod which contains HDWallet? Since it has multi-coin support, we might be able to use it or copy the code from there. The latter may be because the interface only supports seeds and not private keys directly.
i found solution for this
public func sign(_ data: Data, privateKey: Data) throws -> Data {
let signature = UnsafeMutablePointer<secp256k1_ecdsa_signature>.allocate(capacity: 1)
defer {
signature.deallocate(capacity: 1)
}
let status = data.withUnsafeBytes { (ptr: UnsafePointer<UInt8>) in
try? privateKey.withUnsafeBytes { (key: UnsafePointer<UInt8>) in
secp256k1_ecdsa_sign(context, signature, ptr, key, nil, nil)
}
}
guard status == 1 else {
throw CryptoError.signFailed
}
let normalizedsig = UnsafeMutablePointer<secp256k1_ecdsa_signature>.allocate(capacity: 1)
defer {
normalizedsig.deallocate(capacity: 1)
}
secp256k1_ecdsa_signature_normalize(context, normalizedsig, signature)
var length: size_t = 128
var der = Data(count: length)
guard der.withUnsafeMutableBytes({ return secp256k1_ecdsa_signature_serialize_der(context, $0, &length, normalizedsig) }) == 1 else {
throw CryptoError.noEnoughSpace
}
der.count = length
return der
}
|
2025-04-01T04:10:09.406719
| 2021-01-16T05:13:17
|
787351200
|
{
"authors": [
"Altimis",
"ratriik",
"the-data-mogul"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13367",
"repo": "Altimis/Scweet",
"url": "https://github.com/Altimis/Scweet/issues/31"
}
|
gharchive/issue
|
Request to add the Location feature based on Latitude / Longitude or Zipcode
Hi,
Is it possible to replicate the location function (from Twint)? Or is that not possible because of the upgrades made by Twitter?
Thank you!
@the-data-mogul I'll see about that.
@the-data-mogul I'll see about that.
@the-data-mogul can you specify the exact file where the location function exists in twint project ?
@the-data-mogul can you specify the exact file where the location function exists in twint project ?
Thankyou very much for this repository. I've been searching how to scrape tweet in certain date then I found this, it really helps a lot!
Btw, is there any ways to filter tweet based on location using Scweet? Sorry, I'm kinda new. Thankyou.
|
2025-04-01T04:10:09.427257
| 2023-08-10T12:14:01
|
1845059933
|
{
"authors": [
"vscaiceanu-1a"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13368",
"repo": "AmadeusITGroup/otter",
"url": "https://github.com/AmadeusITGroup/otter/issues/657"
}
|
gharchive/issue
|
[Feature]: o3r/testing schematics should handle playwright setup
Context
Currently, o3r/core is adding the Playwright framework.
Proposal
Similarly to https://github.com/AmadeusITGroup/otter/issues/602 (unit tests), Playwright framework (e2) should be setup by the o3r/testing schematics.
Fixed in https://github.com/AmadeusITGroup/otter/pull/663
|
2025-04-01T04:10:09.430042
| 2023-09-04T08:30:28
|
1879778209
|
{
"authors": [
"kpanot",
"matthieu-crouzet"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13369",
"repo": "AmadeusITGroup/otter",
"url": "https://github.com/AmadeusITGroup/otter/issues/718"
}
|
gharchive/issue
|
[Feature]: VSCode intellisense for configuration/styling metadata
Context
As a developper I would like to have help from my IDE to add metadata to the different modules of Otter.
Proposal
[ ] Intellisense for Configuration metadata
[ ] Intellisense for Styling metadata
[ ] EsLint plugin rule to validate configuration metadata
[ ] Support of @o3rWidget
https://github.com/AmadeusITGroup/otter/pull/893
Intellisense for @o3rWidget and @o3rWidgetParam
https://github.com/AmadeusITGroup/otter/pull/896
|
2025-04-01T04:10:09.431801
| 2024-11-19T11:21:44
|
2671813966
|
{
"authors": [
"kpanot"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13370",
"repo": "AmadeusITGroup/otter",
"url": "https://github.com/AmadeusITGroup/otter/pull/2486"
}
|
gharchive/pull-request
|
feat: move schematic to package dependencies to simplify ng-add
Proposed change
move @o3r/schematic package to packages dependencies to simplify ng-add and supporting it outside of Otter project.
Related issues
- No issue associated -
Temporary closed due to the large amount of dependencies bring by it
|
2025-04-01T04:10:09.474534
| 2018-01-09T05:59:49
|
286978971
|
{
"authors": [
"Aminadav",
"Passion08"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13371",
"repo": "Aminadav/screenshot-extension",
"url": "https://github.com/Aminadav/screenshot-extension/issues/135"
}
|
gharchive/issue
|
Visible screen shot is not working
Visible screenshot is not capturing any image inst
ead it is displaying ba blank page
Please test the last version, it should be fixed.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.