added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:55.100818
| 2021-06-30T20:04:53
|
934083147
|
{
"authors": [
"betodealmeida",
"graceguo-supercat"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3711",
"repo": "apache/superset",
"url": "https://github.com/apache/superset/pull/15482"
}
|
gharchive/pull-request
|
feat: more SIP-40 errors
SUMMARY
Make more SQL Lab error message SIP-40 compliant.
BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF
TESTING INSTRUCTIONS
ADDITIONAL INFORMATION
[ ] Has associated issue:
[ ] Changes UI
[ ] Includes DB Migration (follow approval process in SIP-59)
[ ] Migration is atomic, supports rollback & is backwards-compatible
[ ] Confirm DB migration upgrade and downgrade tested
[ ] Runtime estimates and downtime expectations provided
[ ] Introduces new feature or API
[ ] Removes existing feature or API
Hi @betodealmeida this PR cause Sql Lab can not handle some error message correctly:
I have reverted this from airbnb's release branch, but you probably have to fix it in open source master branch.
|
2025-04-01T06:37:55.108066
| 2021-12-20T12:23:08
|
1084729679
|
{
"authors": [
"hbruch",
"villebro"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3712",
"repo": "apache/superset",
"url": "https://github.com/apache/superset/pull/17826"
}
|
gharchive/pull-request
|
fix(deck.gl): update view state on property changes (#17720)
SUMMARY
Updates viewport state of deck.gl on filter updates
BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF
Before (screeshot 1 & 2):
After:
https://user-images.githubusercontent.com/2187389/146766468-a031dad9-b4d1-48ff-b307-dbfa2dd79417.mov
TESTING INSTRUCTIONS
Create a Dashboard containing a deck.gl polygon chart (auto zoom = True) with a column included in filter
Open the dahsboard => the map is zoomed according to show features (screenshot 1)
Change native filter value
The map is now zoomed to the new bounding box (see video)
ADDITIONAL INFORMATION
[x] Has associated issue: fixes #17720
[ ] Required feature flags:
[ ] Changes UI
[ ] Includes DB Migration (follow approval process in SIP-59)
[ ] Migration is atomic, supports rollback & is backwards-compatible
[ ] Confirm DB migration upgrade and downgrade tested
[ ] Runtime estimates and downtime expectations provided
[ ] Introduces new feature or API
[ ] Removes existing feature or API
@hbruch can you rebase this? I believe there may have been some CI issues when you pushed these last changes
@hbruch sorry about this, but there's yet some more flakiness in our CI pipeline: #17918 . Bear with us while we get it sorted..
@villebro Is there anything I could do to have this merged? Thx
|
2025-04-01T06:37:55.115537
| 2019-06-10T07:35:10
|
454037038
|
{
"authors": [
"doychin",
"jeanouii",
"jgallimore"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3713",
"repo": "apache/tomee",
"url": "https://github.com/apache/tomee/pull/482"
}
|
gharchive/pull-request
|
Code to add support for TOMEE-542
Signed-off-by: Doychin Bondzhev<EMAIL_ADDRESS>
@doychin Thanks for the PR
It looks good to me.
Even if it's a small change, I think it should be visible on release notes.
Do you mind creating a JIRA ticket and updating the ticket title with it?
This looks great, and thank you for the PR! I'm just going to run a build with this on a Windows box and get it merged in for you.
|
2025-04-01T06:37:55.122902
| 2020-06-17T21:52:24
|
640765380
|
{
"authors": [
"dpham692",
"ocket8888"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3714",
"repo": "apache/trafficcontrol",
"url": "https://github.com/apache/trafficcontrol/issues/4804"
}
|
gharchive/issue
|
Can mark an empty ip address as service when perform /POST servers API
I'm submitting a ...
[x] bug report
[ ] new feature / enhancement request
[ ] improvement request (usability, performance, tech debt, etc.)
[ ] other
Traffic Control components affected ...
[ ] CDN in a Box
[ ] Documentation
[ ] Grove
[ ] Traffic Control Client
[ ] Traffic Monitor
[x] Traffic Ops
[ ] Traffic Ops ORT
[ ] Traffic Portal
[ ] Traffic Router
[ ] Traffic Stats
[ ] Traffic Vault
[ ] unknown
Current behavior:
User can create a servers with empty IP address and can mark it as service.
For example using this payload:
{
"cachegroupId": 3,
"cachegroup": "infrastructure",
"cdnId": 2,
"cdnName": "testCDN",
"domainName": "test.net",
"hostName": "testingServer",
"interfaceMtu": 1500,
"interfaceName": "eth0",
"ipAddress": "",
"ipGateway": "<IP_ADDRESS>",
"ipNetmask": "<IP_ADDRESS>",
"ip6Address": "0:0:0:0:0:0:0:1",
"ip6Gateway": "::1",
"physLocationId": 1,
"physLocation": "Augusta-ME",
"profileId": 34,
"profile": "ATS_Edge_MKGA",
"statusId": 2,
"typeId": 11,
"updPending": false,
"ipIsService": true,
"ip6IsService": false
}
The server will return 200 OK
Expected / new behavior:
Should return 400 with message like "an empty IP or IPv6 address cannot be marked as a service address"
Minimal reproduction of the problem with instructions:
Anything else:
There's also another problem with that. You said it returns 200 OK, but ignoring the fact that you made an empty IP a service address, it also, therefore, doesn't have any service addresses. It should have at least rejected it on that basis.
@ocket8888 Yeah, you right about that. I just check and server still create without ip and ip6
@ocket8888 Also even if user input the ipv6 and ip address in. The TP still clean out ip of the servers.
|
2025-04-01T06:37:55.129701
| 2019-03-27T12:25:08
|
425930036
|
{
"authors": [
"asfgit",
"smalenfant"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3715",
"repo": "apache/trafficcontrol",
"url": "https://github.com/apache/trafficcontrol/pull/3440"
}
|
gharchive/pull-request
|
BACKPORT: fixes #3223. updated type of SteeringTargetNullable.Value to JSONIntStr
(cherry picked from commit 3e11da43d31ce8e4469e033d633a47976a3f783b)
Which issue is fixed by this PR? If not related to an existing issue, what does this PR do?
Fixes #3223
Can't update or create steering target in 3.0.0 or 3.0.1
Which TC components are affected by this PR?
[ ] Documentation
[ ] Grove
[ ] Traffic Analytics
[ ] Traffic Monitor
[x] Traffic Ops
[ ] Traffic Ops ORT
[x] Traffic Portal
[ ] Traffic Router
[ ] Traffic Stats
[ ] Traffic Vault
[ ] Other _________
What is the best way to verify this PR? Please include manual steps or automated tests.
(If no tests are part of this PR, please provide explanation as to why no tests are included.)
Check all that apply
[ ] This PR includes tests
[ ] This PR includes documentation updates
[ ] This PR includes an update to CHANGELOG.md
[ ] This PR includes all required license headers
[ ] This PR includes a database migration (ensure that migration sequence is correct)
[ ] This PR fixes a serious security flaw. Read more: www.apache.org/security
Refer to this link for build results (access rights to CI server needed):
https://builds.apache.org/job/trafficcontrol-PR/3402/
Test PASSed.
|
2025-04-01T06:37:55.140255
| 2019-11-14T16:44:54
|
522978361
|
{
"authors": [
"asf-ci",
"mhoppa",
"ocket8888"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3716",
"repo": "apache/trafficcontrol",
"url": "https://github.com/apache/trafficcontrol/pull/4114"
}
|
gharchive/pull-request
|
Rewrite current stats from Perl to Golang
What does this PR (Pull Request) do?
[x] This PR fixes #3853
This PR rewrites current_stats from Perl to Golang and adds API documentation for it.
Currently no tests as we do not have Traffic Stats available for TO Api tests.
Which Traffic Control components are affected by this PR?
Documentation
Traffic Golang Control Client
Traffic Ops
What is the best way to verify this PR?
Build traffic ops with PR code and then hit /current_stats, it should return the same data as perl implementation
If this is a bug fix, what versions of Traffic Control are affected?
The following criteria are ALL met by this PR
[x] This PR includes tests OR I have explained why tests are unnecessary
[x] This PR includes documentation OR I have explained why documentation is unnecessary
[x] This PR includes an update to CHANGELOG.md OR such an update is not necessary
[x] This PR includes any and all required license headers
[x] This PR ensures that database migration sequence is correct OR this PR does not include a database migration
[x] This PR DOES NOT FIX A SERIOUS SECURITY VULNERABILITY (see the Apache Software Foundation's security guidelines for details)
Additional Information
Still need to look into why Perl implementation did https://github.com/apache/trafficcontrol/blob/master/traffic_ops/app/lib/Utils/Helper.pm#L39 on each cdn name for the query and if that escaping is handled for us in the Golang client
Refer to this link for build results (access rights to CI server needed):
https://builds.apache.org/job/trafficcontrol-PR/4780/
So then is a work in progress or ready for review?
@ocket8888 this is ready for review. The perl code did replace on ,',",\n in the CDN names prior to going to influxdb but all are blocked from even being apart of a CDN name in the first place -> https://github.com/apache/trafficcontrol/blob/master/traffic_ops/traffic_ops_golang/cdn/cdns.go#L112
Refer to this link for build results (access rights to CI server needed):
https://builds.apache.org/job/trafficcontrol-PR/4808/
Refer to this link for build results (access rights to CI server needed):
https://builds.apache.org/job/trafficcontrol-PR/4847/
Refer to this link for build results (access rights to CI server needed):
https://builds.apache.org/job/trafficcontrol-PR/4848/
|
2025-04-01T06:37:55.152302
| 2020-11-15T08:52:13
|
743212671
|
{
"authors": [
"Dorjpalam",
"rob05c",
"ywkaras",
"zds05"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3717",
"repo": "apache/trafficserver",
"url": "https://github.com/apache/trafficserver/issues/7324"
}
|
gharchive/issue
|
Video streaming issue on traffic server
Hi ,
We are using traffic server as video streaming CDN. We have 2 server each 1,4TB SSD and 10G network interface. During peak hour video stream hard to watch and buffering in client side. Server's load average and Memory are higher than normal state. But streaming bandwidth is around 3Gbps on all 3 server total. We wondering that we have potential only 3Gbps bandwidth on streaming but it has high load on traffic servers. Does those server can handle this load?
load average 106.94, 93.30, 88.32 as 32 core CPU
memory usage total 62GB used 62GB
established tcp connection is around 1200 on each server
traffic server's error log is increasing as following
20201115.16h42m02s CONNECT:[0] could not connect [CONNECTION_CLOSED] to <IP_ADDRESS> for 'http://localhost/vod/encrypt/prod/8a01918b72167aad01722a8007db243e/8a01918b72167aad01722a8007db243e_1500_2/media-88320000.mp4'
20201115.16h42m02s CONNECT:[0] could not connect [CONNECTION_CLOSED] to <IP_ADDRESS> for 'http://localhost/vod/encrypt/prod/8a01918b752cbbb7017577a9a1fc05ba/8a01918b752cbbb7017577a9a1fc05ba_1500/media-46260040.mp4'
20201115.16h42m02s CONNECT:[0] could not connect [CONNECTION_CLOSED] to <IP_ADDRESS> for 'http://localhost/vod/encrypt/prod/8a01918b752cbbb7017549e7cc523b37/8a01918b752cbbb7017549e7cc523b37_1000_2/media-220416000.mp4'
20201115.16h42m02s CONNECT:[0] could not connect [CONNECTION_CLOSED] to <IP_ADDRESS> for 'http://localhost/vod/encrypt/prod/8a01918b69817ebf016a951962ab7ccc/8a01918b69817ebf016a951962ab7ccc_1500/media-4513608.mp4'
Best Regards,
how about disk util,and how much MB every disk,if each sever around 1200 on each server,it is abort 1.2G,the disk may be over used
loadavg is a very broad metric, which includes CPU, memory wait, disk wait, disk usage, and potentially other things.
Can you look at specific metrics on your system, and see what specific things have high load? Is it just CPU? Memory? Disk io_wait? All of the above?
ATS will use as much memory as you tell it to. You can allocate ramdisks and give them to ATS as block devices. Each disk given to ATS also has a memory cache in front of it, the size of which is configurable.
See:
https://docs.trafficserver.apache.org/en/8.0.x/admin-guide/files/records.config.en.html#ram-cache
https://docs.trafficserver.apache.org/en/8.0.x/admin-guide/files/storage.config.en.html
https://docs.trafficserver.apache.org/en/8.0.x/admin-guide/files/volume.config.en.html
ATS does have some known memory leaks, but they're generally pretty small. It shouldn't use much more memory than what you allocated for storage and ram_cache, and the memory shouldn't grow much over time.
Many people run ATS in production with bandwidth much higher than 3Gbps. My company has caches doing in excess of 20Gbps. If you're having trouble achieving those speeds, another possibility is Linux Kernel Parameters. It's common to have to do a lot of tuning of Linux Kernel Parameters to achieve high performance. Though I wouldn't expect a great deal of tuning to be necessary under 10Gbps.
I assume this is somewhat recent hardware, with decent CPUs? We do have some Prod servers that struggle to exceed 10Gbps, from underpowered CPUs with few PCI lanes. Platforms with too few PCI lanes can also cause network bottlenecks like that.
Some SSD brands have poor performance. It may be worth testing their sequential and random speeds, just to be sure that isn't the problem.
traffic server's error log is increasing as following
20201115.16h42m02s CONNECT:[0] could not connect [CONNECTION_CLOSED] to <IP_ADDRESS> for 'http://localhost/vod/encrypt/prod/8a01918b72167aad01722a8007db243e/8a01918b72167aad01722a8007db243e_1500_2/media-88320000.mp4'
I'm not sure I understand. Your initial question is about bandwidth bottlenecks and high loadavg, but this looks like an error? This looks like an origin (on localhost?) is misconfigured, or unable to handle the requests or load?
Are you saying you see a lot of these errors as you approach 3Gbps? That sounds like the Origin server isn't able to handle the load, that the problem is with the Origin, not ATS. Can you verify your Origin itself is capable of the request load?
Are these requests mostly Cache Hits or Misses? For a CDN, ATS should be caching I assume. Is the full traffic going to the Origin? Could that be causing the problem? Maybe the Origin can't handle the full 3Gbps, because everything is a Cache Miss, and you need to set Cache-Control to make ATS cache the content so the Origin can handle it.
Certain SSL Certificates can also cause high CPU usage, especially RSA. Are you using HTTPS?
In short, there are a huge number of factors that can cause bottlenecks like you're seeing. You'll have to narrow it down further, and inspect your hardware usage to figure out what the bottleneck is, and how to fix it. But ATS can definitely do +20Gbps, potentially even 100Gbps, and many large corporations are doing so in Production.
Please try to reproduce this in ATS 9.1, and reopen this issue if you still see it.
|
2025-04-01T06:37:55.154490
| 2024-03-27T18:10:52
|
2211476254
|
{
"authors": [
"bneradt",
"cmcfarlen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3718",
"repo": "apache/trafficserver",
"url": "https://github.com/apache/trafficserver/pull/11190"
}
|
gharchive/pull-request
|
swoc: install swoc_ip_util.h
Adding swoc_ip_util.h to the list of header sources so that it will be installed in the expected swoc include location.
[approve ci autest]
Cherry-picked to v10.0.x
|
2025-04-01T06:37:55.156130
| 2017-11-14T00:09:34
|
273623281
|
{
"authors": [
"bryancall",
"masaori335"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3719",
"repo": "apache/trafficserver",
"url": "https://github.com/apache/trafficserver/pull/2809"
}
|
gharchive/pull-request
|
Remove multiprocessing.Queue.qsize() from traffic_replay
Because this raise NotImplementedError on Unix platforms like Mac OS X.
Details in below.
http://python.readthedocs.io/en/stable/library/multiprocessing.html#multiprocessing.Queue.qsize
[ci approve autest]
[approve ci autest]
|
2025-04-01T06:37:55.160274
| 2016-07-02T11:34:02
|
163514600
|
{
"authors": [
"PSUdaemon",
"atsci",
"bryancall",
"mingzym"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3720",
"repo": "apache/trafficserver",
"url": "https://github.com/apache/trafficserver/pull/786"
}
|
gharchive/pull-request
|
TS-4396: fix number_of_redirections off-by-one
please test
FreeBSD build successful! See https://ci.trafficserver.apache.org/job/Github-FreeBSD/406/ for details.
Linux build successful! See https://ci.trafficserver.apache.org/job/Github-Linux/300/ for details.
FreeBSD build failed! See https://ci.trafficserver.apache.org/job/Github-FreeBSD/412/ for details.
Linux build successful! See https://ci.trafficserver.apache.org/job/Github-Linux/306/ for details.
[approve ci]
Closing since there is a new PR for this #2092
|
2025-04-01T06:37:55.166988
| 2016-06-28T18:23:00
|
162752800
|
{
"authors": [
"jongyoul",
"prabhjyotsingh"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3721",
"repo": "apache/zeppelin",
"url": "https://github.com/apache/zeppelin/pull/1103"
}
|
gharchive/pull-request
|
[ZEPPELIN-1076] Set hbase.client.retries.number for JDBC
What is this PR for?
If a user has "org.apache.phoenix:phoenix-core:4.x.x" jar added as a dependency in JDBC interpreter, and for some reason phoenix was not accessible or not properly configured; then the phoenix tries to for 35 times (which is default for hbase.client.retries.number) and each retires is 8 second apart, before it finally fails.
What type of PR is it?
[Bug Fix]
Todos
[x] - Set phoenix.hbase.client.retries.number for JDBC
What is the Jira issue?
ZEPPELIN-1076
How should this be tested?
In JDBC interpreter add org.apache.phoenix:phoenix-core:4.4.0-HBase-1.0 as dependency, but don't configure phoenix setting.
Then try to run any sql query with any of the configured JDBC driver (like show tables)
Without this it will take slightly more than about 5 mins
With this it should fetch result sooner (in less than a minute)
Screenshots (if appropriate)
Questions:
Does the licenses files need update? n/a
Is there breaking changes for older versions? n/a
Does this needs documentation? n/a
LGTM. That property hbase.client.retries.number will be passed to phoenix jdbc, right?
Yes, phoenix.hbase.client.retries.number, and have tested with and without this string.
Merging this if no more discussion.
|
2025-04-01T06:37:55.172940
| 2018-09-12T11:16:09
|
359433249
|
{
"authors": [
"Savalek",
"deradam",
"mebelousov"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3722",
"repo": "apache/zeppelin",
"url": "https://github.com/apache/zeppelin/pull/3179"
}
|
gharchive/pull-request
|
[ZEPPELIN-3773] - add check permission on write.
What is this PR for?
Sometimes when drawing the result of a paragraph, a call is made commit paragraph. And if the user does not have permission on write paragraph a window appears with a warning.
This PR fix it.
What type of PR is it?
Bug Fix
What is the Jira issue?
ZEPPELIN-3773
Questions:
Does the licenses files need update? no
Is there breaking changes for older versions? no
Does this needs documentation? no
@Savalek
I see an infinite number of GET request like http://localhost:8080/api/helium/suggest/2DMKVSPYC/20180807-154514_2063688624 with response "401 Unauthorized" is case if user don't have permission for write. Could you fix?
@mebelousov, Perhaps you have incorrectly configured shiro.ini.
@felixcheung, there is already a check on the server. Due to some errors, a commit of the paragraph went to the server and an error occurred.
@Savalek
You're right. I fix my shiro.ini.
But I still see a huge amount of background queries (up to 100 per second) connected with Helium.
I can reproduce the issue in development mode.
Tested on Ubuntu 16.04, Chromium (Version 69.0.3497.92) and Firefox Quantum 62.0
Check please.
This seems to be the bug that we are also facing currently. I have a notebook with some users that only have read permissions. These users receive an error message that the do not have the update permission on the notebook and therefore cannot open it.
Any estimate when this will be merged and released?
Need someone else to check this PR. Because I could not reproduce this error.
|
2025-04-01T06:37:55.183662
| 2024-09-29T08:49:38
|
2554816822
|
{
"authors": [
"shanshanying",
"shuoshadow"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3724",
"repo": "apecloud/kubeblocks-addons",
"url": "https://github.com/apecloud/kubeblocks-addons/issues/1070"
}
|
gharchive/issue
|
[Features] support create Kibana for Elasticsearch cluster
Kibana is an essential analysis and visualization platform for Elasticsearch.
Hi @shuoshadow ,
It is a good suggestion. We plan to support by the end of Oct.
|
2025-04-01T06:37:55.186791
| 2020-12-29T13:57:14
|
775905252
|
{
"authors": [
"antoine777",
"mmoayyed"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3725",
"repo": "apereo/cas",
"url": "https://github.com/apereo/cas/pull/5017"
}
|
gharchive/pull-request
|
Support one way hash for OAuth2 client secrets
Add one way hash function support for OAuth2 clientSecret storage mechanism as recommended by OWASP
Thank you for the pull request. It does not look like this would be something we'd want to accept just yet. If need does come up and there is time, we can review this again.
|
2025-04-01T06:37:55.192595
| 2017-08-22T17:22:03
|
252030529
|
{
"authors": [
"MrDys",
"atilling",
"jfritschi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3726",
"repo": "apereo/phpCAS",
"url": "https://github.com/apereo/phpCAS/issues/240"
}
|
gharchive/issue
|
Error ill-formed ticket found in the URL when ticket is encrypted
If the cas server is encrypting the ticket value (cas.ticket.security.cipherEnabled=true) the php client fails with the error: Error ill-formed ticket found in the URL
You have to set this property to false for the client to work.
Related to #180
I don't believe it is related to #180 because php CAS doesn't need to decrypt the ticket instead the criteria for what is a valid ticket needs to accept that the encrypted ticket is valid and just pass it back.
Can you please supply a debug log?
3424 .START (2017-08-22 12:11:13) phpCAS-1.3.5 ****************** [CAS.php:468]
3424 .=> phpCAS::client('3.0', 'castst.conncoll.edu', 443, 'cas') [index.php:12]
3424 .| => CAS_Client::__construct('3.0', false, 'castst.conncoll.edu', 443, 'cas', true) [CAS.php:360]
3424 .| | Starting a new session cq25i8qv8bbq8uqoiud8hh2g75 [Client.php:932]
3424 .| | Session is not authenticated [Client.php:938]
3424 .| | => phpCAS::error('ill-formed ticket found in the URL (ticket=eyJhbGciOiJIUzUxMiJ9.WlhsS05tRllRV2xQYVVwRlVsVlphVXhEU21oaVIyTnBUMmxLYTJGWVNXbE1RMHBzWW0xTmFVOXBTa0pOVkVrMFVUQktSRXhWYUZSTmFsVXlTVzR3TGk1NFYxaDNTbWhOY25KMFoxQm9aM1ExVkc4eGFGVjNMakJhZHpRdE0zcFRVV2hFUlhsaGNYVTVibWw1VlUxT05EazFURkpRWDNkRWVHcERXR0l3YUVGcldXSkdPWGxMTmpWRFIyazBiV2M1VUdocldWbHVVa0l1VVUxWmFtZFVUWGRLWm1VM2NVNVhabVk1VjBsclp3PT0.QCt2Ma0yxcfigVaNE5DYlwog1Vz8bIRB_EzoJjs85wWnXKCEwaxlvQKoIMU7C4HdxFbJya-Pj6URByRfpMwbsg\')') [Client.php:1028] 3424 .| | | ill-formed ticket found in the URL (ticket=eyJhbGciOiJIUzUxMiJ9.WlhsS05tRllRV2xQYVVwRlVsVlphVXhEU21oaVIyTnBUMmxLYTJGWVNXbE1RMHBzWW0xTmFVOXBTa0pOVkVrMFVUQktSRXhWYUZSTmFsVXlTVzR3TGk1NFYxaDNTbWhOY25KMFoxQm9aM1ExVkc4eGFGVjNMakJhZHpRdE0zcFRVV2hFUlhsaGNYVTVibWw1VlUxT05EazFURkpRWDNkRWVHcERXR0l3YUVGcldXSkdPWGxMTmpWRFIyazBiV2M1VUdocldWbHVVa0l1VVUxWmFtZFVUWGRLWm1VM2NVNVhabVk1VjBsclp3PT0.QCt2Ma0yxcfigVaNE5DYlwog1Vz8bIRB_EzoJjs85wWnXKCEwaxlvQKoIMU7C4HdxFbJya-Pj6URByRfpMwbsg') in /cwd/cwassets/httpd/alias/tp/cas/cas5-php-test/index.phpon line 12 [CAS.php:566]
3424 .| | <= ''
Thanks, looks like we simply need to adjust the regexp security filter for the ticket so that it allows all formats including the new encryption.
Has there been any movement on this? I'm running against master, with all of the fixes for ticket/session length (#248, #257, and #224), but enabling cas.ticket.security.cipherEnabled still throws an error.
|
2025-04-01T06:37:55.199784
| 2017-08-09T16:00:52
|
249080454
|
{
"authors": [
"cgarvis",
"jamesramsay",
"tj"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3727",
"repo": "apex/up",
"url": "https://github.com/apex/up/issues/200"
}
|
gharchive/issue
|
Add UP_STAGE environment variable
Technically I think it's more correct to set env vars for the environment, not based on the env such as NODE_ENV. However we need it for logging anyway.
This may have to be a hack, setting the env on first request, since stageVariables is the only way to get at the stage (AFAIK).
It would be handy for UP_STAGE to be made available to the hooks too, allowing particularly the build script to adjust it's behaviour accordingly.
I'm not sure if this issue refers to the application environment (in AWS) only, or making UP_STAGE available everywhere including local environment when building.
Hmm yeah it's a little tricky, since the "ideal" way of deploying would be to stage the application, then promote the staged version to production. UP_STAGE would will be "staging" at that point
Promoting the stage build to production doesn't work for statically built files like next.js does.
@cgarvis yea not if you differentiate config in those two stages. It depends how you structure things I suppose, since you can pass UP_STAGE to your client JS for example and choose config that way.
It would be a more ideal way to deploy since you know exactly what you're getting.
@tj I use webpack's DefinePlugin to replace process.env in my builds. That way I don't leak environment vars by accident. Also makes it more difficult for someone on the team to hit the wrong endpoints.
|
2025-04-01T06:37:55.206824
| 2021-03-14T05:10:33
|
831067091
|
{
"authors": [
"egonelbre",
"kaihendry",
"tj"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3728",
"repo": "apex/up",
"url": "https://github.com/apex/up/issues/825"
}
|
gharchive/issue
|
GLIBC_2.32 not found / CGO_ENABLED=0
Prerequisites
[X] I am running the latest version. (up upgrade)
[X] I searched to see if the issue already exists.
[X] I inspected the verbose debug output with the -v, --verbose flag.
[X] Are you an Up Pro subscriber?
Description
On Arch glibc is not compatible with AWS's. So the workaround is to use CGO_ENABLED=0
However CGO_ENABLED=0 often means poor support with sqlite drivers or example. CGO_ENABLED=0 versions of sqlite drivers exist but they are significantly slower for example.
Perhaps it's having some sort of cross compile option as suggest here https://twitter.com/benbjohnson/status/1370955506471165956 ?
Steps to Reproduce
Try deploy https://github.com/kaihendry/dfts
Hmm I'm not sure I'd want a dependency on Docker, it could get really complicated quick. I think that sort of thing is probably best left to CI like GH Actions where you're building in the target environment already, but something to think about for sure
I'm told this what I need to do: https://gist.github.com/egonelbre/01bbf7ca97d6b5588438da36a2578e7b
I do really want a smooth way to deploy locally without CI.
I have no previous experience with apex/up, but as far as I'm able to deduce it seems that hooks.build is a way to override with arbitrary build command? So it should be possible to pretty much use the same invocation -- maybe modify it to output "server" as the binary. Although, I'm not sure in which computer/server context will run.
The only concern is invoking subcommands such as $(pwd) and $(go env GOPATH) within the command. Maybe it can invoke a separate script? Based on https://github.com/apex/up/blob/9770c5062e3a39563d183f84ce51cec52bf683ec/up.go#L67, I'm guessing it will work?
@egonelbre yep it runs as a shell command, so you should be fine to use stuff like that
@kaihendry yeah I definitely hear you, but the number of variables explodes once you start introducing platform specifics. Using Docker manually locally and putting Up inside would more or less have the same UX as Up trying to use Docker, so I'm not sure there would be much of a benefit, since Up would have to support all the various OSes and shared libraries etc
|
2025-04-01T06:37:55.211942
| 2020-04-06T13:16:28
|
595107157
|
{
"authors": [
"jkohlin",
"junedchhipa",
"oherik"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3729",
"repo": "apexcharts/apexcharts.js",
"url": "https://github.com/apexcharts/apexcharts.js/issues/1472"
}
|
gharchive/issue
|
Always have the fill below the curve in area charts
I'm creating trend line widgets, showing the angle of the current trend line (another request btw) for a given time period. Some chart values are however negative and that makes the chart "flip" the area or even worse, cross the line. Like in this codepen:
https://codepen.io/jkohlin/pen/BaNgbOy
What I want is an option for fill that might look like this:
fill: {
colors: '#0000ff',
opacity: 0.9,
type: 'gradient',
fillTo: -20,
}
Where fillTo is a numeric value that, if set, forces the fill to flow all the way down to that value on the y-axis.
Then you could get this:
instead of this:
Describe alternatives you've considered
At the moment I have to recalculate the y values and add a positive value big enough to force the curve above zero. This works on sparklines, where the y-axis is hidden, but not otherwise of course.
This would be an awesome feature!
I would, too, like to request this. As of now it looks rather strange for negative values, if the area has a gradient:
. It makes it quite difficult to see the lowest values.
Sorry for the late response.
I have added a new option that will extend the area beneath zero-line and fill it till the end.
plotOptions: {
area: {
fillTo: 'end'
}
}
This will be available from v3.19.3
|
2025-04-01T06:37:55.214824
| 2019-04-22T04:55:28
|
435598895
|
{
"authors": [
"bashairm",
"junedchhipa"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3730",
"repo": "apexcharts/apexcharts.js",
"url": "https://github.com/apexcharts/apexcharts.js/issues/533"
}
|
gharchive/issue
|
x-axis labels are not appears
Codepen
https://codepen.io/anon/pen/YMjPEa
added to code-pen
Explanation
What is the behaviour you expect?
the x-axis labels are too long
What is happening instead?
so it is not appearing correctly.
What error message are you getting?
half of it replaced with ... and I need it all to be visible
thanks
Increase the yaxis labels maxWidth by
yaxis: {
labels: {
maxWidth: 200
}
}
|
2025-04-01T06:37:55.252365
| 2022-03-18T12:22:00
|
1173532557
|
{
"authors": [
"acenolaza",
"vponline"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3731",
"repo": "api3dao/airkeeper",
"url": "https://github.com/api3dao/airkeeper/pull/44"
}
|
gharchive/pull-request
|
Airnode functions refactor
This updates wallet functions, loadConfig, logger, getGasPrice, callApi to be imported from airnode-node v0.5
Note: to get this to build locally, you will need to copy the built airnode-node and airnode-utilities /dist folders from Airnode PR 944 into your Airkeeper node_modules/@api3
This probably shouldn't be merged until we can update to v0.5 and then we can temporarily update the package.json to use airnode-node for PR944 as a dependency with:
'git+https://gitpkg.now.sh/api3dao/airnode/packages/airnode-node?ec0340aabc3be174e62f315b9cc0bbe62b43d0da' using this which allows us to use a monorepo subdirectory as a package
@acenolaza Do the example configs work for you in the main branch? I had to change the airnodeAddress in airkeeper.json (these changes) to make them work.
I'm actually using 0xA30CA71Ba54E83127214D3271aEA8F5D6bD4Dace because in secrets.env I have the test mnemonic we are using everywhere else
|
2025-04-01T06:37:55.257928
| 2022-04-07T17:49:39
|
1196376089
|
{
"authors": [
"acenolaza",
"amarthadan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3732",
"repo": "api3dao/airseeker",
"url": "https://github.com/api3dao/airseeker/pull/14"
}
|
gharchive/pull-request
|
BEC258: Terraform ECS
I'm creating this as a WIP because I wanted to get feedback as soon as possible and also because the Airseeker docket image is not yet in docker hub.
Some notes:
Created an ecs cluster, task definition and service to tie them together.
Since I'm using FARGATE I had to add a vpc resource but since we are not exposing a frontend I didn't think it was necessary to deal with internet gateways, load balancers and security groups. The simplest way I found to add a vpc is to just define a default vpc resource with default subnets.
Since I wanted to add logs to the task (awslogs set in the Task Definition) I had to also attach the AmazonECSTaskExecutionRolePolicy.
I've hardcoded the airnode-client docker image url just for testing until airseeker image is ready. Once that happens I guess I'll replace it with a variable.
I wasn't sure if we needed to deploy to different stages so I decided to keep it simple and have all terraform files in the same directory. Also I have not created any modules until I see the need for them.
I kept it as a Draft PR because I wasn't sure if airseeker-dev was the right image to use. Also, running docker pull api3/airseeker-dev:latest doesn't seem to work for me. I need to use a specific task in order to be able to pull the image from docker hub. Same thing happens when I use latest to pull airkeeper image. Maybe @aquarat has any idea what might be happening thinking I did found something about having to set some config to true in docker hub in order to be able to pull using latest.
I think the problem is that there is no @api3/airseeker-dev tagged as latest :smile: Set the default to @api3/airseeker:0.1.0 (even though it doesn't exist yet). You can pass a dev one (with commit hash) as an variable to test that it's actually working.
@amarthadan I feel like this is much better now and it's ready for final review.
@amarthadan I feel like this is much better now and it's ready for final review.
Yes, I'll review it later today :slightly_smiling_face:
I've rebased this branch on top of main and I started getting an issue with the use of console.log in scripts/terraform-fmt.ts which I fixed by just putting the whole scripts folder in the eslintignore file because those scripts are only used for development.
|
2025-04-01T06:37:55.262669
| 2015-06-01T20:50:21
|
83687274
|
{
"authors": [
"danielgtaylor",
"edwardaa",
"freezy-sk",
"jeanregisser",
"pksunkara",
"zdne"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3733",
"repo": "apiaryio/api-blueprint-sublime-plugin",
"url": "https://github.com/apiaryio/api-blueprint-sublime-plugin/issues/23"
}
|
gharchive/issue
|
Incorrect highlighting when MSON attribute description contains underscore
See apiaryio/api-blueprint#202. Sample at line 126.
Same problem here with wrong rendering
+ Request Client credentials (application/json)
+ Attributes
+ grant_type: `client_credentials` (string, required)
+ client_id: `88888888-4444-4444-4444-cccccccccccc` (string, required)
+ client_secret: `clientsecret` (string, required)
+ Body
{
"grant_type": "client_credentials",
"client_id": "88888888-4444-4444-4444-cccccccccccc",
"client_secret": "clientsecret"
}
@danielgtaylor Now that I look at this with fresh eyes, this might not be an issue at all, since underscores in MSON are not allowed and they need to be escaped. If that rule is followed, the highlighter would not make any mistake.
MSON are not allowed and they need to be escaped
This is true for property names and values. Not for description.
Same hightlighting issue happens if you have a JSON payload that contains a key that starts with an underscore.
For instance:
## Videos Collection [/videos]
Provides access to all videos.
+ Model (application/json)
JSON representation of videos
+ Body
{
"_pagination":
{
"next": "/videos?offset=7",
"total_count": 100
}
}
Highlighting is incorrect starting from the underscore.
any update?
|
2025-04-01T06:37:55.267869
| 2018-08-07T16:25:28
|
348395511
|
{
"authors": [
"honzajavorek",
"kylef"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3734",
"repo": "apiaryio/dredd-example",
"url": "https://github.com/apiaryio/dredd-example/pull/29"
}
|
gharchive/pull-request
|
Add Wercker and more
Summer cleanup 😎
Preparing for CircleCI 2.0 support in Dredd's dredd init
Adding Wercker
Using simpler naming for several files
Being more explicit in the README about where all the files are, to which service they belong, etc.
Adding Jenkins to the README even though it's not live-tested
Swagger is dead, long live the OpenAPI 2 (preparing for OpenAPI 3 support in Dredd and elsewhere)
@nadade @kylef Call for help 🙋♂️
I'm trying to make Wercker to npm install --no-optional, but the no optional flag seems to be ignored - it still installs protagonist:
...
export WERCKER_NPM_INSTALL_OPTIONS="--no-optional"
source "/pipeline/npm-install-83ebc85d-4d44-47f3-a21f-75237025acae/run.sh" < /dev/null
Using wercker cache
Creating $WERCKER_CACHE_DIR/wercker/npm
Configuring npm to use wercker cache
Starting npm install, try: 1
npm WARN deprecated<EMAIL_ADDRESS>Broken not support
npm WARN deprecated<EMAIL_ADDRESS>Package no longer supported. Contact<EMAIL_ADDRESS>for more info.
><EMAIL_ADDRESS>install /pipeline/source/node_modules/protagonist
> node-gyp rebuild
...
Even if I tried to create my own step, it was still ignored and Wercker just installed the project as without --no-optional, compiling protagonist. Any ideas?
@honzajavorek I think this may be NPM bug: https://npm.community/t/npm-install-no-optional-not-actually-filtering-optionals-in-cli-6-0-1-or-6-1-0/257 / https://github.com/npm/npm/issues/17633#issuecomment-403938408
Perhaps you can work around the issue with having a package-lock.json generated without protagonist. Probably something you should do regardless due to transient dependency licensing approval.
Oh, package-lock.json is a different can of worms I'll need to resolve later. Thanks for looking into it, I think I'm fine with Wercker installing protagonist now.
|
2025-04-01T06:37:55.293395
| 2024-07-21T19:59:47
|
2421580707
|
{
"authors": [
"Jourdelune",
"janbuchar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3735",
"repo": "apify/crawlee-python",
"url": "https://github.com/apify/crawlee-python/issues/337"
}
|
gharchive/issue
|
httpx.InvalidURL: Invalid non-printable ASCII character in URL
Hey, I try to scrap music but it seems that the crawler with the await context.enqueue_links(strategy="all") add invalid url, I run my code but I have the error:
[crawlee.autoscaling.autoscaled_pool] INFO Waiting for remaining tasks to finish
Traceback (most recent call last):
File "/home/jourdelune/dev/Crawler/src/main.py", line 21, in <module>
asyncio.run(main())
File "/usr/lib/python3.10/asyncio/runners.py", line 44, in run
return loop.run_until_complete(main)
File "/usr/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete
return future.result()
File "/home/jourdelune/dev/Crawler/src/main.py", line 14, in main
await crawler.run(
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/basic_crawler/basic_crawler.py", line 359, in run
await run_task
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/basic_crawler/basic_crawler.py", line 398, in _run_crawler
await self._pool.run()
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/autoscaling/autoscaled_pool.py", line 185, in run
await run.result
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/autoscaling/autoscaled_pool.py", line 336, in _worker_task
await asyncio.wait_for(
File "/usr/lib/python3.10/asyncio/tasks.py", line 408, in wait_for
return await fut
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/basic_crawler/basic_crawler.py", line 734, in __run_task_function
await self._commit_request_handler_result(crawling_context, result)
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/crawlee/basic_crawler/basic_crawler.py", line 653, in _commit_request_handler_result
destination = httpx.URL(request_model.url)
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/httpx/_urls.py", line 115, in __init__
self._uri_reference = urlparse(url, **kwargs)
File "/home/jourdelune/dev/Crawler/env/lib/python3.10/site-packages/httpx/_urlparse.py", line 163, in urlparse
raise InvalidURL("Invalid non-printable ASCII character in URL")
httpx.InvalidURL: Invalid non-printable ASCII character in URL
the invalid url is: https://www.linkedin.com/company/nic-br/
Code:
import re
import urllib.parse
from crawlee.basic_crawler import Router
from crawlee.beautifulsoup_crawler import BeautifulSoupCrawlingContext
from crawlee.playwright_crawler import PlaywrightCrawlingContext
router = Router[PlaywrightCrawlingContext]()
regex = r"https?:\/\/(www\.)?[-a-zA-Z0-9@:%._\+~#=]{1,256}\.[a-zA-Z0-9()]{1,6}\b([-a-zA-Z0-9()!@:%_\+.~#?&\/\/=]*)\.(mp3|wav|ogg)"
@router.default_handler
async def default_handler(context: BeautifulSoupCrawlingContext) -> None:
url = context.request.url
html_page = str(context.soup).replace("\/", "/")
matches = re.finditer(regex, html_page)
audio_links = [html_page[match.start() : match.end()] for match in matches]
for link in audio_links:
link = urllib.parse.urljoin(url, link)
data = {
"url": link,
"label": "audio",
}
await context.push_data(data)
await context.enqueue_links(strategy="all")
hey, thanks you for the answer, the code that imports the router:
"""
main script for the crawler
"""
import asyncio
from crawlee.beautifulsoup_crawler import BeautifulSoupCrawler
from routes import router
from utils import process
async def main() -> None:
"""
Function to launch the crawler
"""
crawler = BeautifulSoupCrawler(
request_handler=router,
)
await crawler.run(
["https://www.cgi.br/publicacao/revista-br-ano-07-2016-edicao-09/"]
)
await crawler.export_data("results.json")
process("results.json")
if __name__ == "__main__":
asyncio.run(main())
I want to crawl the full web to create a dataset of song url (to create an AI music generation model), that why I use strategy="all",
if you run the code, you should get the error.
The url where it get the invalid url is: https://www.cgi.br/publicacao/revista-br-ano-07-2016-edicao-09/
the invalid url is: https://www.linkedin.com/company/nic-br/
Huh, this is getting interesting. I added this to the request handler:
context.log.info(f'links found: {"\n".join([repr(link.attrs.get('href')) for link in context.soup.select('a')])}', )
...and it showed me that the linkedin link in fact contains a line break:
<a class="btn-floating btn-lg btn-li" type="button" role="button" href="https://www.linkedin.com/company/nic-br/
" target="_blank">
<i class="fab fa-linkedin-in"></i>
</a>
However unusual this is, I'll add a .strip() to the enqueue_links implementation.
|
2025-04-01T06:37:55.297635
| 2022-07-06T21:14:03
|
1296481662
|
{
"authors": [
"shrutiparabgoogle"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3736",
"repo": "apigee/registry",
"url": "https://github.com/apigee/registry/issues/643"
}
|
gharchive/issue
|
Styleguides: improvements in styleguide/conformance protos
With the aim of displaying the lint results in the UI, the conformance report protos need the following updates:
The spec field should include the revision ID of the spec.
RuleReport should include the metadata of the rule defined in the styleguide. The metadata fields which should be included in the rule are as follows:
display_name
description
doc_uri
cc @michaelyara
|
2025-04-01T06:37:55.301609
| 2016-02-01T11:56:02
|
130342263
|
{
"authors": [
"brylie",
"elnzv"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3737",
"repo": "apinf/api-umbrella-dashboard",
"url": "https://github.com/apinf/api-umbrella-dashboard/pull/825"
}
|
gharchive/pull-request
|
Feature/api metadata i18n. Closes #779
Added i18n strings to view metadata page. This does not include AutoForm label translations, as that currently seems non-trivial.
@elnzv will you please review thie PR?
It may not be necessary to add the translation strings to the fi.i18n.json. Did you read somewhere that this is a requirement?
@brylie, no, that is not a requirement, but just more comfortable to translate, no need to search for new lines in en.json.
Anyway, I added those myself, resolved conflicts.
Merging.
|
2025-04-01T06:37:55.329689
| 2023-03-06T17:28:45
|
1611897422
|
{
"authors": [
"aplteam",
"rikedyp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3738",
"repo": "aplteam/Cider",
"url": "https://github.com/aplteam/Cider/issues/18"
}
|
gharchive/issue
|
Requires Tatin to intialise before opening project
Don't know whether this is documented and I missed it, but
]CIDER.OpenProject \g\ParquetDotNet
* Command Execution Failed: VALUE ERROR: Undefined name: Tatin
goes away if I first do
]TATIN.Version
┌─────┬───────────┬──────────┐
│Tatin│0.90.0+1485│2023-02-27│
└─────┴───────────┴──────────┘
I would expect not to have to call a different UCMD before being able to use Cider
Tatin it's will be part of a 19.0 standard installation. In older versions it's up to the user to make sure that it's loaded into ⎕SE.
However, the error message should be more explicit about what's required.
Changed my mind. Solved in version 0.23.4
|
2025-04-01T06:37:55.331178
| 2011-12-03T17:50:31
|
2438309
|
{
"authors": [
"dgm"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3739",
"repo": "apneadiving/Google-Maps-for-Rails",
"url": "https://github.com/apneadiving/Google-Maps-for-Rails/issues/124"
}
|
gharchive/issue
|
proxy support
A production server I use has outgoing connections firewalled, and we have to use a proxy. This plugin needs to be able to connect via the proxy for geocoding requests, similar to https://rails.lighthouseapp.com/projects/8994/tickets/2133-activeresource-http-proxy-support
#126 has code attached
|
2025-04-01T06:37:55.334788
| 2016-11-01T21:35:46
|
186654251
|
{
"authors": [
"apocas",
"moonlitSpider"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3740",
"repo": "apocas/dockerode",
"url": "https://github.com/apocas/dockerode/issues/310"
}
|
gharchive/issue
|
Using exec.start()
Hi again,
So I read the examples you referred me to in issue #309. I've even "sort of" run your example. Specifically, rather than create an instance of an Ubuntu container, I did a
docker.getContainer(ID)
The container in question is running MongoDB and I'd like to be able to present to its shell (bash) several commands and get their responses. But I am uncertain about how to accomplish this. As matters stand, based on your example, I can see the usual console messages from MongoDB as its shell starts. These were raised in response to the original Cmd, 'mongo', presented by
container.exec()
But now I need to present a few more commands to the MongoDB shell & parse the responses. These commands don't reside in a file. They are strings defined in the program itself.
Would you be so kind as to provide an example of how to do this with dockerode's exec facility?
Thanks.
Cordially,
Paul
Hmmm...maybe hold off for a bit on any kind of answer. I may have gotten it working. The "trick" appears to be the use of 'hijack: true', 'stdin: true' in the exec.start options as well as 'AttachStdin: true' in the container.exec options.
👍
|
2025-04-01T06:37:55.336534
| 2017-03-12T03:39:51
|
213578998
|
{
"authors": [
"apocas",
"knight42"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3741",
"repo": "apocas/dockerode",
"url": "https://github.com/apocas/dockerode/pull/344"
}
|
gharchive/pull-request
|
Fix inconsistent behaviour of Container.inspect
container.inspect() is supposed to return a Promise not a stringified object
Yup legacy pre-promise code left behind.
Could you please do the same for the other object's inspects? (image, network, etc)
So everything is consistent.
@apocas done!
Published v2.4.0
|
2025-04-01T06:37:55.407059
| 2017-06-28T17:03:43
|
239231375
|
{
"authors": [
"lewisf",
"martijnwalraven",
"tgriesser"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3742",
"repo": "apollographql/apollo-codegen",
"url": "https://github.com/apollographql/apollo-codegen/pull/156"
}
|
gharchive/pull-request
|
TypeScript optional inputs
Similar to #155, but for TypeScript, fixes #81
@tgriesser @lewisf: I'm sorry I'm only getting to this now, but I wanted to publish a new release. It turns out some of the TypeScript snapshot tests are failing however, and it seems the results do not match the expected behavior in this PR. Could you have a look at this?
@martijnwalraven will take a look now
@martijnwalraven this actually looks like a small oversight. Some changes weren't pulled in. I opened up #162 to get the tests passing
|
2025-04-01T06:37:55.448483
| 2022-05-14T12:37:56
|
1235963987
|
{
"authors": [
"denys-meloshyn",
"martinbonnin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3743",
"repo": "apollographql/apollo-kotlin",
"url": "https://github.com/apollographql/apollo-kotlin/issues/4108"
}
|
gharchive/issue
|
OkHttpExecutionContext replacement
Question.
OkHttpExecutionContext was available in version 2 after migration to 3.3.0 this context key was removed and I can't find another way how to get okhttp response from ApolloResponse
Hi 👋 Thanks for reaching out
In 3.x you can use response.executionContext[HttpInfo]. That will give you the status code and headers from your HTTP call. If you need access to the body, I'd recommend using an HttpInterceptor so that you can read the body as it is streamed. If you need something else, let us know and we can investigate together how to get it.
thanks @martinbonnin for quick reply! The reason why I need okhttp Response is because we have backend driven logic when automatically refresh page.
This is an example how we did it previously.
apolloClient.query(...).rx().singleOrError().map {
val okHttpResponse = it.executionContext[OkHttpExecutionContext]?.response
val maxAgeInSeconds = okHttpResponse?.cacheControl?.maxAgeSeconds ?: 0
}
I see, thanks for providing the details!
Looking at the OkHttp code, looks like it's getting this value from a header so you could do something like:
val maxAgeInSeconds = response.executionContext[HttpInfo]
?.headers
?.firstOrNull {
it.name.lowercase() == "max-age"
}
?.value
?.toIntOrNull()
Could that work?
That's so neat! thank you for help and have a nice weekend 😊
Thank you, you too :) !
|
2025-04-01T06:37:55.453460
| 2024-01-04T15:02:27
|
2065815228
|
{
"authors": [
"damianpetla",
"martinbonnin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3744",
"repo": "apollographql/apollo-kotlin",
"url": "https://github.com/apollographql/apollo-kotlin/issues/5506"
}
|
gharchive/issue
|
Firing queries from Android Studio plugin with @nonull directive
Question
Hey, I am using Apollo plugin with Android Studio and I used to test my graphql files directly from IDE. Noticed that adding @nonull does not allow plugin to work and returns error message "message": "directive 'nonnull' is not defined in the schema"
The question is if that can be somehow workaround?
Can you try enabling the "Apollo Kotlin" framework in the GraphQL plugin?
Hey @martinbonnin I had this enabled
IDE support is still being worked on (see https://github.com/JetBrains/js-graphql-intellij-plugin/issues/697) so it'll come but the red underlines for @semanticNonNull are still expected at this stage.
I would still expect @nonnull to be detected though(see https://github.com/JetBrains/js-graphql-intellij-plugin/blob/21d5800921d07176992e3d006728a89a1c1eb242/resources/definitions/ApolloKotlin.graphql#L23)
@nonnull works just fine when editing graphql file, there is support for that. It just fails when executing graphql queries from IDE
Thanks. They work just fine from the app. Been using it for some time now on prod. It's just very convenient sometimes to verify something right from the IDE.
I created https://github.com/apollographql/apollo-kotlin/issues/5507 as a follow up before I forget.
|
2025-04-01T06:37:55.458503
| 2024-05-15T16:18:34
|
2298341080
|
{
"authors": [
"BoD",
"enggazzar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3745",
"repo": "apollographql/apollo-kotlin",
"url": "https://github.com/apollographql/apollo-kotlin/issues/5891"
}
|
gharchive/issue
|
Crash when use policy
Version
3.8.2
Summary
override suspend fun getMerchantsByKeyWord(keyword:String): Flow<Response> {
return apolloClient.query(MerchantsByKeyWordQuery(keyword)).fetchPolicy(FetchPolicy.CacheAndNetwork)
.toFlow()
.map { response ->
try {
//response.data.merchants
Response.Success(response.data?.toRestaurantsSearch() as T)
} catch (e: Exception) {
Response.Failure( e.message ?: "unknown error")
}
}
}
Steps to reproduce the behavior
No response
Logs
(Your logs here)
Hi! It looks like you're having a network issue. This can happen if your device has a connectivity issue, or e.g. if the backend is not responding.
I enabled working offline from cache policy
CacheAndNetwork will both go to the cache and the network, which will throw if the network fails.
I disable wifi to test app working offline
In that case what you're seeing is definitely expected. You should probably add a .catch {} call to the Flow chain, to handle the exception.
|
2025-04-01T06:37:55.471275
| 2017-02-17T18:46:05
|
208523916
|
{
"authors": [
"ajs139",
"helfer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3746",
"repo": "apollographql/graphql-tools",
"url": "https://github.com/apollographql/graphql-tools/issues/281"
}
|
gharchive/issue
|
Not possible to control the types of items in a list
Given a field that returns an array that composed of Union or Interface types, it is not possible to control the type of each item in the list. For example, given the following schema:
union AdvertOrPost = Advert | Post
enum SupportedTypes {
Advert
Post
}
type Advert {
id: ID!
productName: String!
}
type Post {
id: ID!
text: String!
}
type RootQuery {
getAll(type: SupportedTypes): [AdvertOrPost]
}
schema {
query: RootQuery
}
It should be possible to apply the following mocks:
const mockMap = {
RootQuery: () => ({
getAll: (o: any, a: { [key: string]: any }) => new MockList(2, () => ({typename: a['type']})),
}),
Advert: () => ({
productName: 'supercoolproduct',
}),
Post: () => ({
text: 'superlongpost',
}),
};
However, using the above code would result in objects that do not have the type-specific mocks applied, for example a result to querying getAll(type: Advert) would look like this:
{
"getAll": [
{"__typename": "Advert", "productName": "Hello World"},
{"__typename": "Advert", "productName": "Hello World"},
]
}
@ajs139 Sorry for the late reply. I think in order for the server to know which mock to apply, your schema has to define the __resolveType or __ofType functions. I might also be wrong however, in which case a failing test case would be much appreciated.
Sure, please see https://github.com/apollographql/graphql-tools/pull/282/files#diff-b9774ff344f81ba2175bf0fb4973ac7a for a test (this PR also contains a fix, but I closed it because it didn't deal with a nested array return type (e.g. someField: [[Foo]]) and haven't had the chance to revisit.
I think this should work in 1.1.0.
|
2025-04-01T06:37:55.506877
| 2017-11-08T08:54:32
|
272119610
|
{
"authors": [
"bvandewalle",
"sibicramesh"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3747",
"repo": "aporeto-inc/trireme-statistics",
"url": "https://github.com/aporeto-inc/trireme-statistics/pull/33"
}
|
gharchive/pull-request
|
New: Added Unit Tests for Graph Generators
**Merge PR https://github.com/aporeto-inc/trireme-statistics/pull/31 before merging this PR **
--> Added unittests for server functions
--> Added mock for influxdb
Thanks for addressing those changes.
I will let you merge this PR (whenever you feel it is in a ready state)
|
2025-04-01T06:37:55.511625
| 2024-06-27T19:48:36
|
2378974306
|
{
"authors": [
"OliviaHY"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3748",
"repo": "app-sre/qontract-schemas",
"url": "https://github.com/app-sre/qontract-schemas/pull/682"
}
|
gharchive/pull-request
|
Add a new field subnets for machine pool
They can span across more than one subnet for multi-az cluster. This way it's backwards compatible until we are ready to migrate to this new field.
So this will not be useful as we remove the default machine pool which creates nodes in all subnets then create one machine pool designated for each AZ/subnet, and there is no way to provide more than one subnet ids when creating a machine pool through ROSA CLI and OCM API.
|
2025-04-01T06:37:55.516144
| 2018-01-10T20:04:23
|
287560137
|
{
"authors": [
"siddharthlatest"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3749",
"repo": "appbaseio/reactive-manual",
"url": "https://github.com/appbaseio/reactive-manual/issues/65"
}
|
gharchive/issue
|
React-Native: Add Documentation
[x] We need to create a version switcher for React Native. Using a different repo
[x] Getting Started Section
[x] Install and Quickstart
[x] Import Data
[x] ReactiveBase
[x] Components Overview
[x] Component References
One doc per component
[x] Advanced
[x] Add a snack snippet for the ReactiveComponent doc.
[x] Performance Tips @metagrover
[x] Creating Themes, Styles Unsupported in current version
[ ] Link innerStyle to all the components
[ ] Link innerProps to all the components
[ ] Update theme prop for ReactiveBase.
[ ] Add a doc on how to publish the app - With expo and as an ejected app.
|
2025-04-01T06:37:55.571091
| 2016-01-15T14:41:48
|
126885249
|
{
"authors": [
"Astro03",
"BaN4NaJ0e",
"Jonahss",
"TikhomirovSergey",
"trbnb"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3750",
"repo": "appium/appium-dotnet-driver",
"url": "https://github.com/appium/appium-dotnet-driver/pull/132"
}
|
gharchive/pull-request
|
New Zoom/Pinch implementation
After working with the zoom and pinch functionalities I made in my last PR we (myself and a colleague) came across some inconsistencies.
We figured that iOS and Android use very different approaches to these gestures, so we worked on a new implementation for these that works much more consistent.
do you want to take out the generic approach all together?
Well, the problem is that the Appium server needs different parameters for the different platforms. So the only alternative to this approach would be to define an interface with all of those parameters. But this way one implementation would just ignore a half of these parameters and the other implementation would ignore the other half. And I think this doesn't really add anything useful to the code.
Was there something wrong with the multi touch way of doing things instead
of running the script directly?
On Fri, Jan 15, 2016 at 1:30 PM, trbnb<EMAIL_ADDRESS>wrote:
Well, the problem is that the Appium server needs different parameters for
the different platforms. So the only alternative to this approach would be
to define an interface with all of those parameters. But this way one
implementation would just ignore a half of these parameters and the
other implementation would ignore the other half. And I think this
doesn't really add anything useful to the code.
—
Reply to this email directly or view it on GitHub
https://github.com/appium/appium-dotnet-driver/pull/132#issuecomment-172097985
.
Yes, after observing it a bit we found it to be quite unreliable. Especially on iOS we sometimes struggled to make it work.
I'll review this. I've some remark...
Sorry for the late responce, I was busy
I have a a question to @Jonahss. Let him to take a look at your implementation. I have some question to the original code... If it is correct then it makes sense to port your code to Java client.
I'm not really sure how because the gestures don't change the size the view or something like this, just how the content is presented and I don't think that's something we can retrieve with Appium.
Any news on this?
Waiting for @Jonahss (I've told about this PR) because I think that this change could be ported to Java client too.
As I know current gestures were ported from Java client. I dont know why original code was designed the way as it is now. But it is most likely I'm Ok with your changes.
Ok, good to know.
And yes, these gestures were initially ported from the Java client, I made the pull request for that a few weeks/months ago.
@trbnb
i'm ok with your changes.
There are tons of work with appium node 1.5.x. I'm working on java client now. You fix is supposed to be included in <IP_ADDRESS>. There will be many fixes related to node 1.5.x.
Alright, thanks for letting me know.
Just for your information. We (trbnb and myself) ported the new pinch and zoom behavior from the python appium driver. More information can be found here:
https://github.com/appium/python-client/blob/master/appium/webdriver/webdriver.py
Be careful. They are calling pinch and zoom in the code a little bit different. It is internally called pinchClose and pinchOpen. But it does the same.
It is a good idea to port our PR to the java bindings too. Pinch and Zoom is broken there too. We initialy ported the pinch and zoom command from the java bindings to the dotnet bindings without testing if it worked in the Java bindings. Seems that it did not work under iOS at all. So thats the reason we did it again. This time we ported it from the python bindings. And now it should work on Android and iOS.
Guys,
I'm going to play around with proposed changes. I think I'll propose a PR with your authorship if everything is ok.
However, I'm not sure that it is compatible with node server 1.5.x
Sorry.
I need to close it without merging. This functionality is not supported since Appium 1.5.0
https://github.com/appium/appium/blob/master/CHANGELOG.md#general-3
Server is throwing this exception now.
Method has not yet been implemented (WARNING: The server did not provide any stacktrace information)
See also:
https://github.com/appium/appium/issues/2234
Can we add docs for the new way of performing pinch and zoom?
@Jonahss It has to work the same way.
I and @ShrivanTarget are searching for something instead.
|
2025-04-01T06:37:55.913634
| 2024-11-02T12:21:39
|
2630428568
|
{
"authors": [
"atacan",
"dimitribouniol"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3751",
"repo": "apple/swift-algorithms",
"url": "https://github.com/apple/swift-algorithms/issues/245"
}
|
gharchive/issue
|
Warning for Documentation.docc: found 1 file(s) which are unhandled; explicitly declare them as resources or exclude from the target
Building a swift package that has swift-algorithms as dependency outputs warning in the terminal
Swift Algorithms version: 1.2.0
Swift version:
swift-driver version: 1.115 Apple Swift version 6.0 (swiftlang-<IP_ADDRESS>.10 clang-16<IP_ADDRESS>)
Target: arm64-apple-macosx15.0
Checklist
[ ] If possible, I've reproduced the issue using the main branch of this package
[x] I've searched for existing GitHub issues
I see this issue is closed https://github.com/apple/swift-algorithms/pull/237 but I don't see the exclude argument in the main branch as suggested in that issue.
Steps to Reproduce
Build the Hummingbird template project https://github.com/hummingbird-project/template
Expected behavior
No warning for package documentation
Actual behavior
warning: 'swift-algorithms': found 1 file(s) which are unhandled; explicitly declare them as resources or exclude from the target
/Users/atacan/Developer/Repositories/speech-router-hb/.build/checkouts/swift-algorithms/Sources/Algorithms/Documentation.docc
Just chiming in to say I also get this on everyone of my builds and was about to file an issue 😛
|
2025-04-01T06:37:55.916823
| 2022-09-30T00:59:07
|
1391672037
|
{
"authors": [
"lorentey"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3752",
"repo": "apple/swift-collections",
"url": "https://github.com/apple/swift-collections/pull/189"
}
|
gharchive/pull-request
|
[OrderedSet] Tiny doc fixes for isSuperset.
OrderedSet's documentation mistakenly refers to the subset relation when explaining what isSuperset is doing.
Checklist
[X] I've read the Contribution Guidelines
[X] My contributions are licensed under the Swift license.
[X] I've followed the coding style of the rest of the project.
[ ] I've added tests covering all new code paths my change adds to the project (if appropriate).
[ ] I've added benchmarks covering new functionality (if appropriate).
[X] I've verified that my change does not break any existing tests or introduce unexplained benchmark regressions.
[X] I've updated the documentation if necessary.
@swift-ci test
|
2025-04-01T06:37:55.917924
| 2019-10-15T21:48:08
|
507500430
|
{
"authors": [
"gmittert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3753",
"repo": "apple/swift-corelibs-foundation",
"url": "https://github.com/apple/swift-corelibs-foundation/pull/2538"
}
|
gharchive/pull-request
|
[Windows] Search for Dlls in Bundles Properly
On Windows, dlls don't begin with "lib" so the tests should be searching
for "MyDLL.dll", not "libMyDLL.dll"
@swift-ci please test and merge
|
2025-04-01T06:37:55.920865
| 2019-12-12T04:35:29
|
536751915
|
{
"authors": [
"YOCKOW",
"millenomi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3754",
"repo": "apple/swift-corelibs-foundation",
"url": "https://github.com/apple/swift-corelibs-foundation/pull/2585"
}
|
gharchive/pull-request
|
SR-11926: [FileHandle] Ensure the Darwin compatibility of Access Control of new methods.
From Swift 5.0, new APIs of FileHandle are introduced.
While they are defined as open in DarwinFoundation (as you know they are imported from objc), they are public in SwiftFoundation.
They should be defined as open also in SwiftFoundation to ensure compatibility (for example, making subclasses of FileHandle in other modules).
Resolves SR-11926.
Note that at least some of these changes are incorrect. e.g.: write(contentsOf:) is public, not open, on Darwin.
It seems to come to a conclusion.
Unfortunately, I'm going to close this PR.
|
2025-04-01T06:37:55.922384
| 2023-12-01T21:30:50
|
2021609510
|
{
"authors": [
"cachemeifyoucan"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3755",
"repo": "apple/swift-driver",
"url": "https://github.com/apple/swift-driver/pull/1498"
}
|
gharchive/pull-request
|
[Caching] Fix incremental cache build with bridging header
When planning a caching build with bridging header, the consumer of the pre-compiled bridging header needs to know the cache key for the bridging header in order to construct the correct build command, which requires visibility to the job that creates PCH.
Fix the incremental planning for bridging header users so those job can be correctly constructed even the bridging header job is skipped for the incremental build.
rdar://118143215
@swift-ci please test
|
2025-04-01T06:37:55.923715
| 2018-10-14T21:03:57
|
369943413
|
{
"authors": [
"Kaiede",
"dcci"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3756",
"repo": "apple/swift-lldb",
"url": "https://github.com/apple/swift-lldb/pull/971"
}
|
gharchive/pull-request
|
Improve libswiftReflection Linking on Linux
Using CMAKE_SYSTEM_PROCESSOR is a good first step, but for platforms where it maps to something different (say, armv7l to armv7), then you need more complete logic. Ideally, we would want Swift itself to tell us what to link against, but this change is still an improvement, although it is blatantly reusing similar logic from Swift's cmake scripts.
This is definitely better.
@swift-ci test and merge
|
2025-04-01T06:37:55.925843
| 2020-03-04T14:06:14
|
575428250
|
{
"authors": [
"weissi"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3757",
"repo": "apple/swift-nio",
"url": "https://github.com/apple/swift-nio/pull/1433"
}
|
gharchive/pull-request
|
fix macOS build & cleanup Thread
Motivation:
The macOS build was broken by #1424, also the Thread.swift file became
a bit of a mess with #if os(...)/#else everywhere.
Modifications:
fix macOS build
cleanup Thread by splitting it into ThreadPosix & ThreadWindows
Result:
build working again
cleaner code
@swift-nio-bot test this please
|
2025-04-01T06:37:55.927768
| 2019-03-06T16:41:22
|
417902005
|
{
"authors": [
"weissi"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3758",
"repo": "apple/swift-nio",
"url": "https://github.com/apple/swift-nio/pull/878"
}
|
gharchive/pull-request
|
ByteBuffer.viewBytes now returns an Optional
Motivation:
ByteBuffer.viewBytes would unconditionally return a ByteBufferView even
if the requested bytes were not readable.
Modifications:
make it return an Optional<ByteBufferView>
Result:
safer, better APIs
fixed #877
@swift-nio-bot test this please
|
2025-04-01T06:37:55.930976
| 2023-06-21T19:28:56
|
1768335992
|
{
"authors": [
"czechboy0"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3759",
"repo": "apple/swift-openapi-generator",
"url": "https://github.com/apple/swift-openapi-generator/issues/86"
}
|
gharchive/issue
|
Emit a descriptive error when running the generator on iOS, tvOS, or watchOS
The _OpenAPIGeneratorCore library and swift-openapi-generator executable are mostly an implementation detail of the OpenAPIGenerator plugin, all vended by the swift-openapi-generator package.
The generator runs at build time to generate code, but the _OpenAPIGeneratorCore library shouldn't actually be linked by adopters's target, which is documented correctly here (Step 7 of https://swiftpackageindex.com/apple/swift-openapi-generator/0.1.3/tutorials/swift-openapi-generator/clientxcode#Configuring-your-target-to-use-the-Swift-OpenAPI-Generator-plugin), but it's also easy to get it wrong and link it when adding the package in Xcode.
We should make it so that a clear, descriptive error is emitted when someone links the generator module itself to a non-macOS/Linux target.
One example question on the developer forums: https://developer.apple.com/forums/thread/732178
|
2025-04-01T06:37:55.939878
| 2023-12-10T01:57:09
|
2034184238
|
{
"authors": [
"MaxDesiatov",
"lynchsft",
"tishin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3760",
"repo": "apple/swift-package-manager",
"url": "https://github.com/apple/swift-package-manager/issues/7174"
}
|
gharchive/issue
|
Tests Fail to link on Windows when macros are present, jeopardizing Swift's cross-platform support.
Description
Tests fail to link on Windows using swift 5.9.1-RELEASE when macros are present.
This is a trivial repo that reproduces the issue:
https://github.com/lynchsft/swift-windows-macro-testing
The locus of this problem is on Windows platform, but I conclude that it indirectly negatively affects the entire Swift community. As macros proliferate through common libraries, all swift-on-windows tools will either have to abandon testing or wholly abandon windows as a platform. Neither of these outcomes speaks well of the "swift is a cross platform language" story.
Particularly, I'm asserting that this is a rapidly worsening problem because library authors that don't usually concern themselves with Windows are eagerly adopting macros feature. Soon, the libraries that Windows projects currently rely on will be un-ingestiable, effectively locking our dependences at the last-macro-free version.
This in turn, is a security problem as code-currency (up-to-date dependencies) are a crucial topic for averting and responding to security threats.
@neonichu, @MaxDesiatov During construction of that ^^ trivial reproduction repo, I actually encountered a success case, where I WAS able to ingest a macro and use it on Windows. The linker error did not arise.
Including this state in case its a useful resource to have.
Branch: success_case
Repo: https://github.com/lynchsft/swift-windows-macro-testing
Expected behavior
The tests should link and run properly.
Actual behavior
The tests fail to link (compilation succeeds):
Building for debugging...
lld-link: error: duplicate symbol: main
>>> defined at C:\Users\lynchsft\swift-windows-macro-testing\.build\x86_64-unknown-windows-msvc\debug\PowerAssertPlugin.build\PowerAssertPlugin.swift.o
>>> defined at C:\Users\lynchsft\swift-windows-macro-testing\.build\x86_64-unknown-windows-msvc\debug\SwiftWindowsPackageTestingPackageTests.build\runner.swift.o
clang: error: linker command failed with exit code 1 (use -v to see invocation)
[238/239] Linking C:\Users\lynchsft\swift-windows-macro-testing\.build\x86_64-unknown-windows-msvc\debug\SwiftWindowsPackageTestingPackageTests.xctest
error: fatalError
Steps to reproduce
swift --test
Swift Package Manager version/commit hash
5.9.1-RELEASE
Swift & OS version (output of swift --version ; uname -a)
Swift version 5.9.1 (swift-5.9.1-RELEASE)
Target: x86_64-unknown-windows-msvc
MSYS_NT-10.0-22631 lynchal-z01 3.4.6.x86_64 2023-05-18 20:39 UTC x86_64 Msys
Is this reproducible for you with latest development snapshots? (at the moment of writing it's https://download.swift.org/development/windows10/swift-DEVELOPMENT-SNAPSHOT-2023-11-27-a/swift-DEVELOPMENT-SNAPSHOT-2023-11-27-a-windows10.exe)
It is reproducible with the latest snapshot:
> swift --version
Swift version 5.11-dev (LLVM d50917983d84235, Swift 3f06b5efd0ebcfb)
Target: x86_64-unknown-windows-msvc
I have a slightly different example though:
https://github.com/tishin/MacroTestIssue
It has two targets - an empty macro implementation target with a @main attribute (which I believe causes this issue)
import SwiftCompilerPlugin
import SwiftSyntaxMacros
@main
struct MyMacroPlugin: CompilerPlugin {
let providingMacros: [Macro.Type] = [
]
}
And a trivial test target:
import XCTest
final class MyMacroTests: XCTestCase {
func testMacro() throws {
XCTAssertTrue(true)
}
}
Running swift test -v using the latest snapshot (as well as 5.9.2 release) fails with the duplication error:
lld-link: error: duplicate symbol: main
>>> defined at C:\Projects\MacroTestIssue\.build\x86_64-unknown-windows-msvc\debug\MyMacroMacros.build\MyMacroMacro.swift.o
>>> defined at C:\Projects\MacroTestIssue\.build\x86_64-unknown-windows-msvc\debug\MyMacroPackageTests.build\runner.swift.o
Thank you @tishin .
@MaxDesiatov, @neonichu Are you able to provide a read on the situation? Have I concluded wrongly that as time goes on this bug will strand Windows projects in the past?
|
2025-04-01T06:37:55.941585
| 2018-08-24T23:28:13
|
353957402
|
{
"authors": [
"aciidb0mb3r"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3761",
"repo": "apple/swift-package-manager",
"url": "https://github.com/apple/swift-package-manager/pull/1760"
}
|
gharchive/pull-request
|
[Workspace] Use topological sort when creating DependencyManifests
rdar://problem/43700179 isResolutionRequired is non-deterministic when a graph contains both revision and version based dependency requirements
Need to write a test for this
@swift-ci smoke test
|
2025-04-01T06:37:55.943883
| 2016-03-07T00:51:41
|
138856224
|
{
"authors": [
"kostiakoval",
"modocache"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3762",
"repo": "apple/swift-package-manager",
"url": "https://github.com/apple/swift-package-manager/pull/178"
}
|
gharchive/pull-request
|
[bootstrap] Remove deprecated "--build-tests"
The "--build-tests" option is marked "deprecated" in the SwiftPM bootstrap script and is no longer used by the script. It is also no longer used by any systems calling the bootstrap script; its only usage in the Swift build script has been removed in https://github.com/apple/swift/pull/1547.
Remove the option to reduce confusion.
This is trivial and safe change so I'm going to merge it.
I've tested that ./Utilities/bootstrap test works as expected both on MacOS and Linux and we don't need "--build-tests"
Thanks, @kostiakoval! :bow:
|
2025-04-01T06:37:55.946453
| 2020-01-10T22:33:41
|
548323051
|
{
"authors": [
"aciidb0mb3r",
"hartbit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3763",
"repo": "apple/swift-package-manager",
"url": "https://github.com/apple/swift-package-manager/pull/2511"
}
|
gharchive/pull-request
|
Binary targets workspace implementation
This implements the download and unarchiving implementation of the binary targets proposal.
I have several diagnostics in the PR reported using the emit(error: "description of error") instead of emit(.errorName(arguments)). I thought the second form was preferred, but I also saw a few instances of the first form using in Workspace. @aciidb0mb3r what do you think?
@hartbit The string variant is ok for one-off cases where we don't need to emit the same error from somewhere else but emit(.errorName(arguments)) is probably better/preferable.
@swift-ci smoke test self hosted
@swift-ci smoke test
|
2025-04-01T06:37:55.949352
| 2023-05-22T23:16:23
|
1720788610
|
{
"authors": [
"TTOzzi",
"ahoppen",
"kimdv"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3764",
"repo": "apple/swift-syntax",
"url": "https://github.com/apple/swift-syntax/issues/1691"
}
|
gharchive/issue
|
Parser should emit a single diagnostic if both open and close quote are missing from string literal
Issue Kind
Bad Diagnostic Produced
Source Code
@_expose(Cxx, baz) func foo() {}
Description
From https://github.com/apple/swift-syntax/pull/1688/files#r1201105564
In the above example the parser produces two diagnostics: One for the missing opening " before baz and one for the missing closing " after baz. It should emit a single diagnostic instead that offers to add both.
Tracked in Apple’s issue tracker as rdar://109685043
Go for it @TTOzzi 🙌
Thank you! 😉
|
2025-04-01T06:37:55.954955
| 2024-07-31T11:49:13
|
2439854085
|
{
"authors": [
"grynspan",
"kateinoigakukun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3765",
"repo": "apple/swift-testing",
"url": "https://github.com/apple/swift-testing/pull/584"
}
|
gharchive/pull-request
|
Complete WebAssembly/WASI Support
Motivation:
Trying to support WebAssembly/WASI before shipping swift-testing as a part of toolchain
Modifications:
Thanks to prior WASI compatibility efforts by @grynspan, most of the parts work as is now. Here are a few items I changed:
Remove SWT_NO_FILE_IO definition since wasi-libc provides all IO functionalities we need
Stop including signal.h
Fix metadata accessor function pointer representation on WASI to use absolute pointer instead of relative pointer
Define primitive lock type as Void for WASI
It's not the optimal representation but I couldn't find a better way 😓
Result:
Tested with DEVELOPMENT-SNAPSHOT-2024-07-15-a toolchain
$ swift sdk install https://github.com/swiftwasm/swift/releases/download/swift-wasm-DEVELOPMENT-SNAPSHOT-2024-07-16-a/swift-wasm-DEVELOPMENT-SNAPSHOT-2024-07-16-a-wasm32-unknown-wasi.artifactbundle.zip
$ swift package init --name Hello --enable-experimental-swift-testing
# Replace swit-testing dependency entry in Package.swift with:
# .package(url: "https://github.com/kateinoigakukun/swift-testing.git", branch: "yt/wasi-support"),
$ swift build --swift-sdk DEVELOPMENT-SNAPSHOT-2024-07-16-a-wasm32-unknown-wasi --build-tests --enable-experimental-swift-testing
$ wasmkit-cli run --dir . ./.build/debug/HelloPackageTests.wasm
◇ Test run started.
↳ Testing Library Version: 3702b3a53779ccca0f8205b7a76a6f5038197120
◇ Test example() started.
✔ Test example() passed after 0.041244319 seconds.
✔ Test run with 1 test passed after 0.081816901 seconds.
Checklist:
[x] Code and documentation should follow the style of the Style Guide.
[x] If public symbols are renamed or modified, DocC references should be updated.
@swift-ci test
Thanks for taking a look at all this! We really appreciate it. :)
@swift-ci test
One other thing! Can you modify the project readme to include this line in the supported OS table? 😁
| **Wasm** | | | Experimental |
@swift-ci please test
Ready to merge when you are.
⛵
|
2025-04-01T06:37:55.961958
| 2017-09-12T20:55:57
|
257179226
|
{
"authors": [
"Rostepher",
"davezarzycki",
"tishansteve"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3766",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/11882"
}
|
gharchive/pull-request
|
Move build-script apply default arguments into separate module
Purpose
This PR moves the apply_default_arguments logic out of the main build-script code into the driver_arguments module in order to better stage for #11872 and eventually #11880. The diff will show that the move was not strictly a copy-paste, the chunk for determining if ninja ought be built requires the toolchain information which only exists in the BuildScriptInvocation class. I've opted instead to single out that single case and keep it where it was, everything else was moved verbatim.
rdar://problem/34336890
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
As soon as the builds for this branch pass I'll merge it, then onwards to #11872!
This broke the build-script on my machine. It can no longer find ninja in /usr/local/bin. What do you need to help debug this?
@davezarzycki I've made a PR to address the issue #11919. I'll have it merged tonight or in the morning, Sorry for the inconvenience.
Can you unsubscribe me from this email service cause I don’t want any
emails from git hub
On Thu, Sep 14, 2017 at 11:32 AM Ross Bayer<EMAIL_ADDRESS>wrote:
@davezarzycki https://github.com/davezarzycki I've made a PR to address
the issue #11919 https://github.com/apple/swift/pull/11919. I'll have
it merged tonight or in the morning, Sorry for the inconvenience.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/apple/swift/pull/11882#issuecomment-329382112, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AQ86b_vQseMcea4728BLdWX66v9Jcs63ks5siMF0gaJpZM4PVPK6
.
@tishansteve You can change your email preferences in your Github account settings. I can't do anything about it.
|
2025-04-01T06:37:55.986419
| 2018-01-20T13:15:13
|
290191466
|
{
"authors": [
"CodaFi",
"futurejones",
"hartbit",
"itaiferber",
"mortenbekditlevsen",
"norio-nomura",
"parkera",
"shahmishal",
"theblixguy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3767",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/14039"
}
|
gharchive/pull-request
|
Add useSnakeCasedKeys to JSONDecoder.KeyDecodingStrategy
Adds JSONDecoder.KeyDecodingStrategy.useSnakeCasedKeys.
This will make following code works.
import Foundation
struct S: Codable {
var myURLProperty: String
}
let encoder = JSONEncoder()
encoder.keyEncodingStrategy = .convertToSnakeCase
let data = try encoder.encode(S(myURLProperty: "property"))
let decoder = JSONDecoder()
decoder.keyDecodingStrategy = .useSnakeCasedKeys
let decoded = try decoder.decode(S.self, from: data)
/cc: @itaiferber
Resolves SR-6629.
@itaiferber For what's its worth, I tried using convertToSnakeCase/convertFromSnakeCase for the first time today and was really surprised that the behaviour was not what @norio-nomura is proposing. Converting the CodingKeys to snake_case instead of converting the JSON keys to camelCase seems like the only solution to avoid all those round-trip errors.
@hartbit @norio-nomura I forgot to update this PR (and the SR) with some additional discussion and notes for thought. Even if we decide to special-case convertToSnakeCase and convertFromSnakeCase here and perform the conversions in the "favorable" direction, there are still other cases which can't necessarily be fixed.
Consider KeyedDecodingContainer<Key>.allKeys, which maps the contents of what is in the payload to a list of Keys. This is done by iterating the keys in the payload and creating a Key with the value of every found key via CodingKey.init?(stringValue:).
Right now, we do this by having two conversions — Key-to-String (Key.stringValue → case conversion) and String-to-Key (case conversion → Key.init?(stringValue:)). If we "fix" the issue by only using Key-to-String conversions as suggested, there's no real way to map from something already in the payload to a Key. One way to do this could be to require all CodingKeys to be CaseIterable and on every request for allKeys, take Key, iterate every case, converts every key to a string, applies the case conversion, and then iterate all of the keys in the payload and try to look them up in the converted table. This is both expensive, and not necessarily possible as not all CodingKeys can be CaseIterable (think _DictionaryCodingKey and other dynamic keys).
We could decide to "fix" this but still leave allKeys as-is, but this is still bad because you can't rely on iterating through allKeys to really know what's in the payload.
And at that, this is a special case for convertToSnakeCase/convertFromSnakeCase; we could not apply this to strategies which are even more lossy (e.g. .custom, for which we really only have one direction of conversion). There's a balance here between doing the "right" (i.e. magic) thing for some API consumers and being consistent.
One thing to keep in mind, I think, is that these conversions are by nature lossy. There are always going to be edge cases where we've lost too much information to round-trip the keys. The question is where is the balance between preserving more cases and being consistent across the board. I currently favor being consistent here.
Seems like a loose-loose situation where there is not great solution.
One way to do this could be to require all CodingKeys to be CaseIterable and on every request for allKeys, take Key, iterate every case, converts every key to a string, applies the case conversion, and then iterate all of the keys in the payload and try to look them up in the converted table.
Couldn't KeyedDecodingContainer cache the converted keys?
How about special casing the snake_case to camelCase conversion to recognise common abbreviations often found in Swift code and uppercasing them as the naming conventions suggests? The obvious that come to mind are: HTTP, URL, UUID, XML, JSON
Yes, caching is possible, though unfortunately this doesn't solve the problem for keys which are not iterable.
Recognizing special words is certainly an option (and can be done transparently without updated API); the question is — what abbreviations are important enough to make it into that list? 🙂 All the ones you listed, I agree with. If we're doing HTTP, should we do HTTPS? (And what about other schemes?) It gets a bit hairier... What if people have other abbreviations they want that are not on our list? One option is to make an extensible list of known initialisms that anyone can extend in their process and just do that.
(I'm not trying to say that just because there isn't a "best" solution we shouldn't do anything; it's just about trying to figure out the best way going forward.)
Another solution is simply exposing JSONEncoder.KeyEncodingStrategy.toSnakeCase() and JSONDecoder.KeyDecodingStrategy.fromSnakeCase() and letting folks more easily do what they need to in .custom conversions.
I noticed another problem as a consequence of the current behaviour: resolving the round-trip issue in @norio-nomura's example requires giving the CondingKey a rawValue of myUrlProperty. I think this is very unintuitive and not easily discoverable as no domain (wether the Swift domain or JSON domain) ever declares a key with the value myUrlProperty: it feels outlandish to both domains.
The more I think about it the more I think that @norio-nomura's useSnakeCasedKeys has to be part of the solution, whatever the cost is of resolving KeyedDecodingContainer. By the way, isn't KeyedDecodingContainer.keys invisible to users to JSONEncoder and JSONDecoder? Can't we just return an empty array?
No, we can't do that, unfortunately. JSONDecoder.container(keyedBy:) returns a KeyedDecodingContainer, which exposes allKeys. That value has to be consumable for types which iterate through the possible keys in the payload in order to decide how to decode; we can't simply return an empty array for the benefit of making these edge cases work.
Perhaps we should take this to the Swift Forums for discussion on possible ways forward. At this point, we're not going to be able to introduce new API into Swift 4.2, but maybe there are other directions we can take this.
I started a thread on the Swift forums to discuss this in a more public setting. :)
Thx @itaiferber! Would it worth moving the discussion to Swift Evolution?
I think it would be good to gather feedback first about potential directions before suggesting API to Swift Evolution, but eventually, yes
Was any conclusion on the issue reached through the discussion?
There were three responses to the thread, all supportive of this approach; if we consider that to be enough signal to go ahead with this, we'll need to figure out how we want to review and schedule this API change. Given the amount of things going on at the moment, I'm not sure this will make the 4.2 release.
@parkera Any thoughts here?
Let's aim for 5.0. master is still the branch for that, but since we need time to do additional API review here, we will leave this open for a bit longer.
Hi there. I hope that it is ok that I bump this thread.
@parkera you mention that you hope to aim for 5.0. Now that we are nearing 5.0 do you think that there is any hope that this could make it?
Any updates on this? Could we merge this and cherry pick into 5.1 perhaps? Seems like there's a lot of support for this so it would be a shame if this gets abandoned.
Any changes made here would need to go through API review, which won't make it into Swift 5.1, unfortunately. We will try to push on this internally, given bandwidth.
Okay! I hope it makes into Swift 5.2 then :)
FYI, I opened https://github.com/jpsim/Yams/pull/200 as PoC that improved from this PR.
It's been a while, so let's get this moving again. Because this change is API and ABI breaking (not that I anticipate there are many clients switching over this enum), it requires a formal proposal to Swift evolution. During that process, it's possible you may hear from the core team that this change is OK to take without a formal review. Luckily, you've already done the heavy lifting by providing an implementation.
I'm going to tag this as pending evolution discussion.
@codafi Since JSONDecoder does not live in the standard library, this should not require full Swift evolution, but does require Foundation review. /CC @parkera to weigh in on changes here
I have checked, and I stand corrected. The overlays are Foundation’s and are not subject to evolution. Hopefully @parkera can get this reviewed in time.
cc @bendjones on this one
Can one of the admins verify this patch?
Please update the base branch to main by Oct 5th otherwise the pull request will be closed automatically.
How to change the base branch: (Link)
More detail about the branch update: (Link)
Hi there,
Apologies for the bump, but unfortunately this PR was automatically closed - and I don't think it got the attention that it deserved.
The comments do mention internal reviews, so I am hoping that perhaps the issue is not yet off of the radar.
Otherwise: might I reopen a copy of the PR against main?
I am certain that the issue is very well understood, but I think that the current state causes a lot of confusion - even for experienced developers. The intention was to keep the key conversion in the encoder and decoder, but the current state is that this often 'leaks' into the CodingKey definition where you have to 'translate' something like imageURL to imageUrl for it to roundtrip to snake case correctly. I've seen quite a few bugs due to the misunderstanding that it would help to translate the key directly into snake case as:
case imageURL = "image_url"
As mentioned elsewhere, the big issue is that the 'imageUrl' representation belongs neither to the source or the destination domains but must be added for the sake of the key coding strategy.
@parkera , @CodaFi , @bendjones
|
2025-04-01T06:37:55.989370
| 2018-05-23T15:51:36
|
325769018
|
{
"authors": [
"compnerd"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3768",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/16795"
}
|
gharchive/pull-request
|
IRGen: adjust for LLVM API change
SVN r332881 introduced an additional parameter to emit a DWO file.
Replace this paragraph with a description of your changes and rationale. Provide links to external references/discussions if appropriate.
Resolves SR-NNNN.
CC: @bob-wilson
|
2025-04-01T06:37:55.991891
| 2018-08-18T17:38:44
|
351838084
|
{
"authors": [
"eeckstein",
"gottesmm"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3769",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/18822"
}
|
gharchive/pull-request
|
[passmanager] Add the ability to verify before/after/around a specifi…
…c transform.
This helps speed up triaging failures caught by -sil-verify-all since this
allows one to trigger the -sil-verify-all verification around specific passes in
the pipeline rather than after every pass run. Was useful for me when tracking
down missing pass manager notification.
@swift-ci smoke test
Actually, I am going to centralize this logic a little bit before the review. Erik, I'll ping you again when I am ready.
I'm wondering what's the motivation for this. Is -sil-verify-all really too slow for debugging?
@eeckstein I found that the PassManagerNotificationVerificationAnalysis was too expensive to run with sil-verify-all since it has to run over all functions. I found an alternative solution (namely, verifyFull), but this seemed like generally useful functionality.
|
2025-04-01T06:37:55.994430
| 2018-08-20T23:24:13
|
352330823
|
{
"authors": [
"rjmccall",
"swift-ci"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3770",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/18858"
}
|
gharchive/pull-request
|
Fix some edge cases when inlining coroutines
The current inlining strategy doesn't support inlining coroutines when there are multiple end_apply or abort_apply instructions in the caller, so refuse to inline such cases. Also, handle the case where there are no yield instructions in the callee, which can happen if e.g. the callee calls a no-return function.
I also simplified the code somewhat by removing the vestiges of the code that tried to unify control flow with switches.
As an unrelated fix, suppress function signature optimization for coroutines for now.
Build failed
Swift Test Linux Platform
Git Sha -<PHONE_NUMBER>ed929b51267d0473aa8ab488cfea69
@swift-ci Please clean test Linux.
|
2025-04-01T06:37:56.359823
| 2016-03-30T08:55:49
|
144509029
|
{
"authors": [
"benlangmuir",
"gparker42",
"gribozavr",
"jckarter",
"nadavrot",
"shahmishal",
"shawnce",
"swift-ci",
"tinysun212"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3771",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/1950"
}
|
gharchive/pull-request
|
[runtime] enhanced and refactored recently added Mutex abstraction
added read / write lock support
added non-fatal error support to allow use of mutex in fatal error reporting pathway
isolated pthread implementation to it own header/cpp file pair
expanded unit tests to cover new code as well as better test existing mutex
removed a layer of complexity that added no real value
added static allocation friendly versions of Mutex, ConditionVariable, and ReadWriteLock
Resolved bug number: (SR-946)
Before merging this pull request to apple/swift repository:
[ ] Test pull request on Swift continuous integration.
Triggering Swift CI
The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are:
Smoke Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please smoke test
OS X platform | @swift-ci Please smoke test OS X platform
Linux platform | @swift-ci Please smoke test Linux platform
Validation Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please test
OS X platform | @swift-ci Please test OS X platform
Linux platform | @swift-ci Please test Linux platform
Note: Only members of the Apple organization can trigger swift-ci.
Looking at conflict now.
Not a review, just triggering the CI.
@swift-ci Please test
Offline until morning (PST).
Looks like both Linux and Mac passed thanks for the early CI run.
Think I dealt with most comments with an open question or two for folks. If possible getting a CI and benchmark run would be helpful.
@swift-ci Please test
@swift-ci Please benchmark
@swift-ci Please test
!!! Couldn't read commit file !!!
In addition to CI, some should verify that this doesn't introduce any additional -Wglobal-constructors or -destructors warnings.
Looks like a lingering use of Condition instead of ConditionVariable not sure why my local benchmark run didn't flag it.
This is not a review of anything in this patch, but when writing tests for your synchronization primitives please bear in mind that 1) sleeping for XX ms is not sufficient to always have all your threads started and 2) gtest ASSERT_* macros won't force the test to stop running if they happen inside a thread, since they work by using 'return'.
I just fixed a race in the previous tests with fa77a7a.
Yeah that test has been reworked in this patch for that very reason. I had made a comment on the prior PR about that issue after it was reported to be seen on Linux.
I'm looking at Mutex::lockOrWait(). Its body is equivalent to below code, is it right?
template <typename CriticalSection>
void lockOrWait(ConditionVariable &condition,
CriticalSection criticalSection) {
lock([&] {
while (!criticalSection()) {
wait(condition);
}
});
}
lock();
while (!criticalSection()) {
ConditionPlatformHelper::wait(condition.Handle, Handle);
}
unlock();
According to their manuals, all three condition variable's wait function (pthread, <condition_variable>, Win32) release the associated lock. I think the second while loop will be running in unlocked.
Upon successful return, the mutex shall have been locked and shall be owned by the calling thread.
I believe the above is universal for condition wait behavior (quoted from http://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_cond_timedwait.html). You must hold the lock before calling wait, then before waiting (blocking calling thread) the wait function drops the lock, and before wait returns (assuming no error / timeout) it must reacquire the lock.
So while looping the lock is held except when inside of the wait call.
Oh I understand. I was wrong about pthread, <condition_variable>, but Win32 SleepConditionVariableSRW() may differ.
I see the following community addition to the docs about SleepConditionVariableSRW(). I would be very surprised if it didn't acquire the lock before returning.
Even after the wait time expires, this function does not return until it acquires the lock again no matter how long it takes. This thread must always release the lock regardless of the return value. The return value indicates that the conditional variable was set (true) or wait timed out (false + ERROR_TIMEOUT).
Yes, I confirmed with test code. It locks again !. MSDN document was poor. Thanks for reference.
I wrote MutexWin32.cpp(.h) and they passed your test code unittests/runtime/Mutex.cpp. It was very simple. Thanks you.
I think this is finally good to go. It will need a CI & benchmark before any attempt to merge.
I have held off removing ReadWriteLock::readWriteLock([]{}, []{}) for the moment since unlock got separated into readUnlock and writeUnlock in support of Win32 version of this code. The abstraction helps avoid mistakes with calling the wrong unlock (granted likely easily detected in testing) and of course avoids duplication of code in typical fast path / slow path type usage of read write locks. I will remove it folks want.
@swift-ci Please test
@swift-ci Please benchmark
@swift-ci Please test
!!! Couldn't read commit file !!!
@shawnce There's some linker heartburn:
Undefined symbols for architecture x86_64:
"swift::Condition::notifyAll()", referenced from:
__swift_allocBox_ in libswiftRuntime.a(HeapObject.cpp.o)
_swift_getGenericMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getObjCClassMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getFunctionTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getTupleTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getExistentialMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
...
"swift::Condition::Condition()", referenced from:
swift::Lazy<swift::MetadataCache<(anonymous namespace)::BoxCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(HeapObject.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::ObjCClassCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::FunctionCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::TupleCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::MetatypeCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::GenericCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::WitnessTableCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
...
"swift::MutexImpl::lock()", referenced from:
__swift_allocBox_ in libswiftRuntime.a(HeapObject.cpp.o)
_swift_getGenericMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getObjCClassMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getFunctionTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getTupleTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getExistentialMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
...
"swift::MutexImpl::wait(swift::Condition&)", referenced from:
__swift_allocBox_ in libswiftRuntime.a(HeapObject.cpp.o)
_swift_getGenericMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getObjCClassMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getFunctionTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getTupleTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getExistentialMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
...
"swift::MutexImpl::unlock()", referenced from:
__swift_allocBox_ in libswiftRuntime.a(HeapObject.cpp.o)
_swift_getGenericMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getObjCClassMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getFunctionTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getTupleTypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
_swift_getExistentialMetatypeMetadata in libswiftRuntime.a(Metadata.cpp.o)
...
"swift::MutexImpl::MutexImpl(bool)", referenced from:
swift::Lazy<swift::MetadataCache<(anonymous namespace)::BoxCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(HeapObject.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::ObjCClassCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::FunctionCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::TupleCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::MetatypeCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<(anonymous namespace)::ForeignTypeState>::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
swift::Lazy<swift::MetadataCache<(anonymous namespace)::GenericCacheEntry> >::defaultInitCallback(void*) in libswiftRuntime.a(Metadata.cpp.o)
...
ld: symbol(s) not found for architecture x86_64
@jckarter I spotted that I can't explain why that is happening. Can a plain CI run be done to confirm what I am seeing on my end (I don't see any linker issues, etc.). I will rerun benchmarks locally to confirm I see no issues (didn't see any last time).
@swift-ci Please test
@jckarter the issue looks specific to what the benchmark operation is doing. I can build --benchmark locally without issues. I suspect somehow a reference branch is being made from some invalid merge point that is resulting in a mix of old and new code trying to be compiled.
Any place I can look to see how a given CI build is setup, to understand the command executed, etc.?
@shahmishal Any idea why the benchmark runner would fail with linker errors when the normal CI works?
I looked the failed benchmark consol output https://ci.swift.org/job/swift-PR-osx-perf/35/console. I think the problem is in the ci build process.
The commit a0452be removed stdlib/public/runtime/Mutex.cpp and added stdlib/public/runtime/MutexPThread.cpp, but the benchmark compiled both.
It seems it tested without MutexPThread.cpp, and older source set. There is no MutexImpl class in current commit, isn't it?
[505/949] Building CXX object stdlib/public/runtime/CMakeFiles/swiftRuntime-macosx-x86_64.dir/Mutex.cpp.o
.....
[30/167] Building CXX object stdlib/public/runtime/CMakeFiles/swiftRuntime-macosx-x86_64.dir/MutexPThread.cpp.o
.....
"swift::MutexImpl::lock()", referenced from:
@shahmishal @jckarter Do the build commands exist in git someplace for what the benchmark CI plan does? ..or does a lot of that live in Jenkins? Would a new pull request potentially side step some state that may have a hand in an incorrect merge being used?
@swift-ci Please benchmark
@shawnce Benchmark bot:
Checkout master branch (Targeted Branch)
run swift/utils/build-script -R --no-assertions --benchmark
Checkout origin/pr/1950/merge
run swift/utils/build-script -R --no-assertions --benchmark
Upload the results to GitHub pull request.
Build comment file:
Optimized (O)
Regression (1)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ClassArrayGetter
12
13
+8.3%
0.92x
Improvement (1)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeFromNSArrayAnyObjectToString
166404
155969
-6.3%
1.07x(?)
No Changes (113)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeToNSSet
16349
15737
-3.7%
1.04x(?)
DictionaryLiteral
3418
3277
-4.1%
1.04x
ObjectiveCBridgeFromNSDictionaryAnyObjectToString
190926
185476
-2.9%
1.03x
ObjectiveCBridgeStubFromArrayOfNSString
113176
109936
-2.9%
1.03x(?)
Array2D
2024
1981
-2.1%
1.02x(?)
StringInterpolation
10080
9920
-1.6%
1.02x
Walsh
311
306
-1.6%
1.02x
ObjectiveCBridgeToNSArray
29678
29149
-1.8%
1.02x(?)
ObjectiveCBridgeToNSDictionary
16343
16057
-1.8%
1.02x(?)
DictionaryRemove
4799
4718
-1.7%
1.02x(?)
NSStringConversion
616
606
-1.6%
1.02x
Chars
670
660
-1.5%
1.02x(?)
ArraySubscript
1338
1330
-0.6%
1.01x
ObjectiveCBridgeFromNSArrayAnyObjectForced
3654
3617
-1.0%
1.01x
ErrorHandling
2866
2829
-1.3%
1.01x(?)
ArrayLiteral
993
987
-0.6%
1.01x(?)
ProtocolDispatch2
163
162
-0.6%
1.01x
DictionarySwapOfObjects
7176
7116
-0.8%
1.01x(?)
StrComplexWalk
2808
2783
-0.9%
1.01x(?)
ArrayOfRef
3454
3434
-0.6%
1.01x(?)
ObjectiveCBridgeFromNSArrayAnyObject
63195
62477
-1.1%
1.01x(?)
NSDictionaryCastToSwift
8554
8436
-1.4%
1.01x(?)
HashTest
1807
1791
-0.9%
1.01x(?)
LinkedList
8997
8931
-0.7%
1.01x(?)
MonteCarloPi
42564
42587
+0.1%
1.00x
StackPromo
18991
18990
-0.0%
1.00x(?)
RecursiveOwnedParameter
1829
1829
+0.0%
1.00x
ObjectiveCBridgeStubToNSString
1373
1375
+0.1%
1.00x(?)
Histogram
652
652
+0.0%
1.00x
SortStrings
1753
1747
-0.3%
1.00x
DictionaryBridge
3528
3530
+0.1%
1.00x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectForced
4550
4567
+0.4%
1.00x(?)
ArrayOfGenericPOD
207
208
+0.5%
1.00x
StringWithCString
81188
81245
+0.1%
1.00x(?)
ObjectiveCBridgeFromNSStringForced
2202
2194
-0.4%
1.00x(?)
Prims
729
730
+0.1%
1.00x(?)
SortLettersInPlace
1003
1001
-0.2%
1.00x
PopFrontArray
1228
1230
+0.2%
1.00x(?)
Dictionary3OfObjects
1184
1185
+0.1%
1.00x(?)
ByteSwap
1
1
+0.0%
1.00x
SuperChars
282895
283773
+0.3%
1.00x(?)
XorLoop
359
359
+0.0%
1.00x
ObjectiveCBridgeStubFromNSStringRef
327
327
+0.0%
1.00x
ProtocolDispatch
2879
2885
+0.2%
1.00x(?)
ObjectAllocation
141
141
+0.0%
1.00x
TypeFlood
0
0
+0.0%
1.00x
ObjectiveCBridgeFromNSSetAnyObject
93758
94213
+0.5%
1.00x(?)
Dictionary3
726
727
+0.1%
1.00x(?)
Dictionary2
2399
2400
+0.0%
1.00x(?)
SetIntersect_OfObjects
2728
2719
-0.3%
1.00x(?)
NSError
353
352
-0.3%
1.00x(?)
DictionaryOfObjects
2687
2693
+0.2%
1.00x(?)
PopFrontUnsafePointer
8506
8501
-0.1%
1.00x(?)
PolymorphicCalls
61
61
+0.0%
1.00x
114
2243449
2236732
-0.3%
1.00x
Hanoi
3431
3414
-0.5%
1.00x(?)
OpenClose
49
49
+0.0%
1.00x
DictionaryRemoveOfObjects
19410
19485
+0.4%
1.00x(?)
UTF8Decode
257
256
-0.4%
1.00x
SetIsSubsetOf
480
479
-0.2%
1.00x
NopDeinit
34609
34660
+0.1%
1.00x(?)
SetIntersect
1134
1133
-0.1%
1.00x(?)
RGBHistogramOfObjects
22346
22384
+0.2%
1.00x
StringBuilder
1369
1365
-0.3%
1.00x(?)
Phonebook
9288
9299
+0.1%
1.00x(?)
Sim2DArray
413
411
-0.5%
1.00x(?)
SetExclusiveOr_OfObjects
8855
8839
-0.2%
1.00x(?)
MonteCarloE
10048
10034
-0.1%
1.00x(?)
SetUnion_OfObjects
7013
7020
+0.1%
1.00x(?)
RC4
251
251
+0.0%
1.00x
SetIsSubsetOf_OfObjects
540
540
+0.0%
1.00x
ArrayAppend
730
728
-0.3%
1.00x(?)
RGBHistogram
3359
3352
-0.2%
1.00x(?)
ArrayValueProp4
5
5
+0.0%
1.00x
ArrayOfPOD
172
172
+0.0%
1.00x
SetUnion
3242
3234
-0.2%
1.00x
StrToInt
3622
3632
+0.3%
1.00x(?)
BitCount
1
1
+0.0%
1.00x
AngryPhonebook
3367
3358
-0.3%
1.00x(?)
SevenBoom
1267
1269
+0.2%
1.00x(?)
StringWalk
6438
6452
+0.2%
1.00x(?)
ArrayValueProp
5
5
+0.0%
1.00x
GlobalClass
0
0
+0.0%
1.00x
Memset
222
222
+0.0%
1.00x
ArrayAppendReserved
508
508
+0.0%
1.00x
ArrayValueProp2
5
5
+0.0%
1.00x
ArrayValueProp3
5
5
+0.0%
1.00x
ObjectiveCBridgeToNSString
1021
1028
+0.7%
0.99x(?)
ObjectiveCBridgeFromNSDictionaryAnyObject
184245
186150
+1.0%
0.99x(?)
DictionarySwap
757
762
+0.7%
0.99x(?)
Integrate
228
230
+0.9%
0.99x(?)
ObjectiveCBridgeStubFromNSString
697
701
+0.6%
0.99x(?)
Join
428
433
+1.2%
0.99x
PopFrontArrayGeneric
1255
1262
+0.6%
0.99x
CaptureProp
3752
3779
+0.7%
0.99x(?)
MapReduce
329
332
+0.9%
0.99x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToStringForced
122660
124131
+1.2%
0.99x(?)
Dictionary
959
973
+1.5%
0.99x
ArrayInClass
82
83
+1.2%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObjectToStringForced
106501
107077
+0.5%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObjectToString
169579
171590
+1.2%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObjectForced
4023
4070
+1.2%
0.99x(?)
DeadArray
176
177
+0.6%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
159423
160487
+0.7%
0.99x(?)
Dictionary2OfObjects
3969
4005
+0.9%
0.99x
TwoSum
1713
1727
+0.8%
0.99x
RangeAssignment
262
266
+1.5%
0.98x(?)
ObjectiveCBridgeStubToArrayOfNSString
29702
30248
+1.8%
0.98x(?)
SetExclusiveOr
3782
3844
+1.6%
0.98x
ArrayOfGenericRef
3475
3542
+1.9%
0.98x(?)
ObjectiveCBridgeFromNSString
1529
1553
+1.6%
0.98x(?)
StaticArray
2598
2672
+2.9%
0.97x(?)
Calculator
32
33
+3.1%
0.97x
**Unoptimized (Onone)**
Regression (1)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
TypeFlood
142
151
+6.3%
0.94x(?)
Improvement (0)
No Changes (114)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeFromNSSetAnyObjectToStringForced
121775
116408
-4.4%
1.05x(?)
OpenClose
451
433
-4.0%
1.04x(?)
DictionaryOfObjects
7282
7049
-3.2%
1.03x
StringWalk
42766
41327
-3.4%
1.03x(?)
StringInterpolation
13674
13419
-1.9%
1.02x(?)
StrComplexWalk
7871
7686
-2.4%
1.02x
PopFrontArrayGeneric
12442
12203
-1.9%
1.02x
CaptureProp
55130
54267
-1.6%
1.02x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToStringForced
127072
124314
-2.2%
1.02x
StackPromo
80945
80335
-0.8%
1.01x(?)
Histogram
13859
13768
-0.7%
1.01x(?)
PopFrontArray
12262
12107
-1.3%
1.01x(?)
ErrorHandling
7241
7181
-0.8%
1.01x(?)
ObjectiveCBridgeStubFromNSString
1138
1132
-0.5%
1.01x
ProtocolDispatch
5602
5568
-0.6%
1.01x(?)
MapReduce
45175
44618
-1.2%
1.01x(?)
Integrate
348
346
-0.6%
1.01x
UTF8Decode
39402
39202
-0.5%
1.01x(?)
Dictionary
4532
4500
-0.7%
1.01x
ObjectiveCBridgeStubToArrayOfNSString
30424
30172
-0.8%
1.01x(?)
StringBuilder
8268
8153
-1.4%
1.01x(?)
TwoSum
9273
9193
-0.9%
1.01x(?)
ArraySubscript
4450
4448
-0.0%
1.00x(?)
DictionarySwapOfObjects
25809
25767
-0.2%
1.00x(?)
ObjectiveCBridgeFromNSDictionaryAnyObject
192412
191679
-0.4%
1.00x(?)
RecursiveOwnedParameter
10283
10293
+0.1%
1.00x(?)
ObjectiveCBridgeStubToNSString
1672
1676
+0.2%
1.00x
ClassArrayGetter
1264
1264
+0.0%
1.00x
Array2D
697294
697454
+0.0%
1.00x(?)
SortStrings
5672
5652
-0.3%
1.00x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectForced
7074
7042
-0.5%
1.00x(?)
MonteCarloPi
50224
50226
+0.0%
1.00x(?)
StringWithCString
713012
712645
-0.1%
1.00x(?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
7029
7038
+0.1%
1.00x(?)
Prims
13669
13646
-0.2%
1.00x(?)
SortLettersInPlace
2319
2315
-0.2%
1.00x(?)
DictionarySwap
11593
11587
-0.1%
1.00x(?)
ByteSwap
9
9
+0.0%
1.00x
XorLoop
18434
18445
+0.1%
1.00x(?)
ObjectiveCBridgeStubFromNSStringRef
651
651
+0.0%
1.00x
StaticArray
17031
16984
-0.3%
1.00x(?)
ArrayLiteral
1033
1035
+0.2%
1.00x(?)
Walsh
12244
12243
-0.0%
1.00x(?)
Dictionary3
2099
2098
-0.1%
1.00x(?)
Dictionary2
5147
5143
-0.1%
1.00x(?)
SetIntersect_OfObjects
18893
18918
+0.1%
1.00x(?)
Join
1340
1334
-0.5%
1.00x
ArrayOfRef
9929
9908
-0.2%
1.00x(?)
PopFrontUnsafePointer
95350
95047
-0.3%
1.00x(?)
Calculator
932
931
-0.1%
1.00x
DictionaryRemoveOfObjects
60971
60879
-0.1%
1.00x(?)
SetIsSubsetOf
2952
2948
-0.1%
1.00x(?)
NopDeinit
53314
53273
-0.1%
1.00x(?)
SetIntersect
18837
18853
+0.1%
1.00x(?)
SetExclusiveOr
48566
48583
+0.0%
1.00x(?)
RGBHistogramOfObjects
97745
98173
+0.4%
1.00x(?)
ArrayInClass
3648
3647
-0.0%
1.00x
ArrayOfGenericRef
10569
10563
-0.1%
1.00x(?)
Phonebook
59729
59869
+0.2%
1.00x(?)
Sim2DArray
13143
13139
-0.0%
1.00x(?)
SetExclusiveOr_OfObjects
66912
66855
-0.1%
1.00x(?)
MonteCarloE
98538
98873
+0.3%
1.00x
SetUnion_OfObjects
51158
51146
-0.0%
1.00x(?)
GlobalClass
0
0
+0.0%
1.00x
HashTest
6016
6036
+0.3%
1.00x
SetIsSubsetOf_OfObjects
2835
2847
+0.4%
1.00x(?)
ArrayAppend
2874
2865
-0.3%
1.00x(?)
DictionaryRemove
30422
30428
+0.0%
1.00x(?)
LinkedList
26390
26358
-0.1%
1.00x
RGBHistogram
49481
49605
+0.2%
1.00x(?)
ArrayOfPOD
2335
2328
-0.3%
1.00x(?)
SetUnion
34827
34854
+0.1%
1.00x(?)
Chars
54412
54391
-0.0%
1.00x(?)
DeadArray
128681
128252
-0.3%
1.00x(?)
BitCount
77
77
+0.0%
1.00x
AngryPhonebook
5880
5887
+0.1%
1.00x(?)
SevenBoom
1417
1422
+0.3%
1.00x
ArrayValueProp
2225
2227
+0.1%
1.00x(?)
RC4
8737
8753
+0.2%
1.00x(?)
Memset
20655
20659
+0.0%
1.00x(?)
Dictionary2OfObjects
6936
6918
-0.3%
1.00x(?)
ArrayAppendReserved
2697
2688
-0.3%
1.00x(?)
ArrayValueProp2
18192
18189
-0.0%
1.00x(?)
ArrayValueProp3
2580
2574
-0.2%
1.00x(?)
ObjectiveCBridgeToNSString
1048
1056
+0.8%
0.99x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToString
193233
194649
+0.7%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
159865
160995
+0.7%
0.99x(?)
ObjectiveCBridgeFromNSStringForced
2566
2581
+0.6%
0.99x(?)
Dictionary3OfObjects
2725
2740
+0.6%
0.99x
RangeAssignment
15535
15615
+0.5%
0.99x
SuperChars
691413
694997
+0.5%
0.99x(?)
114
5537837
5570509
+0.6%
0.99x
ObjectAllocation
1292
1301
+0.7%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObject
99870
101059
+1.2%
0.99x
ProtocolDispatch2
421
424
+0.7%
0.99x
ObjectiveCBridgeToNSSet
16368
16559
+1.2%
0.99x(?)
ObjectiveCBridgeToNSArray
29459
29780
+1.1%
0.99x(?)
NSError
648
653
+0.8%
0.99x(?)
DictionaryLiteral
20071
20229
+0.8%
0.99x(?)
Hanoi
22703
22848
+0.6%
0.99x
ArrayOfGenericPOD
3426
3475
+1.4%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObject
66231
66937
+1.1%
0.99x(?)
NSDictionaryCastToSwift
9559
9662
+1.1%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObjectToString
189658
190659
+0.5%
0.99x(?)
NSStringConversion
2553
2570
+0.7%
0.99x(?)
ArrayValueProp4
2522
2535
+0.5%
0.99x(?)
DictionaryBridge
3534
3623
+2.5%
0.98x(?)
PolymorphicCalls
1133
1154
+1.9%
0.98x
StrToInt
18521
18920
+2.1%
0.98x
ObjectiveCBridgeFromNSString
4939
5036
+2.0%
0.98x(?)
ObjectiveCBridgeToNSDictionary
16261
16616
+2.2%
0.98x(?)
ObjectiveCBridgeStubFromArrayOfNSString
108740
111878
+2.9%
0.97x(?)
ObjectiveCBridgeFromNSSetAnyObjectForced
7453
7790
+4.5%
0.96x
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
158374
164849
+4.1%
0.96x(?)
**Hardware Overview**
Model Name: Mac mini
Model Identifier: Macmini7,1
Processor Name: Intel Core i7
Processor Speed: 3 GHz
Number of Processors: 1
Total Number of Cores: 2
L2 Cache (per Core): 256 KB
L3 Cache: 4 MB
Memory: 16 GB
@nadavrot Do these regressions look OK to you? It looks potentially legitimate, since both of the affected benchmarks are type-metadata-access-heavy.
@jckarter I am seeing one improvement and one regression. The numbers we are seeing could be noise. In any case, we'll benchmark this change again with the internal setup of the performance benchmark. I think that this change can go in.
Also looking at the numbers it seems like some of the benchmarks result in low values that can cause what look like larger percentage changes. It seems like they should be enhanced to always achieve a threshold of say 100 if they are going to be measured in the way they are.
bump
Any remaining objections @gparker42 @nadavrot ?
I'm going to be out next week; @gparker42, if this looks good, do you mind merging?
I am building with your code on Cygwin. I found errors.
In file included from /cygdrive/c/Work/swift_msvc/swift/stdlib/public/runtime/Casting.cpp:25:
In file included from /cygdrive/c/Work/swift_msvc/swift/include/swift/Runtime/Mutex.h:24:
/cygdrive/c/Work/swift_msvc/swift/include/swift/Runtime/MutexPThread.h:49:32: error: constexpr function never produces a constant expression [-Winvalid-constexpr]
static constexpr MutexHandle staticInit() {
^
/cygdrive/c/Work/swift_msvc/swift/include/swift/Runtime/MutexPThread.h:50:12: note: cast that performs the conversions of a reinterpret_cast is not allowed in a constant expression
return PTHREAD_MUTEX_INITIALIZER;
^
/usr/include/pthread.h:52:35: note: expanded from macro 'PTHREAD_MUTEX_INITIALIZER'
#define PTHREAD_MUTEX_INITIALIZER PTHREAD_NORMAL_MUTEX_INITIALIZER_NP
^
/usr/include/pthread.h:50:45: note: expanded from macro 'PTHREAD_NORMAL_MUTEX_INITIALIZER_NP'
#define PTHREAD_NORMAL_MUTEX_INITIALIZER_NP (pthread_mutex_t)19
^
@tinysun212 Any suggestions to how to correct? I don't currently have a usable Windows install with Cygwin (it will take me a bit to get it going). I guess Cygwin doesn't support staticly allocated pthread mutex?
I extracted some definitions from pthread.h in Cygwin.
I think you can reproduce above compile error with temporarily changing #include <pthread.h> to #include "swift/Runtime/_pthread.h" in Mutex.h/MutexPThread.h.
typedef __uint32_t pthread_mutex_t; /* identify a mutex */
typedef __uint32_t pthread_cond_t; /* identify a condition variable */
typedef __uint32_t pthread_rwlock_t; /* POSIX RWLock Object */
#define PTHREAD_COND_INITIALIZER (pthread_cond_t)21
#define PTHREAD_MUTEX_INITIALIZER PTHREAD_NORMAL_MUTEX_INITIALIZER_NP
#define PTHREAD_NORMAL_MUTEX_INITIALIZER_NP (pthread_mutex_t)19
#define PTHREAD_RWLOCK_INITIALIZER (pthread_rwlock_t)22
int pthread_cond_broadcast (pthread_cond_t *);
int pthread_cond_destroy (pthread_cond_t *);
int pthread_cond_init (pthread_cond_t *, const pthread_condattr_t *);
int pthread_cond_signal (pthread_cond_t *);
int pthread_cond_wait (pthread_cond_t *, pthread_mutex_t *);
int pthread_mutexattr_init (pthread_mutexattr_t *);
int pthread_mutexattr_settype (pthread_mutexattr_t *, int);
int pthread_mutexattr_destroy (pthread_mutexattr_t *);
int pthread_mutex_destroy (pthread_mutex_t *);
int pthread_mutex_init (pthread_mutex_t *, const pthread_mutexattr_t *);
int pthread_mutex_lock (pthread_mutex_t *);
int pthread_mutex_trylock (pthread_mutex_t *);
int pthread_mutex_unlock (pthread_mutex_t *);
int pthread_rwlock_destroy (pthread_rwlock_t *rwlock);
int pthread_rwlock_init (pthread_rwlock_t *rwlock, const pthread_rwlockattr_t *attr);
int pthread_rwlock_rdlock (pthread_rwlock_t *rwlock);
int pthread_rwlock_tryrdlock (pthread_rwlock_t *rwlock);
int pthread_rwlock_trywrlock (pthread_rwlock_t *rwlock);
int pthread_rwlock_unlock (pthread_rwlock_t *rwlock);
int pthread_rwlock_wrlock (pthread_rwlock_t *rwlock);
I guess we will have to predicate the use of constexpr on not Cygwin. It won't make cygwin any worse then it apparently was (doesn't have a "true" static allocated mutex, etc.) while allowing other platforms to correctly honor static allocation (under C++).
Sorry, I captured wrong header. In include/cygwin/types.h, the typedef part is as follows,
typedef struct __pthread_condattr_t {char __dummy;} *pthread_condattr_t;
typedef struct __pthread_cond_t {char __dummy;} *pthread_cond_t;
typedef struct __pthread_rwlock_t {char __dummy;} *pthread_rwlock_t;
typedef struct __pthread_rwlockattr_t {char __dummy;} *pthread_rwlockattr_t;
I wrote simple source to test.
$ cat tt.cpp
typedef struct __pthread_cond_t {char __dummy;} *pthread_cond_t;
#define PTHREAD_COND_INITIALIZER (pthread_cond_t)21
typedef pthread_cond_t ConditionHandle;
static constexpr ConditionHandle staticInit() {
return PTHREAD_COND_INITIALIZER;
};
$ clang++ -c -std=c++11 tt.cpp
tt.cpp:7:34: error: constexpr function never produces a constant expression [-Winvalid-constexpr]
static constexpr ConditionHandle staticInit() {
^
tt.cpp:8:10: note: cast that performs the conversions of a reinterpret_cast is not allowed in a constant expression
return PTHREAD_COND_INITIALIZER;
^
tt.cpp:3:34: note: expanded from macro 'PTHREAD_COND_INITIALIZER'
#define PTHREAD_COND_INITIALIZER (pthread_cond_t)21
^
1 error generated.
Do you have any idea to keep constexpr?
@tinysun212 I just pushed a change that should deal with the issue you reported under Cygwin on Windows. If possible can you try it out to verify it works as expected?
It is successfully compiled without any warning or error. Thanks for your patch.
@tinysun212 cool, you will need to update your Win32 implementation have the following in its header.
#define CONDITION_SUPPORTS_CONSTEXPR 1
#define MUTEX_SUPPORTS_CONSTEXPR 1
#define READWRITELOCK_SUPPORTS_CONSTEXPR 1
@gparker42 @jckarter ok I think this ready to go after a CI
@swift-ci Please test
@swift-ci Please benchmark
Build comment file:
Optimized (O)
Regression (1)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeToNSDictionary
16422
17423
+6.1%
0.94x
Improvement (3)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeFromNSArrayAnyObjectToString
166571
157769
-5.3%
1.06x
Calculator
35
33
-5.7%
1.06x
StringHasSuffix
714
675
-5.5%
1.06x(?)
No Changes (117)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ObjectiveCBridgeToNSSet
16729
15967
-4.5%
1.05x(?)
ObjectiveCBridgeFromNSDictionaryAnyObject
196894
189306
-3.9%
1.04x(?)
DictionaryOfObjects
2264
2204
-2.6%
1.03x
Dictionary
731
713
-2.5%
1.03x
SetExclusiveOr_OfObjects
8077
7816
-3.2%
1.03x
SetUnion_OfObjects
6748
6546
-3.0%
1.03x
SortStrings
1762
1730
-1.8%
1.02x
ObjectiveCBridgeStubToNSString
1415
1392
-1.6%
1.02x
AngryPhonebook
3389
3337
-1.5%
1.02x(?)
DictionaryRemove
5015
4907
-2.1%
1.02x(?)
ObjectiveCBridgeFromNSSetAnyObjectToString
166038
162739
-2.0%
1.02x(?)
ObjectiveCBridgeStubFromArrayOfNSString
109552
107570
-1.8%
1.02x
ObjectiveCBridgeToNSString
1062
1055
-0.7%
1.01x(?)
PopFrontArray
1060
1054
-0.6%
1.01x(?)
Array2D
2024
2012
-0.6%
1.01x(?)
DictionaryBridge
3636
3609
-0.7%
1.01x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectForced
4574
4544
-0.7%
1.01x(?)
Prims
714
710
-0.6%
1.01x
DictionarySwap
757
752
-0.7%
1.01x(?)
RangeAssignment
273
270
-1.1%
1.01x
StrComplexWalk
3023
2993
-1.0%
1.01x(?)
ObjectAllocation
143
142
-0.7%
1.01x(?)
Dictionary2
1918
1904
-0.7%
1.01x
ArrayOfRef
3451
3429
-0.6%
1.01x(?)
ObjectiveCBridgeStubToArrayOfNSString
30347
30111
-0.8%
1.01x(?)
Phonebook
7221
7130
-1.3%
1.01x
NSStringConversion
612
606
-1.0%
1.01x
Chars
618
609
-1.5%
1.01x(?)
StrToInt
4936
4866
-1.4%
1.01x
DeadArray
173
172
-0.6%
1.01x
120
2294275
2269762
-1.1%
1.01x
ArraySubscript
1325
1321
-0.3%
1.00x(?)
MonteCarloPi
42667
42668
+0.0%
1.00x(?)
StackPromo
18693
18709
+0.1%
1.00x(?)
RecursiveOwnedParameter
1826
1827
+0.1%
1.00x(?)
Integrate
233
232
-0.4%
1.00x
ClassArrayGetter
12
12
+0.0%
1.00x
Histogram
639
639
+0.0%
1.00x
StringWithCString
75595
75613
+0.0%
1.00x(?)
ObjectiveCBridgeFromNSStringForced
2187
2186
-0.1%
1.00x(?)
SortLettersInPlace
940
943
+0.3%
1.00x(?)
ArrayOfGenericPOD
207
208
+0.5%
1.00x
Dictionary3OfObjects
850
846
-0.5%
1.00x
ByteSwap
1
1
+0.0%
1.00x
SuperChars
276889
275774
-0.4%
1.00x
XorLoop
359
359
+0.0%
1.00x
ObjectiveCBridgeStubFromNSStringRef
327
327
+0.0%
1.00x
StaticArray
2648
2643
-0.2%
1.00x(?)
ProtocolDispatch
2892
2888
-0.1%
1.00x(?)
TypeFlood
0
0
+0.0%
1.00x
ProtocolDispatch2
151
151
+0.0%
1.00x
Dictionary3
497
498
+0.2%
1.00x(?)
StringHasPrefix
576
577
+0.2%
1.00x(?)
SetIntersect_OfObjects
2360
2358
-0.1%
1.00x(?)
ErrorHandling
2859
2845
-0.5%
1.00x(?)
ObjectiveCBridgeToNSArray
30182
30037
-0.5%
1.00x(?)
NSError
359
359
+0.0%
1.00x
PopFrontArrayGeneric
1049
1051
+0.2%
1.00x(?)
PopFrontUnsafePointer
8657
8655
-0.0%
1.00x(?)
CaptureProp
3741
3742
+0.0%
1.00x(?)
PolymorphicCalls
61
61
+0.0%
1.00x
RC4
250
250
+0.0%
1.00x
MapReduce
327
327
+0.0%
1.00x
Hanoi
3405
3405
+0.0%
1.00x
OpenClose
48
48
+0.0%
1.00x
DictionaryRemoveOfObjects
19503
19471
-0.2%
1.00x(?)
UTF8Decode
260
260
+0.0%
1.00x
SortStringsUnicode
8994
9013
+0.2%
1.00x(?)
SetIsSubsetOf
548
549
+0.2%
1.00x(?)
NopDeinit
34619
34609
-0.0%
1.00x
ObjectiveCBridgeFromNSArrayAnyObject
61546
61842
+0.5%
1.00x
SetIntersect
1300
1298
-0.1%
1.00x(?)
SetExclusiveOr
3830
3836
+0.2%
1.00x(?)
RGBHistogramOfObjects
21199
21271
+0.3%
1.00x(?)
StringBuilder
1389
1390
+0.1%
1.00x(?)
ArrayInClass
85
85
+0.0%
1.00x
ArrayOfGenericRef
3519
3524
+0.1%
1.00x(?)
Sim2DArray
410
412
+0.5%
1.00x
MonteCarloE
9996
10032
+0.4%
1.00x
StringHasSuffixUnicode
62311
62479
+0.3%
1.00x(?)
HashTest
1819
1816
-0.2%
1.00x(?)
SetIsSubsetOf_OfObjects
617
617
+0.0%
1.00x
ArrayAppend
730
728
-0.3%
1.00x(?)
StringHasPrefixUnicode
14981
15017
+0.2%
1.00x(?)
LinkedList
6847
6843
-0.1%
1.00x
RGBHistogram
3233
3238
+0.1%
1.00x(?)
ObjectiveCBridgeFromNSSetAnyObjectForced
3981
3968
-0.3%
1.00x(?)
ArrayAppendReserved
507
507
+0.0%
1.00x
ArrayOfPOD
171
171
+0.0%
1.00x
SetUnion
3385
3398
+0.4%
1.00x(?)
BitCount
1
1
+0.0%
1.00x
ArrayLiteral
979
981
+0.2%
1.00x(?)
SevenBoom
1276
1274
-0.2%
1.00x
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
159858
159122
-0.5%
1.00x(?)
StringWalk
6008
6014
+0.1%
1.00x(?)
ArrayValueProp
5
5
+0.0%
1.00x
GlobalClass
0
0
+0.0%
1.00x
Memset
222
223
+0.5%
1.00x
Dictionary2OfObjects
3293
3283
-0.3%
1.00x(?)
ArrayValueProp4
5
5
+0.0%
1.00x
TwoSum
1642
1643
+0.1%
1.00x(?)
ArrayValueProp2
5
5
+0.0%
1.00x
ArrayValueProp3
5
5
+0.0%
1.00x
DictionarySwapOfObjects
6606
6658
+0.8%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
3609
3633
+0.7%
0.99x(?)
StringInterpolation
9999
10054
+0.6%
0.99x(?)
Join
439
445
+1.4%
0.99x(?)
ObjectiveCBridgeStubFromNSString
694
700
+0.9%
0.99x
StringEqualPointerComparison
7102
7160
+0.8%
0.99x
ObjectiveCBridgeFromNSString
1561
1572
+0.7%
0.99x
ObjectiveCBridgeFromNSSetAnyObjectToStringForced
101734
103174
+1.4%
0.99x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToString
180143
182976
+1.6%
0.98x(?)
ObjectiveCBridgeFromNSSetAnyObject
95459
97047
+1.7%
0.98x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToStringForced
114004
116779
+2.4%
0.98x(?)
NSDictionaryCastToSwift
8281
8422
+1.7%
0.98x(?)
DictionaryLiteral
3006
3113
+3.6%
0.97x(?)
Walsh
317
329
+3.8%
0.96x
**Unoptimized (Onone)**
Regression (1)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
TypeFlood
145
156
+7.6%
0.93x(?)
Improvement (3)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
StringHasPrefix
1630
1516
-7.0%
1.08x
ByteSwap
10
9
-10.0%
1.11x
StringWalk
26926
23276
-13.6%
1.16x
No Changes (117)
TEST
OLD_MIN
NEW_MIN
DELTA (%)
SPEEDUP
ArrayOfPOD
2256
2152
-4.6%
1.05x
StringHasSuffix
1561
1487
-4.7%
1.05x
ObjectiveCBridgeToNSSet
16775
16073
-4.2%
1.04x(?)
NopDeinit
60929
59005
-3.2%
1.03x
ObjectiveCBridgeFromNSSetAnyObjectToString
184614
179171
-3.0%
1.03x(?)
OpenClose
452
442
-2.2%
1.02x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectToStringForced
122508
119650
-2.3%
1.02x(?)
StringHasPrefixUnicode
16769
16441
-2.0%
1.02x
ObjectiveCBridgeFromNSSetAnyObjectToStringForced
115560
113711
-1.6%
1.02x(?)
ArrayAppendReserved
2701
2659
-1.6%
1.02x(?)
Chars
53814
52725
-2.0%
1.02x(?)
PopFrontArray
11670
11546
-1.1%
1.01x
SortStrings
14096
13993
-0.7%
1.01x
SortLettersInPlace
4158
4134
-0.6%
1.01x(?)
ObjectiveCBridgeFromNSDictionaryAnyObject
199443
197410
-1.0%
1.01x(?)
Integrate
347
345
-0.6%
1.01x(?)
Join
1370
1363
-0.5%
1.01x(?)
ErrorHandling
7549
7464
-1.1%
1.01x(?)
PopFrontArrayGeneric
11885
11821
-0.5%
1.01x(?)
PopFrontUnsafePointer
96350
95260
-1.1%
1.01x(?)
StringEqualPointerComparison
9779
9645
-1.4%
1.01x
ArrayOfGenericPOD
3493
3449
-1.3%
1.01x(?)
DictionaryRemoveOfObjects
60808
60183
-1.0%
1.01x
UTF8Decode
38869
38347
-1.3%
1.01x
SortStringsUnicode
23042
22860
-0.8%
1.01x
StrToInt
19217
18965
-1.3%
1.01x(?)
ArrayAppend
3130
3112
-0.6%
1.01x(?)
LinkedList
26288
26108
-0.7%
1.01x(?)
StringBuilder
8351
8250
-1.2%
1.01x(?)
ObjectiveCBridgeStubFromArrayOfNSString
113665
112449
-1.1%
1.01x(?)
ArrayValueProp4
2484
2449
-1.4%
1.01x
ArraySubscript
5398
5380
-0.3%
1.00x(?)
ObjectiveCBridgeToNSString
1052
1052
+0.0%
1.00x
ObjectiveCBridgeStubToNSString
1663
1667
+0.2%
1.00x
ClassArrayGetter
1182
1183
+0.1%
1.00x(?)
Array2D
759861
760971
+0.1%
1.00x
Histogram
13495
13503
+0.1%
1.00x(?)
MonteCarloPi
50444
50489
+0.1%
1.00x(?)
StringWithCString
73856
73828
-0.0%
1.00x(?)
ObjectiveCBridgeFromNSStringForced
2605
2597
-0.3%
1.00x(?)
Prims
14107
14082
-0.2%
1.00x(?)
Dictionary3OfObjects
2653
2652
-0.0%
1.00x(?)
RangeAssignment
12817
12807
-0.1%
1.00x(?)
SuperChars
720170
720255
+0.0%
1.00x(?)
XorLoop
18026
18042
+0.1%
1.00x(?)
ObjectiveCBridgeStubFromNSStringRef
649
649
+0.0%
1.00x
StaticArray
17222
17296
+0.4%
1.00x(?)
ProtocolDispatch
5395
5398
+0.1%
1.00x(?)
ObjectAllocation
1282
1277
-0.4%
1.00x(?)
ArrayLiteral
1039
1038
-0.1%
1.00x(?)
HashTest
6172
6158
-0.2%
1.00x(?)
Dictionary3
2024
2023
-0.1%
1.00x(?)
Dictionary2
5100
5094
-0.1%
1.00x(?)
StrComplexWalk
8557
8558
+0.0%
1.00x(?)
SetIntersect_OfObjects
18947
18892
-0.3%
1.00x(?)
ArrayOfRef
8836
8796
-0.5%
1.00x
ObjectiveCBridgeToNSArray
30251
30255
+0.0%
1.00x(?)
DictionaryOfObjects
6904
6931
+0.4%
1.00x(?)
CaptureProp
52080
52161
+0.2%
1.00x(?)
RC4
8433
8430
-0.0%
1.00x(?)
MapReduce
41976
41773
-0.5%
1.00x
Hanoi
18879
18903
+0.1%
1.00x(?)
SetIsSubsetOf
3012
3002
-0.3%
1.00x
SetIntersect
18957
18953
-0.0%
1.00x(?)
SetExclusiveOr
48910
48885
-0.1%
1.00x(?)
RGBHistogramOfObjects
114910
114596
-0.3%
1.00x(?)
ArrayInClass
3562
3560
-0.1%
1.00x(?)
ArrayOfGenericRef
9464
9493
+0.3%
1.00x(?)
Sim2DArray
13160
13155
-0.0%
1.00x(?)
SetExclusiveOr_OfObjects
67180
67096
-0.1%
1.00x(?)
MonteCarloE
102055
102489
+0.4%
1.00x(?)
SetUnion_OfObjects
51266
51226
-0.1%
1.00x(?)
StringHasSuffixUnicode
64403
64445
+0.1%
1.00x(?)
Walsh
11993
11994
+0.0%
1.00x(?)
SetIsSubsetOf_OfObjects
2836
2840
+0.1%
1.00x
DictionaryRemove
30238
30204
-0.1%
1.00x(?)
NSStringConversion
2596
2608
+0.5%
1.00x(?)
RGBHistogram
65805
65844
+0.1%
1.00x(?)
SetUnion
35038
35061
+0.1%
1.00x(?)
BitCount
92
92
+0.0%
1.00x
AngryPhonebook
6087
6080
-0.1%
1.00x(?)
ObjectiveCBridgeFromNSArrayAnyObjectToStringForced
159358
158771
-0.4%
1.00x(?)
ArrayValueProp
2176
2172
-0.2%
1.00x(?)
120
5250686
5254294
+0.1%
1.00x
GlobalClass
0
0
+0.0%
1.00x
Memset
18649
18643
-0.0%
1.00x(?)
ArrayValueProp2
17864
17913
+0.3%
1.00x
SevenBoom
1436
1430
-0.4%
1.00x(?)
StackPromo
161665
164036
+1.5%
0.99x
RecursiveOwnedParameter
7555
7607
+0.7%
0.99x
ObjectiveCBridgeFromNSDictionaryAnyObjectToString
195015
196489
+0.8%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObjectToString
156329
157882
+1.0%
0.99x(?)
ObjectiveCBridgeFromNSDictionaryAnyObjectForced
7059
7137
+1.1%
0.99x(?)
ObjectiveCBridgeFromNSArrayAnyObjectForced
6948
7012
+0.9%
0.99x(?)
DictionarySwap
11425
11532
+0.9%
0.99x(?)
StringInterpolation
15303
15419
+0.8%
0.99x(?)
ObjectiveCBridgeStubFromNSString
1122
1130
+0.7%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObject
103827
104526
+0.7%
0.99x(?)
ProtocolDispatch2
425
428
+0.7%
0.99x
NSError
634
639
+0.8%
0.99x(?)
PolymorphicCalls
1137
1151
+1.2%
0.99x(?)
Calculator
938
952
+1.5%
0.99x
DictionaryLiteral
20228
20461
+1.1%
0.99x(?)
ObjectiveCBridgeFromNSString
5072
5105
+0.7%
0.99x(?)
Phonebook
84595
85233
+0.8%
0.99x(?)
ObjectiveCBridgeFromNSSetAnyObjectForced
7430
7504
+1.0%
0.99x(?)
DeadArray
136053
137218
+0.9%
0.99x(?)
ArrayValueProp3
2503
2519
+0.6%
0.99x
Dictionary2OfObjects
6866
6903
+0.5%
0.99x(?)
TwoSum
9227
9288
+0.7%
0.99x(?)
DictionarySwapOfObjects
25014
25456
+1.8%
0.98x
DictionaryBridge
3669
3753
+2.3%
0.98x(?)
Dictionary
4130
4199
+1.7%
0.98x(?)
ObjectiveCBridgeFromNSArrayAnyObject
65215
66489
+1.9%
0.98x(?)
NSDictionaryCastToSwift
9712
9926
+2.2%
0.98x
ObjectiveCBridgeStubToArrayOfNSString
30388
31338
+3.1%
0.97x
ObjectiveCBridgeToNSDictionary
16430
16901
+2.9%
0.97x(?)
**Hardware Overview**
Model Name: Mac mini
Model Identifier: Macmini7,1
Processor Name: Intel Core i7
Processor Speed: 3 GHz
Number of Processors: 1
Total Number of Cores: 2
L2 Cache (per Core): 256 KB
L3 Cache: 4 MB
Memory: 16 GB
@swift-ci Please test
It looks like the Linux build issues continues to affect things?
ERROR: testREPL (TestREPLArray.REPLArrayTestCase)
@swift-ci Please test
I noted a few stylistic things that should be changed, but this is good enough to accept now and polish later.
@gparker42 I will work on the changes outlined and open a new pull request when ready. Thanks
|
2025-04-01T06:37:56.368855
| 2019-04-20T07:46:45
|
435367887
|
{
"authors": [
"MaxDesiatov",
"compnerd",
"stephentyrone",
"tbkka",
"zhuowei"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3772",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/24181"
}
|
gharchive/pull-request
|
[stdlib] Fix return type of swift_{uint64,int64,float*}ToString
The return type of these functions are uint64_t in Stubs.cpp but UInt in the Swift code; this changes the Swift code to match the C++ return type.
Found when compiling the stdlib for WebAssembly, which requires that all return types match: UInt maps to i32 while uint64_t maps to i64, so functions calling these functions fail the validation.
CC: @airspeedswift
@airspeedswift would you be able to have a look please? Is there anyone who else who could be pinged about reviewing this PR?
@swift-ci please smoke test
Thanks @zhuowei. Sorry that we didn't see this earlier. FWIW, your best bet is to tag me or @tbkka for this stuff.
N.B. this was a "safe" return type mismatch because the actual implementation has a signature compatible with the signature used; the high 32b of the result were simply ignored by the caller on 32b targets, but they are semantically guaranteed to be zero anyway.
The return value here is the length of the formatted string, so a 32-bit value would be preferable to avoid overhead on 32-bit platforms. I believe this is all internal, so changing the type won't break the ABI.
Of course, that might mean touching more places to get the types to line up all the way down.
Does it mean that this PR needs any more changes to pass the review and be merged? I'm a co-maintainer of @swiftwasm, and as far as I'm aware @zhuowei is not actively maintaining it currently, but I have access to the PR branch and I'm ready to push it forward and to apply any changes if needed
I agree with @tbkka that it would be better if these just returned int32. But, the cpp functions are marked SWIFT_RUNTIME_STDLIB_API, so we formally can't modify them (they are API), and the difference in efficiency is minimal.
@MaxDesiatov since we know that the result is always in-range, can you replace the Int inits (line 172 and similar) with Int(truncatingIfNeeded: ...)? This will eliminate the check so that the only overhead on 32b platforms is zeroing out a register, which is sufficiently cheap to not worry about.
Actually, I'll just merge this and make the other change myself as a follow on. Thanks all!
|
2025-04-01T06:37:56.375421
| 2016-06-01T22:58:10
|
158025619
|
{
"authors": [
"CodaFi",
"harlanhaskins",
"slavapestov"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3773",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/2828"
}
|
gharchive/pull-request
|
Fix off-by-one in @noescape fixit
What's in this pull request?
This
func noescapeD(@noescape f: () -> Bool) {}
used to become this
func noescapeD( f: @noescape() -> Bool) {}
It now becomes
func noescapeT(f: @noescape () -> Bool) {}
Moreover this
func autoclosureD(@autoclosure f: () -> Bool) {}
used to become this
func autoclosureD( f: @autoclosure@noescape() -> Bool) {}
It now becomes
func autoclosureD(f: @autoclosure @noescape () -> Bool) {}
Before merging this pull request to apple/swift repository:
[ ] Test pull request on Swift continuous integration.
Triggering Swift CI
The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are:
Smoke Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please smoke test
All supported platforms | @swift-ci Please smoke test and merge
OS X platform | @swift-ci Please smoke test OS X platform
Linux platform | @swift-ci Please smoke test Linux platform
Validation Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please test
All supported platforms | @swift-ci Please test and merge
OS X platform | @swift-ci Please test OS X platform
OS X platform | @swift-ci Please benchmark
Linux platform | @swift-ci Please test Linux platform
Lint Testing
Language | Comment
------------ | -------------
Python | @swift-ci Please Python lint
Note: Only members of the Apple organization can trigger swift-ci.
@swift-ci Please fix
@slavapestov I wish it worked that way...
@swift-ci please test
Oh man, this is not the first time I asked CI to 'please fix'. I guess I'm just begging our AI overlords to automate me out of a job...
The Linux failure is unrelated. @shahmishal?
Otherwise this LGTM.
|
2025-04-01T06:37:56.376995
| 2020-07-03T14:06:56
|
650619834
|
{
"authors": [
"CodaFi",
"Jumhyn",
"varungandhi-apple"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3774",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/32701"
}
|
gharchive/pull-request
|
Update README for new target names
@swift-ci smoke test and merge
@swift-ci smoke test and merge
Hmm. Why did the smoke tests not run? 😕
@swift-ci please smoke test and merge
|
2025-04-01T06:37:56.380210
| 2020-07-17T01:56:31
|
658748273
|
{
"authors": [
"gottesmm",
"theblixguy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3775",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/32940"
}
|
gharchive/pull-request
|
[ast] Rename VarPattern -> BindingPattern.
VarPattern is today used to implement both 'let' and 'var' pattern bindings, so
today is already misleading. The reason why the name Var was chosen was done b/c
it is meant to represent a pattern that performs 'variable binding'. Given that
I am going to add a new 'inout' pattern binding to this, it makes sense to
give it now a better fitting name before I make things more confusing.
Its actually a NFC.
@swift-ci test
TSPL refers to it as “value-binding pattern”. Should it be reflected here (i.e. ValueBindingPattern) or should TSPL be updated? Or does it not matter at all?
I don't think it matters. And I am not sure the larger name helps. I am trying not to change things in a large way. Only what I need.
@theblixguy the bigger change here is that I am going to change how it represents lets/vars to use a true kind + exhaustive switches instead of isLet(). That will make it easier to update this later. This is just the first in a series of changes.
|
2025-04-01T06:37:56.387989
| 2016-07-21T23:44:59
|
166942389
|
{
"authors": [
"dabrahams"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3776",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/3671"
}
|
gharchive/pull-request
|
id-as-Any support
What's in this pull request?
Resolved bug number: (SR-)
Before merging this pull request to apple/swift repository:
[ ] Test pull request on Swift continuous integration.
Triggering Swift CI
The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are:
Smoke Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please smoke test
All supported platforms | @swift-ci Please smoke test and merge
OS X platform | @swift-ci Please smoke test OS X platform
Linux platform | @swift-ci Please smoke test Linux platform
A smoke test on macOS does the following:
Builds the compiler incrementally.
Builds the standard library only for macOS. Simulator standard libraries and
device standard libraries are not built.
lldb is not built.
The test and validation-test targets are run only for macOS. The optimized
version of these tests are not run.
A smoke test on Linux does the following:
Builds the compiler incrementally.
Builds the standard library incrementally.
lldb is built incrementally.
The swift test and validation-test targets are run. The optimized version of these
tests are not run.
lldb is tested.
Validation Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please test
All supported platforms | @swift-ci Please test and merge
OS X platform | @swift-ci Please test OS X platform
OS X platform | @swift-ci Please benchmark
Linux platform | @swift-ci Please test Linux platform
Lint Testing
Language | Comment
------------ | -------------
Python | @swift-ci Please Python lint
Note: Only members of the Apple organization can trigger swift-ci.
Apparently direct casting is all working now!
But indirect casting
(bouncing through Any) fails in really strange ways, without reaching
any known stdlib entry points.
@swift-ci Please test Linux platform
@swift-ci Please test and merge
@swift-ci Please test and merge
@swift-ci Please test and merge
@swift-ci Please test and merge
|
2025-04-01T06:37:56.391690
| 2021-07-01T18:23:19
|
935106047
|
{
"authors": [
"mikeash",
"swift-ci"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3777",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/38203"
}
|
gharchive/pull-request
|
[Reflection] Fix iterateAsyncTaskAllocations.
Fix the declaration of AsyncTask and add a test for iterateAsyncTaskAllocations. Reflection's declaration of AsyncTask had fallen out of sync with the real thing. The test that was supposed to catch this was never actually committed, oops.
While we're in there, switch out uses of Builtin.getCurrentAsyncTask for the less error prone _getCurrentAsyncTask.
rdar://80035307
@swift-ci please test
Build failed
Swift Test Linux Platform
Git Sha - f0e41d7414ae943e156e68664e08620cab74c723
Build failed
Swift Test OS X Platform
Git Sha - f0e41d7414ae943e156e68664e08620cab74c723
@swift-ci please test
@swift-ci please test
@swift-ci please test
|
2025-04-01T06:37:56.398192
| 2016-10-05T03:22:33
|
181056465
|
{
"authors": [
"atrick",
"bob-wilson",
"shahmishal",
"swift-ci"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3778",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/5129"
}
|
gharchive/pull-request
|
Add descriptive doc comments to withMemoryRebound(to:capacity:_:) and…
… bindMemory(to:capacity:).
SR-2480: Improve documentation for withMemoryRebound(to:capacity:_:)
rdar://problem/28440528
––– CCC Information –––
• Explanation: Users migrating to Swift 3 often hit a compiler error with a diagnostic pointing them to this API. I have interacted with several developers who don't understand how to use the API. A more detailed description was necessary in the doc comments.
• Scope of Issue: Developers migrating to Swift 3 and using UnsafePointers.
• Origination: SE-0107 UnsafeRawPointer--terse doc comments.
• Risk: None whatsoever.
• Reviewed By: Dave Abrahams
• Testing: None needed.
• Directions for QE: None
@swift-ci test.
@swift-ci test OS X.
@swift-ci test OS X
Build failed
Jenkins build - Swift Test OS X Platform
Git Commit - 36f3921f65b3afb00582c4341800963bf0fb2c0e
Test requested by - @bob-wilson
This failed because it was merged before testing was completed.
|
2025-04-01T06:37:56.399594
| 2022-11-28T19:36:12
|
1466984469
|
{
"authors": [
"aschwaighofer"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3779",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/62280"
}
|
gharchive/pull-request
|
Fix and re-enable test resilient_debug_value.sil
In non-asserts compiler we don't seem to get named llvm values.
rdar://102535969
@swift-ci test
|
2025-04-01T06:37:56.403847
| 2023-02-04T01:50:46
|
1570700483
|
{
"authors": [
"kavon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3780",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/63427"
}
|
gharchive/pull-request
|
allow move-only types to conform to Sendable
Part of the reason why we do not want to
permit conformance to protocols for
move-only types is that they're
fundamentally wrong: all existing protocols
assume the type is copyable, so we'd be
allowing people to write conformances to
things that are not actually true.
The other aspect of it is that we may
need to change the runtime representation
for conformance descriptors of move-only
types. So we can't have these conformances
leaving residue at runtime. Luckily, that
means marker protocols would be OK,
since they leave no residue at runtime.
So for, now we're going to specifically
permit move-only types to conform to the
marker protocol Sendable since it's needed
for move-only types to work with concurrency.
All of the existing rules about mixing
move-only types with generics still apply.
That means you still can't turn it into the
existential any Sendable in any way at all,
despite it conforming. The purpose of the
conformance is purely to allow the concrete
instances to cross actor isolation boundaries
if it is actually Sendable.
resolves rdar://104987062
@swift-ci please smoke test and merge
@swift-ci please smoke test and merge
|
2025-04-01T06:37:56.405690
| 2023-06-12T10:30:57
|
1752428601
|
{
"authors": [
"hamishknight"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3781",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/66551"
}
|
gharchive/pull-request
|
[SILGen] Emit block after unreachable when emitting if/switch expressions
When emitting the underlying switch statement for a switch expression, we emit an unreachable if the subject is uninhabited. Statement emission code can handle this, but expression emission expects an RValue to handed back. To remedy this, emit an unreachable block that we can emit the rest of the expression emission code into. The SILOptimizer will then drop this unreachable block.
Resolves #66490
@swift-ci please test
@swift-ci please test Windows
|
2025-04-01T06:37:56.407327
| 2023-10-09T17:39:51
|
1933519623
|
{
"authors": [
"hamishknight"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3782",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/69074"
}
|
gharchive/pull-request
|
NFC: Remove ClangModuleLoader.h include from ExtInfo.h
This was unnecessarily pulling in a whole bunch of Clang headers when all was needed was a forward declaration.
@swift-ci please smoke test
@swift-ci please smoke test
@swift-ci please smoke test
|
2025-04-01T06:37:56.409210
| 2023-11-29T18:53:41
|
2017217045
|
{
"authors": [
"tshortli"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3783",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/70095"
}
|
gharchive/pull-request
|
Frontend: Add -experimental-skip-non-inlinable-function-bodies-is-lazy
This option causes the -experimental-lazy-typecheck and -experimental-skip-non-exportable-decls options to be inferred from the presence of -experimental-skip-non-inlinable-function-bodies. This new option is meant to be a temporary testing aid that allows lazy typechecking to be tested on projects without full build system support for passing the other flags to the right jobs.
Also, ignore -experimental-lazy-typecheck during interface verification jobs, since the flag does not make sense during that action but the driver will pass the flag down regardless.
Resolves rdar://118938251
@swift-ci please smoke test
|
2025-04-01T06:37:56.420035
| 2017-02-04T01:04:07
|
205309496
|
{
"authors": [
"aciidb0mb3r",
"jrose-apple",
"slavapestov",
"swift-ci"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3784",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/7241"
}
|
gharchive/pull-request
|
Handle implicit member references inside extensions of nested types
Within an extension, references to other members of the extended type are permitted without qualification. This is intended to work even when the extended type was a nested type, although members of the enclosing type are not visible in this case. In order to implement this, the type checker pre-checks to see if there are any members with this name and then rewrites the unqualified reference to a qualified one, based on an unresolved TypeExpr with the name of the enclosing type. Unfortunately, if the enclosing type is a nested type, that isn't going to work very well—we find the correct declaration, but fail to map it into context by virtue of not realizing where it came from. Fix this by explicitly checking for this case.
SR-3847
@swift-ci Please test
Oh I see, thanks.
Does that mean the fix makes sense? I admit I'm not very familiar with this part of name resolution.
Build failed
Jenkins build - Swift Test OS X Platform
Git Commit - ac5cb46c2c2e14201fc8fcb93bf74842c04da34b
Test requested by - @jrose-apple
Build failed
Jenkins build - Swift Test Linux Platform
Git Commit - ac5cb46c2c2e14201fc8fcb93bf74842c04da34b
Test requested by - @jrose-apple
@swift-ci Please test
Build failed
Jenkins build - Swift Test OS X Platform
Git Commit - ac5cb46c2c2e14201fc8fcb93bf74842c04da34b
Test requested by - @jrose-apple
Build failed
Jenkins build - Swift Test Linux Platform
Git Commit - ac5cb46c2c2e14201fc8fcb93bf74842c04da34b
Test requested by - @jrose-apple
mapTypeIntoContext() translates interface types to archetypes. Here, I think the code path is resolving an unbound generic reference, which implicitly gets generic parameters from context.
LGTM, by the way. Longer term, we should try to remove findDeclContextForType() altogether. It duplicates the DeclContext walk that's already done in AST/NameLookup.cpp, because we just throw out that information in UnqualifiedLookup.
Test case added.
@swift-ci Please smoke test
Unrelated crash in serialization. @aciidb0mb3r, have you seen this before? Did you file it already?
3. While type-checking expression at [/home/buildnode/jenkins/workspace/swift-PR-Linux-smoke-test/branch-master/swiftpm/Sources/PackageGraph/PackageGraphLoader.swift:111:35 - line:111:118] RangeText="externalModuleDependencies.lazy.map({$0.name}).duplicates(modules.lazy.map{$0.name})"
4. While loading members for 'LazyMapRandomAccessCollection' in module 'Swift'
5. While deserializing 'init' (ConstructorDecl #11643)in 'Swift'
6. While deserializing 'transform' (ParamDecl #22123)in 'Swift'
@swift-ci Please smoke test Linux
@jrose-apple Haven't seen that one, seems new
Another failure.
swift: /home/buildnode/disk2/workspace/swift-PR-Linux-smoke-test/branch-master/swift/lib/AST/ArchetypeBuilder.cpp:2484: auto swift::ArchetypeBuilder::getGenericEnvironment(swift::GenericSignature *)::(anonymous class)::operator()(swift::ArchetypeBuilder::PotentialArchetype *) const: Assertion `(inContext->isEqual(repInContext) || inContext->hasError() || repInContext->hasError()) && "Potential archetype mapping differs from representative!"' failed.
@swift-ci Please smoke test Linux
|
2025-04-01T06:37:56.422508
| 2024-03-28T10:39:33
|
2212888136
|
{
"authors": [
"al45tair"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3785",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/72653"
}
|
gharchive/pull-request
|
[Demangler] Fix assertion failure.
It's illegal to call node->addChild() with a NULL child argument; it's possible to construct unexpected Node trees by passing invalid manglings, and in this case that was causing popTypeAndGetChild() to fail (because the top node was not a Type node), which then meant that the call to addChild had a NULL child argument.
The simplest fix is to use createWithChildren() to do the node construction, because that function checks its arguments for NULLs.
rdar://125350219
@swift-ci Please smoke test
|
2025-04-01T06:37:56.424695
| 2024-06-12T08:05:57
|
2348105932
|
{
"authors": [
"gottesmm"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3786",
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/74326"
}
|
gharchive/pull-request
|
[6.0][region-isolation] Dont crash when processing global actor isolated init accessors.
This just means that I stopped treating it like an actor instance isolated thing. This was fun to track down since ActorIsolation has a union in it that was being misinterpreted, leading to memory corruption... my favorite! = ).
rdar://129256560
(cherry picked from commit f035590784023f16136abd7b593151b3f0375829)
Going to add CCC in the morning
@swift-ci test
Original: https://github.com/apple/swift/pull/74321
|
2025-04-01T06:37:56.462041
| 2017-07-29T14:57:16
|
246526055
|
{
"authors": [
"tamalsaha"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3787",
"repo": "appscode/searchlight",
"url": "https://github.com/appscode/searchlight/issues/166"
}
|
gharchive/issue
|
Fix HTTP client with incluster config
Remove inclusterconfig option https://github.com/appscode/searchlight/blob/master/docs/guides/cluster-alerts/json_path.md
Fixed
|
2025-04-01T06:37:56.473380
| 2023-05-04T11:52:21
|
1695858029
|
{
"authors": [
"danciaclara",
"jnikhila"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3788",
"repo": "appsmithorg/appsmith-docs",
"url": "https://github.com/appsmithorg/appsmith-docs/issues/1208"
}
|
gharchive/issue
|
[Docs]: Upgrade to Business Edition - Kubernetes
Is there an existing issue for this?
[X] I have searched the existing issues
Documentation Link
https://docs.appsmith.com/getting-started/setup/upgrade-to-business-edition/kubernetes
Discord/slack/intercom Link
No response
Describe the problem and improvement.
Rehaul the doc
fixed by https://github.com/appsmithorg/appsmith-docs/pull/1324
|
2025-04-01T06:37:56.479741
| 2023-12-29T06:36:46
|
2059122181
|
{
"authors": [
"CLAassistant",
"DevDengChao"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3789",
"repo": "appsmithorg/appsmith-docs",
"url": "https://github.com/appsmithorg/appsmith-docs/pull/2016"
}
|
gharchive/pull-request
|
Update disable-user-signup.md
Checklist
I have:
[x] run the content through Grammarly
[ ] linked to sample apps when relevant
[ ] added the meta description for each page in the PR
[x] minimized the callouts and added only when necessary
[ ] added the queryString parameter to the Tabs (if used)
[ ] masked PII in images. For example, login credentials, account details, and more
[ ] added images only when necessary
[ ] deleted the images that are no longer used for the updated pages in the PR
[ ] followed the image file naming convention while renaming or adding new images. (Use lowercase letters, dashes between words, and be as descriptive as possible)
[ ] used the <figure/> tag instead of a markdown representation for images
[ ] added the <figcaption/> tag to add a caption to the image
[ ] added the alt attribute in the <img/> tag
[ ] verified and updated the cross-references or created redirect rules for the changed or removed content
[x] reviewed and applied the style changes for UI formatting. For example, Bold the UI elements(Buttons on screen) used in the doc.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T06:37:56.482577
| 2022-03-22T06:03:56
|
1176306880
|
{
"authors": [
"Parthvi12",
"sharat87"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3790",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/12063"
}
|
gharchive/issue
|
Test-event-driver docker command
Is there an existing issue for this?
[X] I have searched the existing issues
SubTasks
Docker commands for ted:
docker run --name appsmithted -d -p 2222:22 -p 5001:5001 -p 3306:3306 -p 28017:27017 -p 5432:5432 -p 25:25 -v `pwd`/git-server/keys:/git-server/keys -v `pwd`/git-server/repos:/git-server/repo --pull always appsmith/test-event-driver
Here's the repository link : https://github.com/appsmithorg/TestEventDriver
The issue here is that connecting to MySQL from the TED image doesn't work, and throws an SSL error sporadically. Disabling SSL also doesn't help.
Currently on-hold since MySQL doesn't seem to be coming up in the latest TED image which @Parthvi12 is looking into. Once that is running, I'll investigate the original issue again.
|
2025-04-01T06:37:56.487274
| 2023-05-25T07:21:11
|
1725239694
|
{
"authors": [
"Nikhil-Nandagopal",
"aritraroy",
"marks0351",
"narhzih"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3791",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/23719"
}
|
gharchive/issue
|
[Bug]: Auto-layout Jank issues
Is there an existing issue for this?
[X] I have searched the existing issues
Description
I have a few radio buttons in my app, and based on which radio button is selected a different input widget is shown - https://share.cleanshot.com/ZyXJsn86
If you see the GIF carefully, you will see that the input widget and date picker widget shifts a bit whenever the radio button selection is changed. This jank makes the UI look a bit bad. The logic to toggle between the input widget and date picker is very simple and is working perfectly.
I can confirm that this jank was not there before using auto layout. But after I started using auto layout, I am seeing this issue. Hopefully, this UI issue can be fixed.
Steps To Reproduce
Try a similar setup as described in the GIF here,
https://share.cleanshot.com/ZyXJsn86
Public Sample App
No response
Environment
Production
Issue video log
No response
Version
Cloud & Self - Hosted > 1.9.19
@narhzih Request you to correct the app version to Appsmith Community v<IP_ADDRESS>
@aswathkk I believe this is happening coz of row gap issue. Can you confirm?
Closing as we have a new layout builder
|
2025-04-01T06:37:56.490364
| 2023-11-10T09:43:41
|
1987271759
|
{
"authors": [
"ankitsrivas14",
"deepikaappsmith"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3792",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/28788"
}
|
gharchive/issue
|
[Task]: FE changes for Homepage experience - Search bar
Is there an existing issue for this?
[X] I have searched the existing issues
SubTasks
[ ] 1. Search bar UI
[ ] 2. API integration
[ ] 3. Real-time filtering to server side search with debounce
[ ] 4. Cypress
Reduced 0.5sp from this ticket as we are tracking cypress changes in this ticket https://github.com/appsmithorg/appsmith/issues/28980
fixed and merged in release. Hence closing the ticket
|
2025-04-01T06:37:56.491642
| 2024-02-05T09:09:54
|
2118025381
|
{
"authors": [
"ichik",
"jsartisan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3793",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/30896"
}
|
gharchive/issue
|
bug: adjust WDS caption text style
Currently it's bigger than body text, should be smaller than that (close or identical to footnote if we have range between that and body).
@ichik Is this issue ready to pick up? If yes, can you tell me the value we need for caption? I'll update it.
@jsartisan discussed on a call: just one (or two) steps smaller than body text.
|
2025-04-01T06:37:56.493537
| 2021-09-15T05:55:15
|
996687527
|
{
"authors": [
"SatishGandham",
"ayushpahwa",
"somangshu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3794",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/7461"
}
|
gharchive/issue
|
[Task] Move the dataTree diff calculations back to worker
Diff calculation takes about 100ms and is directly proportional to data, so could be very high in some cases.
This was moved to main thread from worker to unblock a user. This was misdiagnosed as the diffs on workers being the reason.
Moving the diffs to main thread does effectively the same thing, there was no issue reported related to it in the last 30days.
@somangshu The data tree diff is taking 10-25 ms on the main thread (checked on multiple apps, depending on the size of the data tree). Moving this to worker thread should bring down time by the same.
Moving this to the icebox and out of the sprint, We do not plan to work on this right now.
|
2025-04-01T06:37:56.497122
| 2023-10-25T19:07:58
|
1962052821
|
{
"authors": [
"Aishwarya-U-R"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3795",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/28379"
}
|
gharchive/pull-request
|
test: Cypress | SetWidget property validations + CI Stabilize
Description
This PR added validations for SetWidget properties
Flaky fixes below:
/Widgets/Others/IconButton_2_spec.ts
Type of change
Script fix (non-breaking change which fixes an issue)
Testing
How Has This Been Tested?
[X] Cypress CI runs
Checklist:
QA activity:
[X] Added Test Plan Approved label after Cypress tests were reviewed
/ci-test-limit runId=6641552552
/ci-test-limit runId=6641552552
/ci-test-limit runId=6641552552
|
2025-04-01T06:37:56.506106
| 2023-11-02T08:03:00
|
1973710322
|
{
"authors": [
"sumitsum"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3796",
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/28574"
}
|
gharchive/pull-request
|
chore: add analytics event for query failure post schema fetch pass
Description
Add event to capture scenario where the schema fetch succeeds but data fetch fails for schema preview page.
Fix a bug where the error message and error type values were interchanged before the Appsmith plugin exception was returned to client.
Fixes #28535 #28618
Media
Type of change
Bug fix (non-breaking change which fixes an issue)
Chore (housekeeping or task changes that don't impact user perception)
Testing
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Also list any relevant details for your test configuration.
Delete anything that is not relevant
[x] Manual
Not aware how to add automation test for this at the moment.
Test Plan
Add Testsmith test cases links that relate to this PR
Issues raised during DP testing
Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR)
Checklist:
Dev activity
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
[ ] PR is being merged under a feature flag
QA activity:
[ ] Speedbreak features have been covered
[ ] Test plan covers all impacted features and areas of interest
[ ] Test plan has been peer reviewed by project stakeholders and other QA members
[ ] Manually tested functionality on DP
[ ] We had an implementation alignment call with stakeholders post QA Round 2
[ ] Cypress test cases have been added and approved by SDET/manual QA
[ ] Added Test Plan Approved label after Cypress tests were reviewed
[ ] Added Test Plan Approved label after JUnit tests were reviewed
/ok-to-test
/ok-to-test
|
2025-04-01T06:37:56.512396
| 2019-09-19T07:57:02
|
495635089
|
{
"authors": [
"gireeshpunathil",
"kewegner",
"kylegc"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3797",
"repo": "appsody/controller",
"url": "https://github.com/appsody/controller/issues/31"
}
|
gharchive/issue
|
Adopt a volume-mounting Docker syntax that allows colons in path names
Is your feature request related to a problem? Please describe.
Please refer to https://github.com/appsody/appsody/issues/31
Describe the solution you'd like
Please refer to https://github.com/appsody/appsody/issues/31
Describe alternatives you've considered
Please refer to https://github.com/appsody/appsody/issues/31
Additional context
https://github.com/appsody/appsody/pull/338 is raised in appsody, a similar change in controller in this repo would compliment it
cc @kewegner
@gireeshpunathil why would the controller need this? The controller does not run docker commands.
The controller uses the APPSODY_MOUNTS environment variable in the case there are no APPSODY_WATCH_DIRS.
If we change the value of APPSODY_MOUNTS to have something other than ";" as a separator, the logic in the controller would have to change, because it splits on ";". See line 269 of main.go.
Ok, so to clarify, adopting the docker --mount syntax to allow colons does NOT need controller changes.
However if as a result we change the APPSODY_MOUNTS separator, then yes the controller needs to support this.
When this issue is done, we would also want to consider the separator for the APPSODY_WATCH_DIRs and the APPSODY_WATCH_IGNOREDIRS (PATHS). Since those are in a list using ';' as the separator.
We should address https://github.com/appsody/controller/issues/31 at the same time. Also the changes we made for single mount testing might be affected.
I have started to code this up and will work with @rykal-nate prior to Friday.
|
2025-04-01T06:37:56.519189
| 2020-02-14T15:38:50
|
565402269
|
{
"authors": [
"Kamran64",
"claassistantio",
"raguibm"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3798",
"repo": "appsody/stacks",
"url": "https://github.com/appsody/stacks/pull/673"
}
|
gharchive/pull-request
|
Add ARG to allow appsody to pass docker-options for build-arg - WIP DO NOT MERGE
Checklist:
[ ] Read the Code of Conduct and Contributing Guidelines.
[ ] Followed the commit message guidelines.
[ ] Stack adheres to Appsody stack structure.
Modifying an existing stack:
[ ] Updated the stack version in stack.yaml
Contributing a new stack:
Describe how application dependencies are managed:
Explain how Appsody file watcher is utilized:
Describe other Appsody environment variables defined by the stack image:
Describe any limitations and known issues:
Related Issues:
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Ragu Kattinakere seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
@raguibm - Thanks for raising this. Please could you raise this as two separate PR's? (one for microprofile and one for open liberty)
|
2025-04-01T06:37:56.578505
| 2024-04-14T15:01:44
|
2242209936
|
{
"authors": [
"Honzoraptor31415",
"ItzNotABug",
"ernstmul"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3799",
"repo": "appwrite/console",
"url": "https://github.com/appwrite/console/issues/1033"
}
|
gharchive/issue
|
🐛 Bug Report: Project creates two times
👟 Reproduction steps
Go to the console https://cloud.appwrite.io/console
Click "Create project"
Go trough the setup proccess
👍 Expected behavior
It should take me to the new project page right after clicking the button (of course with some delay because of some async stuff).
👎 Actual Behavior
At the end after clicking the "Create" button, nothing happens even after waiting a few seconds, but after I click it the second time, it takes me to the new project page. However now when I go to the console again, I can see two projects named the same, even tho I just wanted created one.
👀 Have you spent some time to check if this issue has been raised before?
[X] I checked and didn't find similar issue
🏢 Have you read the Code of Conduct?
[x] I have read the Code of Conduct
Also I apologise for accidentally creating the issue in the website repo, I had to move it here
Can reproduce this, probably a state management issue which re-enables the Create button before the setup is completed.
I proceeded to click the Create three times and 3 projects were created. Also note, if you provide a Project ID then the multiple projects will not be created and an error will be thrown.
Can reproduce this, probably a state management issue which re-enables the Create button before the setup is completed.
I proceeded to click the Create three times and 3 projects were created. Also note, if you provide a Project ID then the multiple projects will not be created and an error will be thrown.
Probable source -
https://github.com/appwrite/console/blob/737457c1446318c4f39fcc373024537af1bc09b6/src/lib/layout/wizard.svelte#L103-L105
Does the setTimeout() function have any purpose? Because I feel like this part of code doesn't make sense with the timeout there.
This is fixed in a recent release
|
2025-04-01T06:37:56.592183
| 2020-03-06T13:28:46
|
576931848
|
{
"authors": [
"LennartCode",
"aprilandjan",
"bomrafinha",
"galer7",
"sz-p"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3800",
"repo": "aprilandjan/vscode-ascii-tree-generator",
"url": "https://github.com/aprilandjan/vscode-ascii-tree-generator/issues/9"
}
|
gharchive/issue
|
Ignore file or directory
(Translated) It would be interesting if there was a way to ignore files or directories within a main directory. Something like the git .gitignore file. Example: Ignore the project's .git directory.
(Original) Seria interessante se houvesse uma maneira de ignorar arquivos ou diretórios dentro de um diretório principal. Algo como o arquivo .gitignore do git. Exemplo: Ignorar o diretório .git do projeto.
@bomrafinha thanks for your advice! Currently this extension does not ignore any thing. However, I did considered to add configuration for users to specify ignore patterns, for example, node_modules and .git folders, etc. I'll do that as soon as I got some spare time 😺
emmm any news here?
I see the code load directoryIgnore in vscode.getConfiguration from config.ts
Could we load directoryIgnore from .gitignore default ?
should a .vscode/settings.json that has:
{
"asciiTreeGenerator.directoryIgnore": [
".git",
"node_modules",
"venv",
"__pycache__",
".vscode",
".idea",
".next",
".nuxt",
"dist",
"build",
"coverage"
]
}
for for line
directoryIgnore: vscode.workspace.getConfiguration().get<string[]>('asciiTreeGenerator.directoryIgnore'),
?
As this repository seems unmaintained, you might want to try another extension that has the feature you are asking for. It supports own ignore files, as well as using the .gitignore.
As this repository seems unmaintained, you might want to try another extension that has the feature you are asking for. It supports own ignore files, as well as using the .gitignore.
BTW:
Look this pr: https://github.com/aprilandjan/vscode-ascii-tree-generator/pull/20
It is have another future like fileTreeExtractor.showFileComment.
Maybe you can pick it to your repo.
|
2025-04-01T06:37:56.594828
| 2023-04-24T14:27:59
|
1681408537
|
{
"authors": [
"turner-aptible"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3801",
"repo": "aptible/docker-alpine",
"url": "https://github.com/aptible/docker-alpine/pull/22"
}
|
gharchive/pull-request
|
Testing Wolfi base image
Doing a test with Wolfi Alpine base.
Both curl and git cloning is failing due network related issues? Works fine locally, but in travis it is failing.
|
2025-04-01T06:37:56.598216
| 2023-04-24T07:16:45
|
1680611386
|
{
"authors": [
"UserNotFound",
"mkielar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3802",
"repo": "aptible/supercronic",
"url": "https://github.com/aptible/supercronic/issues/126"
}
|
gharchive/issue
|
Vulnerabilities reported against "golang.org/x/sys/unix".
Context
In one of our projects we're using pact-broker docker image, which uses supersonic. We're re-packaging the pact-broker image to install some additional tooling (security / monitoring / etc) and harden the image. Then we scan that image with snyk.io.
The problem
snyk.io reports vulnerability in v0.0.0-20220615213510-4f61da869c0c version of golang.org/x/sys/unix which is a transitive vulnerability of supersonic. The detailed snyk report looks like this:
✗ Medium severity vulnerability found in golang.org/x/sys/unix
Description: Incorrect Privilege Assignment
Info: https://security.snyk.io/vuln/SNYK-GOLANG-GOLANGORGXSYSUNIX-3310442
Introduced through<EMAIL_ADDRESS> From<EMAIL_ADDRESS> Fixed in: 0.1.0
Organization: [REDACTED]
Package manager: gomodules
Target file: /usr/local/bin/supercronic-linux-amd64
Project name: github.com/aptible/supercronic
Docker image: [REDACTED].pact-broker
Licenses: enabled
Proposed solution
It seems that upgrading golang.org/x/sys/unix to 0.1.0 fixes the vulnerability.
v0.2.24
|
2025-04-01T06:37:56.631727
| 2024-11-26T05:23:16
|
2693156919
|
{
"authors": [
"CLAassistant",
"tom1299"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3803",
"repo": "aquasecurity/trivy-operator",
"url": "https://github.com/aquasecurity/trivy-operator/pull/2332"
}
|
gharchive/pull-request
|
docs: Fix mage build command
Description
Replace mage build:all with mage build:binary.
Checklist
[ x ] I've read the guidelines for contributing to this repository.
[ ] I've added tests that prove my fix is effective or that my feature works.
[ x ] I've updated the documentation with the relevant information (if needed).
[ ] I've added usage information (if the PR introduces new options)
[ ] I've included a "before" and "after" example to the description (if the PR is a user interface change).
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T06:37:56.636456
| 2024-04-14T08:49:15
|
2242049828
|
{
"authors": [
"CLAassistant",
"naortalmor1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3804",
"repo": "aquasecurity/trivy-plugin-aqua",
"url": "https://github.com/aquasecurity/trivy-plugin-aqua/pull/780"
}
|
gharchive/pull-request
|
Update Plugin Artifacts Links for v0.166.0
Automated changes by create-pull-request GitHub action
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T06:37:56.692375
| 2015-11-16T07:33:50
|
117068682
|
{
"authors": [
"dbuxton",
"evgenyidf"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3805",
"repo": "arachnys/cabot",
"url": "https://github.com/arachnys/cabot/issues/287"
}
|
gharchive/issue
|
Sending alerts when graphite server is not availible
Cabot sending tons of alerts when graphite server is unavailable or authentication has been failed.
It should send one information email and stop alerting per each check..
I agree that this can be annoying but it's difficult to know how big an issue no information is for you.
Do you have a proposal of what sort of notification you'd like to get? What about frequency? What if there's a general outage that kills your graphite machine as well as your whole infrastructure? Wouldn't you want to be notified?
Genuine question - for us this would be a critical issue and so we don't mind too many alerts (especially with the acknowledgement mechanism that recently landed) but would like to establish parameters around this.
I think that such alert emails should be sent to "administrator" but not to every single user which run graphite check ..
I think if the services Cabot is monitoring are mission-critical then a failure to be able to check if the services are working should itself be considered a critical failure... What you are describing wouldn't work for us, for example.
If you were to submit a PR for some ability to configure this I'd be happy to review and in theory to merge, but I don't think we should change this globally.
Closing this. The "official" workaround is to set up an HTTP check for graphite being up, and base your reaction on that. If you have a very specific solution I'd be happy to reopen.
|
2025-04-01T06:37:56.700087
| 2019-05-17T08:31:33
|
445326317
|
{
"authors": [
"araffin",
"tatsubori"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3806",
"repo": "araffin/rl-baselines-zoo",
"url": "https://github.com/araffin/rl-baselines-zoo/pull/22"
}
|
gharchive/pull-request
|
Enabled MiniGrid gyms.
Congratulations of this very very handy work making everyone happy. :->
Just want to start discussing this quick hack enabling to use MiniGrid gyms.
https://github.com/maximecb/gym-minigrid
but not limited to this.
Instead of hard-coding imports, I have provided a command-line option to dynamically import specified gym packages for train.py and enjoy.py.
Adaptation of observation space (Dict) is hard-coded.
I am not sure if this breaks others, while I tried to be careful not to.
Cheers,
Mich
Hello,
Thank you for your interest.
I like the idea. Please don't forget to update the readme with the new instructions ;)
The CI is failing because the docker image does not have minigrid installed... please add an exception in the test script so your trained models are not tested.
The CI is failing because the docker image does not have minigrid installed... please add an exception in the test script so your trained models are not tested.
Added --gym-packages gym_minigrid if "-MiniGrid-" in trained_model.
pip install gym-minigrid for docker scripts (cpu & gpu) as well.
Good, I'll test that on my local machine before pushing a new docker image (so the CI can pass)
I got qt segmentation fault on my machine trying to render the scene :/
Anyway, it seems ok otherwise.
I would also use the flat obs wrapper from minigrid: https://github.com/maximecb/gym-minigrid/blob/master/gym_minigrid/wrappers.py#L167
Better to exclude the cases from the test case for now?
Better to exclude the cases from the test case for now?
yep! (that was my first remark ;))
@tatsubori is that ok if I merge now and in another PR, you'll include a better support for minigrid? (using FlatObsWrapper for instance)
Yes, please proceed to merge. I have CNN impl with it at hand but it won’t converge. I am trying CNN+Lstm but it may take time.
Will continue enhancing it.
StableBaselines to support (not to fail) Dict is better to handle mission text, I believe, though.
StableBaselines to support (not to fail)
I agree, but the support is far from trivial ;) (however we would appreciate PR to solve that issue)
|
2025-04-01T06:37:56.705256
| 2018-12-31T16:18:39
|
395001865
|
{
"authors": [
"bingen"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3807",
"repo": "aragon/hack",
"url": "https://github.com/aragon/hack/pull/73"
}
|
gharchive/pull-request
|
Add instructions to skip initialization on aragon dao install
If install command doesn't find initialization method, it won't
initialize the newly installed app, but this is not obvious from the
help instructions.
Although it's "dangerous", sometimes it's useful to install an app
without initializing. For instance, Token Manager: to initialize it
the token must be provided, and this token must have this new Token
Manager app as controller. Therefore token controller must be changed
in between creation and initialization of Token Manager app.
See https://github.com/aragon/aragon-cli/pull/312
I personally think it's fine for such a small PR like these one (and the one in aragon-cli) which doesn't even touch code. When I add 3 reviewers I'm not necessarily meaning "all of them" ("AND"), in this case it was more like "any of them" ("OR"). But not sure what the correct procedure is.
|
2025-04-01T06:37:56.727356
| 2019-09-26T08:23:23
|
498734467
|
{
"authors": [
"Darkheir",
"aveiros",
"gnusi",
"iurii-i-popov",
"maxkernbach"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3808",
"repo": "arangodb/arangodb",
"url": "https://github.com/arangodb/arangodb/issues/10090"
}
|
gharchive/issue
|
Arangosearch: PHRASE function combined with other filter not working
My Environment
ArangoDB Version: 3.5.0
Storage Engine: RocksDB
Deployment Mode: Single Server
Deployment Strategy: anual Start in Docker
Infrastructure: local
Total RAM in your machine: 16Gb
Disks in use: SSD
Used Package: Docker - official Docker library
Component, Query & Data
Affected feature: ArangoSearch
Search view:
{
"links": {
"my_collection": {
"analyzers": [
"identity"
],
"fields": {
"description": {
"analyzers": [
"text_en",
"identity"
]
},
"type": {},
"name": {
"analyzers": [
"text_en",
"identity"
]
}
},
"includeAllFields": false,
"storeValues": "none",
"trackListPositions": false
}
}
}
AQL query (if applicable):
FOR item IN my_view
SEARCH PHRASE(item.name, "foo bar", 'text_en') AND item.type == "my_type"
RETURN item
AQL explain (if applicable):
Execution plan:
Id NodeType Est. Comment
1 SingletonNode 1 * ROOT
2 EnumerateViewNode 1 - FOR item IN my_view SEARCH ((item.`type` == "my_type") && PHRASE(item.`name`, "foo bar", "text_en")) /* view query */
3 ReturnNode 1 - RETURN item
Indexes used:
none
Optimization rules applied:
Id RuleName
1 handle-arangosearch-views
Steps to reproduce
The query above returns no results and should return one.
The following queries all return the expected result:
Only one word in PHRASE search:
FOR item IN my_view
SEARCH PHRASE(item.name, "foo", 'text_en') AND item.type == "my_type"
RETURN item
No filter on type:
FOR item IN my_view
SEARCH PHRASE(item.name, "foo bar", 'text_en')
RETURN item
Using TOKEN instead of PHRASE:
FOR item IN objects_search_view
SEARCH ANALYZER(item.name in TOKENS("foo bar", 'text_en'), "text_en") AND item.type == "my_type"
RETURN item
Hi @darkheir,
Thank you for submitting the issue. Could you please share the dataset you use so we can reproduce the problem on our side?
I managed to reproduce the issue with this small set:
[
{
"_key":"identity--357447d7-9229-4ce1-b7fa-f1b83587048e",
"_id":"my_collection/identity--357447d7-9229-4ce1-b7fa-f1b83587048e",
"_rev":"_ZUy9LSS---",
"name":"SEKOIA",
"type":"identity"
},
{
"_key":"identity--1621d4d4-b67d-41e3-9670-f01faf20d111",
"_id":"my_collection/identity--1621d4d4-b67d-41e3-9670-f01faf20d111",
"_rev":"_ZUy9LSS--A",
"name":"Adversary Bravo",
"type":"identity"
}
]
I notice that the _rev fields look very similar. Could it be the issue ?
If you insert into the collection using this order, you get 0 results
doc1 => { "type": "person", "text": "foo" }
doc2 => { "type": "person", "text": "foo bar" }
Reversing insertion results, you get 1 result.
I agree with you @aveiros If I insert the documents in the opposite order then the search is working fine.
@Darkheir @aveiros thank you for your effort to reproducing this, we'll address the issue soon. I will keep you updated.
The fix will be available in the next minor release.
tested against docker 3.5 preview build and looks good :)
Hi @Darkheir and @aveiros,
Version 3.5.3, which includes a fix for this issue, has been released.
From the Changelog:
Fixed issue #10090: fix repeatable seek to the same document in
SEARCH operations for ArangoSearch views.
Please find the latest release in the download section.
Thank you for the release! We installed it and it fixes our issue :-)
|
2025-04-01T06:37:56.734887
| 2017-04-11T03:13:18
|
220827499
|
{
"authors": [
"ansarizafar",
"dothebart",
"hkernbach"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3809",
"repo": "arangodb/arangodb",
"url": "https://github.com/arangodb/arangodb/issues/2420"
}
|
gharchive/issue
|
Admin UI design/Color palette and Typography
ArangoDB is a great idea. Market for NoSql database with relation/transaction support is wide open. ArangoDb has a potential to replace Mongodb as a mainstream database. I would like to suggest CSS design refresh for admin ui to give it more professional look. Color palette and typography also need improvements. Please consider using material design guidelines or check rethinkDB admin ui for a reference.
@hkernbach Semantic-Ui can also be used to create great looking ui. SUI also available for react.
Thanks for your comment. Optimizing look & feel and also updating the underlying JS technologies is on our roadmap. But currently it is not a top priority feature.
I would like to suggest CSS design refresh for admin ui to give it more professional look. Color palette and typography also need improvements.
Do you have examples for 'color palettes' and/or 'typography' ? What exactly do you dislike?
Please check Rethinkdb admin ui, Semantic-Ui and Material design. Are you using any front-end framework? VueJs is my personal favorite nowadays.
On 04/18/17 19:08, Zafar Ansari wrote:
Please check Rethinkdb admin ui https://www.rethinkdb.com/docs/,
Semantic-Ui https://semantic-ui.com and Material design
https://getmdl.io/components/index.html. Are you using any front-end
framework? VueJs is my personal favorite nowadays.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/arangodb/arangodb/issues/2420#issuecomment-294913204,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AEx-Vbwtfer3xyQj30aQcjHLz2-J1igJks5rxO4PgaJpZM4M5kUt.
Hallo Claudius,
wie moechtest du denn mit dem Herren in diesem Punkt umgehen? Vielleicht
nimmst du dich dem besser an.
Gruss,
Willi
|
2025-04-01T06:37:56.741980
| 2020-10-09T09:30:49
|
717998818
|
{
"authors": [
"KVS85",
"jsteemann"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3810",
"repo": "arangodb/arangodb",
"url": "https://github.com/arangodb/arangodb/pull/12813"
}
|
gharchive/pull-request
|
always show documents for a collection in web interface
Scope & Purpose
Always show documents for a collection in web interface, regardless of what collection.count() says. This fixes a potential race between fetching the count and fetching the documents.
[x] :hankey: Bugfix
[ ] :pizza: New feature
[ ] :hammer: Refactoring
[ ] :book: CHANGELOG entry made
[x] :muscle: The behavior in this PR was manually tested
[ ] :computer: The behavior change can be verified via automatic tests
Backports:
[ ] No backports required
[ ] Backports required for: (Please specify versions)
Testing & Verification
[x] This change is a trivial rework / code cleanup without any test coverage.
http://<IP_ADDRESS>:8080/view/PR/job/arangodb-matrix-pr/12219/
Tests blue.
|
2025-04-01T06:37:56.747494
| 2023-10-18T17:08:05
|
1950250126
|
{
"authors": [
"Tails32",
"aratakileo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:3811",
"repo": "aratakileo/quick-harvest-plugin",
"url": "https://github.com/aratakileo/quick-harvest-plugin/issues/1"
}
|
gharchive/issue
|
Add Option to Break Harvest Block
Hi there! I hope it's fine to request a feature here.
I would like to have an option to have the plugin break the harvest block and drop the items to the ground in order to collect them instead of putting them straight into the inventory. It is very convenient the way the plugin works right now but it is feeling far less rewarding.
I don't think this option would defeat the purpose of the plugin since automatic replanting itself and being able to hold down right click is still a breeze for harvesting large fields. It should probably not be the default configuration though as most people probably like the current behavior best.
Hi there! I think I can fulfill your request 👌
Thank you so much for adding the option I've requested! Also it's amazing how quickly you acted on it. Thank you.
I've had a chance to test it out and I think it adds exactly what I've been missing in the original implementation. I have a few other suggestions to further improve upon it though. I hope it's fine to add my thoughts here:
Add Block Breaking Effect
I think it needs a bit more work to appear just like vanilla's block breaking. Firstly the particle animation is missing entirely and secondly the configured sound from the configuration still plays. Instead the block's breaking sound should play.
In order to help a bit with this change and not just ask I've researched a bit and I think this fuction adds at least the particles, maybe even the correct sound.
https://hub.spigotmc.org/javadocs/spigot/org/bukkit/World.html#playEffect(org.bukkit.Location,org.bukkit.Effect,T)
https://hub.spigotmc.org/javadocs/spigot/org/bukkit/Effect.html#STEP_SOUND
The block's Material can be accessed by calling getType on the Block and the Effect should probably be STEP_SOUND judging by the Enum's descriptions. The effect should be called on the world to make it visible to all players. This would override the custom sound which then probably only applies to dispensers. But since this might have the potential to cause some lag it would be best to make it configurable. Maybe by default the effect should be played only for the player themselves and optionally it's called on the world?
These are just some thoughts and possible implementation ideas. I've never written a server plugin before but seeing your commits helped me understand it all a bit better. I hope this is a help. Thank you again for your work!
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.