Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
First you should follow the instructions from the GitHub help to create a fork. After you do this, you will have a local repo and a personal clone (or fork) on the GitHub servers. The local repo has two "remotes". origin points to your personal GitHub repo. upstream points to the original main repo. Now when you want ...
I have a repository on git, which I am cloning from into a server that compiles and runs it. Since that server is local on my network, I can further clone the project that is accessible via samba or even open it on my preferable editor and work on it from there, doing the changes I need or anything else. Here is where...
Update modified git repo?
-2TLS and SSL are the same thing. To explain, TLS is a higher-level protocol of SSL that has more security. Anyways, the situation you are having is because the browser trusts your cert and is expecting it to be valid. You should get your new certificate installed in the browser and then change the config of your serve...
I'm using SSL Let's Encrypt. It is revoked 2 days ago. I've generated a new Lets Encrypt SSL. When I do a cURL, the page returns successfully. When I access the website via browsers, it returns a NET::ERR_CERT_REVOKED error. How can I fix this problem?
SSL reissued but still having NET::ERR_CERT_REVOKED
Because that's not how you apply a diff stored in a gist. The gist itself is a repo, with each file in the gist as a separate file. The only way what you tried would work is if the repo is actually a fork of yours.Instead you just want to download the diff file from the gist and pass it togit apply.
Somebody provided me with a Gist on Github that includes a small fix I want to include. On my machine, which contains a clone of the Git repository where I want to apply the change to, I did the following:git remote add fix [gisturl] git merge fix/masterNow I have an additional file "gistfile1.diff" in my local repo b...
Why is my Gist not getting merged but creates a gistfile1.diff instead?
1 I think what you are seeing is due to interleave option for memory allocation under NUMA. There's a very good paper which talks about NUMA under Linux. Without the details of the hardware that you are using it's very difficult to guess how the memory interleaving option c...
I found that physical memory is split into ranks as follows (Memory Interleaving): rank0: [0-512KB] [2048KB-2560KB] [4096KB-4608KB] ... rank1: [512KB-1024KB] [2560KB-3072KB] [4608KB-5120KB] ... rank2: [1024KB-1536KB] [3072KB-3584KB] [5120KB-... rank3: [1536KB-2048KB] [3584KB-4096KB] ... Linux kernel is g...
How to use contiguous memory in linux kernel?
So you used a different email address as user.email on the original commits?If yes, then I think you can fix that by associating that email address with your GitHub account (seehereon how to do that).
Recently I noticed that all my commits for a git repository in gitHub are showing a wrong name (not my account user name, which normally a hyperlink to my profile)So there will be ablue question markwhich has a tooltip saying"Invalid author email. If this commit was made by you, check your local git email setting."So I...
User name of the GitHub account is not displayed as the author of the commits
Apparently it's because THP are swappable, and the Linux swap subsystem can only handle base page sizes and not HugePages. This creates a race condition which has a significant impact on system performance. This Oracle blog dates from 2014 but still seems to be the final word on the topic.
I'm curious to know the reason why transparent huge pages are not recommended for the Databases like Oracle and MongoDB? How THP affect the performance of any Database?
Why THP (Transparent Huge Pages) are not recommended for Databases like Oracle and MongoDB?
1 A general approach is to avoid reloading the whole table if only one cell is changed. In such case, just get the reference to the UITableViewCell you want to "refresh" and invoke a [self setNeedsDisplay] from the main thread on it to trigger its refresh (will call the dra...
currently I'm struggling with this problem: I got a UITableViewController that displays a tableView with different custom cells. One custom cell displays a number (by a label). If you click on this cell, the navigationController moves to a UIPicker where the user can select the number to be displayes. If the user move...
IPhone custom UITableViewCell Reloading
The only way you can consistently do this is if you are using https. If not you have no way to enforce the browser to not use a cached page. There are the hacks you mentioned about but they are not full proof. If it is really important, use https because each request will force a reload.
Basically all pages on this site I am building cannot be accessed when the user clicks on "Back" (or with key control) in the browser, and the page should expire if one is trying to navigate back in history. I put into Global.asax::Application_BeginRequest Response.Cache.SetCacheability(HttpCacheability.NoCache) ...
ASP.Net: Expiring a page when navigating back
1 You can edit your ~/.bashrc or ~/.bash_profile, add there eval $(dinghy env) and it executes that eval for each new session. Share Improve this answer Follow answered Mar 3, 2017 at 20:34 ...
I installed dinghy, docker on mac os x. If I try this, it works: $dinghy up Starting the dinghy VM... Starting NFS daemon, this will require sudo Password: Waiting for NFS daemon... Mounting NFS /Users/xxxx Starting the FsEvents daemon Starting DNS and HTTP proxy xxxxxxxxxxxxx VM: running NFS: running FSEV: runn...
Have to set up eval $(dinghy env) everytime. Cannot connect to the Docker daemon. Is the docker daemon running on this host?
Changing the Access tier to Azure Archive Storage(if storing data in Blobs) would be your best option. A few notes: The Archive storage tier is only available at the blob level and not at the storage account level. Archive storage is offline and offers the lowest storage costs but also the highest access costs Hot, ...
We are migrating from an on-premises virtual machine to Azure cloud. The virtual machine will eventually be decommissioned and we have many files and folders that we don't want to lose, like old websites and databases, scripts, programs etc. We use an Azure storage account for storing and retrieving images via blob co...
Backup files to Azure Storage
Yes. The K8s API is limited when it comes to this, I believe the abstractions for container image metadata are at lower level and probably left out for a reason. You can always look at theCRI specto see what's supported (note that the doc is out of date so you might have to look at the code).If the end goal is to use ...
I try to use the Kubernetes API to read metadata via annotations from container images. The metadata is applicable to every instance of the respecting image and is needed in order to run any resulting container properly. Following this SOquestionit is not possible to read Docker image labels from the kubernetes API dir...
Kubernetes API custom image metadata
The solution for me was to edit App Repository URL in the OpsWorks stack. Change it from:https://github.com/user_name/repository.gitto[email protected]:user_name/repository.git
I am trying to deploy rails 3 application on using AWS instance. For deployment, I am using opsworks services and also accessing private github repository for deployment. When start the instance, getting following errors.[2015-03-10T04:34:32+00:00] INFO: Running queued delayed notifications before re-raising exception ...
Amazon AWS - Fatal: could not read Username for 'https://github.com': No such device or address
2 Question 1: From my experience the ASP forms authentication would be enough. No reason to send credentials as POST and certainly not GET. You can use that for a change password or account info method. You might want to look into Membership and Roles. Question 2: I would ...
I need to place my app business logic into a WCF service. The service shouldn't be dependent on ASP.NET and there is a lot of data regarding the authenticated user which is frequently used in the business logic hence it's supposed to be cached (probably using a distributed cache). As for authentication - I'm going to ...
asp.net, wcf authentication and caching
Cron always runs in an environment different to what you think :-)I always have my cronjobs set up like:*/10 * * * * ( date ; python /webapps/foo/manage.py fetch_articles ) >>/tmp/fetch.out 2>&1to ensure that there's something logged that I can look at.This will narrow your problem down to either:cron, if the temp file...
thanks for helping me setting my cron jobs, crontab has really been a gold mine for me.Unfortunately I have a problem, and have no idea what so ever what it might be... basically a job does not start while the neighbour jobs do. I'll explainThis is my crontabs job list:*/10 * * * * python /webapps/foo/manage.py fetch_...
Crontab job does not start... ideas?
1 You can try following possible solutions: Update the dtype of the columns : Pandas (by default) try to infer dtypes of the datatype of columns when it creates a dataframe. Certain data types can result in large memory allocation. You can reduce it by updating the dtypes ...
I have a flask app running in a container on EC2. On starting the container, the docker stats gave memory usage close to 48MB. After making the first API call (reading a 2gb file from s3), the usage rises to 5.72GB. Even after completion of the api call, the usage does not go down. On hitting the request, the usage go...
Memory Leak - After every request hit on Flask API running in a container
Using Google Chrome:Openchrome://flagsDisableExperimental Web Platform featuresRelaunch the browserIf this doesn't work, openchrome://flags, but instead tryReset allbutton at the top.Source:https://github.com/orgs/community/discussions/50295
Firefox and Edge do not have this problem.When using Google Chrome browser, I am trying to select code or text. The text is selected, but it is not visible. The usual highlighting of the background is not visible. Most other websites seem fine. GitHub website behaves like this.It appears as if it's a bug of the GitHub ...
Why is text selection not visible in GitHub or other websites using the Google Chrome browser?
You can't call a string variable as a method, but you can try this:steps { script { "${INSTANCE}adaptation"() } }
In my Jenkins pipeline I need to call several methods based on the parameter that I get at the run time. For example, If I give the parameter as "Development", it should call the method "Developmentadaptation" similarly for other parameters as well. Below is the code which i tried where INSTANCE is the parameter for th...
Method call in Jenkins pipeline
I found the issue. I had to deploy kubeless to the Kubernetes cluster I had to do this for that:$ export RELEASE=$(curl -s https://api.github.com/repos/kubeless/kubeless/releases/latest | grep tag_name | cut -d '"' -f 4) $ kubectl create ns kubeless $ kubectl create -f https://github.com/kubeless/kubeless/releases/down...
I am trying to deploy a kubeless function using serverless. I created a kubernetes cluster using minikube and I am trying to follow thislinkfollowing whichI installed serverlesscreated a template kubeless-nodejsinstalled plugins withnpm installand tried to deploy usingserverless deploy -vbut I am getting an error/home/...
Error while deploying kubeless function to kubernetes cluster by using serverless framework
From Generating Presigned URLs: import boto3 import requests from botocore import client # Get the service client. s3 = boto3.client('s3', config=client.Config(signature_version='s3v4')) # Generate the URL to get 'key-name' from 'bucket-name' url = s3.generate_presigned_url( ClientMethod='get_object', Param...
In Boto, I used to generate a signed URL using the below function. import boto conn = boto.connect_s3() bucket = conn.get_bucket(bucket_name, validate=True) key = bucket.get_key(key) signed_url = key.generate_url(expires_in=3600) How do I do the exact same thing in boto3? I searched through boto3 GitHub codebase but ...
Generate Signed URL in S3 using boto3
If you set up your SSH successfully, then use your SSH key ID to clone the repo:git clone ssh://<YOUR_SSH_KEY_ID>@git-codecommit.us-east-2.amazonaws.com/v1/repos/<MY_REPO_NAME>Hope that helps.
Below public key (modified) when I upload to IAM changes to a different format:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDZeqQA7Uc+dsdfasWqqqqqqqqqqqqqq6hvmFLZHfoVsebiGkW5t0g6wCaaaaaaaaaaaaaauD0rsW/IoWsYC95Ax5Kdo5N1yGkIs2m1l6DP0JCx1FuZN6DPp4HcccccVmfA3s88BTC84890qdfffffffffffffffff1/0HSqJyBTGrplZcFSesX7cVUGRdiOEF1iBW743 xx...
unable to ssh to AWS codecommit repository, public key contents change
Perl's autoincrement will work on letters as well as digits, in exactly the manner you describe We may as well tidy your nginx-cache increment as well while we're at it I assume SITENAME holds the name of the file to be modified? It would look like this. I have to assign the capture $1 to an ordinary variable $n to in...
I have a string that contains fastcgi_cache_path /var/run/nginx-cache15 levels=1:2 keys_zone=MYSITEP:100m inactive=60m; One of the goals of this script is to increment nginx-cache two digits based on the value find on previous file. For doing that I used this code: # Replace cache_path PREV=$(ls -t /etc/nginx/sites-...
Replace and increment letters and numbers with awk or sed
1 I added semicolons to the ends of the lines in your example. The mysql database now builds and runs for me. CREATE USER 'slave'@'%' IDENTIFIED BY 'password'; GRANT REPLICATION SLAVE ON *.* TO 'slave'@'%'; As mentioned in the comment above, the mbind error is unrelated: h...
I am currently working with Docker to create a database replication using MySQL. I need to create the user and grant replication when the Docker container starts. I have looked on the Dockerhub for MySQL and it appears that you can copy files to a directory and then they would get executed in alphabetical order. See h...
Docker | Create new user on fresh instance start
In SQL Server 2008, you can useouter applyto search for the latest history record. For example, this searches for the last rate before the transaction date:select * from TransactionUnderInvestigation tui outer apply ( select top 1 * from RateHistory rh where rh.EmployeeId = tu...
I am attempting to perform an audit compare the actual RATE for the day to the historical RATE that should have been paid on that day.I have created a temp table to get the information needed from history and called this PPDAUDIT. Here is the audit SQLSELECT EMPLOYEE, PAY_SUM_GRP, HOURS, RATE, TR_DATE, DATE, CORRECT_RA...
Compare history effective closest to transaction date
For anything like this you need real information not guesses. Use a profiler (Netbeans and most other IDEs have one built in) and that can tell you exactly where the memory has gone. For what it's worth unless you have something very odd in your setup logging is unlikely to cause the issue. Whatever happened in the exc...
I'm gettingjava.lang.OutOfMemoryError: GC overhead limit exceededin my production server.The only thing which I noticed in the log file is that there are too many full stacked trace exceptions thrown by a method which is not able to find some id (which is part of business logic).I'm usingorg.slf4jfor logging.So my ques...
Can extensive exception logging cause java.lang.OutOfMemoryError: GC overhead limit exceeded
As discussed in the comments, the way to go here is to use a service-resource as this will give you a static DNS within the cluster to reach all the pods that a targeted by that service.The DNS for the service is:the services name if you access it from within the same namespace<my-service-name>.<namespace-name>.svc.clu...
I have a pod running in a statefulset but it needs to know the hostname or address of all pods running in another statefulset to communicate with them. The second statefulset is being created by a separate helm chart. Can the pod work this out dynamically? Can I inject this information into the pod through an env simil...
Is it possible for a pod running in a satrefulset to get the hostname of the all the pod running in different statefulset?
1 grep worker_processes /etc/nginx/nginx.conf worker_processes auto; change worker processes from "1" to "auto" use this in your config ssl_session_cache shared:SSL:10m; ssl_session_timeout 1h; ssl_buffer_size 4k; Share Follow ...
I am getting this error log in /var/log/nginx. When I run the site I get a 502 bad gateway after exactly 30 seconds of loading up. If i remove some (time extensive) lines of code everything works fine. I edited the nginx config file but I dont think thats the problem. I went to: /etc/nginx/sites-available/django and t...
(upstream prematurely closed connection while reading response header from upstream) ..... && 502 bad gateway
0 I didn't find anything related to disabling this behavior in GitHub's help pages, so I think that it may not be possible to disable this. The best you can do is ask their support team to see if this feature exists or, if that doesn't work, ask for it — the support team's...
GitHub seems to be creating releases out of tags whenever I push those tags up to a GitHub repository that I've cloned to my own machine. How do I prevent GitHub from doing this? If I wanted a release, I'd go into GitHub and create one from a pre-existing tag, but I don't want all of my tags to be associated with re...
Why Does GitHub Create Releases for New Tags When I Push Them to It?
For (1) logically the maximum parallelism would be limited by the least-parallel part of the pipeline which is the 10 LFBs, and this is probably strictly true for demand-load parallelism when prefetching is disabled or can't help. In practice, everything is more complicated once your load is at least partly helped by p...
I am asking this question regarding Haswell Microarchitetcure(Intel Xeon E5-2640-v3 CPU). From the specifications of the CPU and other resources I found out that there are 10 LFBs and Size of the super queue is 16. I have two questions related to LFBs and SuperQueues:1) What will be the maximum degree of memory level p...
What is the semantics for Super Queue and Line Fill buffers?
__weak typeof(self) weakSelf = self;
I see old sample code use this code a lot: __unsafe_unretained __block AssetItem *weakSelf = (AssetItem *)self; I remember that something has changed and this became easier. So is there now a modernized way of doing this?
Is there a modernized way of getting a weak self reference for blocks?
I guess the problem is that there’s an additional]in yourRewriteCond’s pattern and that there is a%missing when referencingHTTP_HOSTinRewriteRule’s substitution. So try this:RewriteCond %{HTTP_HOST} ^([0-9a-z-]+)\.([0-9a-z-]+)$ [NC] RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]And to cover Cags’ concerns a li...
I'm trying to make a general rewrite rule to redirect all domain.com requests to www.domain.com.RewriteCond %{HTTP_HOST} ^([0-9a-z-]+)\.([0-9a-z-]+])$ [NC] RewriteRule ^(.*)$ http://www.{HTTP_HOST}/$1 [R=301,L]The problem is that this rewrite rule doesn't match anything. How can I change it? Thanks
.htaccess RewriteRule non-www to www without explicity specifying domain
Although the problem seems to be solved already (based on the time the question had been asked), I'll add some links for people stumbling across here.Decide, which certificate you want to useUsefoo.com's certificateIf you want to use the same certificate as used for your page, you have to add the IP address to the cert...
I have a websitehttps://www.foo.comand on this site I connect to other machines with a websocket (some random ip address of a machine that has our software running on it). i.e. I connect to a websocket running onws://123.45.67:80The problem is that this site is served over HTTPS and the websocket server that I want to ...
Secure websocket (WSS) certificate signing
5 Kmemleak and Kmemcheck perform different tasks, none is better than the other. 1. Kmemleak checks if some memory blocks were allocated by the kernel but were not freed (that is, checks for memory leaks in the kernel, hence the name). The performance overhead is usually ac...
Is there any special usage/advantage over each other (kmemleak and kmemcheck) ? Can I enable these tools on Android operating system (not Linux OS) please guide me how. Ref: https://www.kernel.org/doc/Documentation/kmemcheck.txt https://www.kernel.org/doc/Documentation/kmemleak.txt
what is the difference between kmemleak and kmemcheck? and How to enable these tools on Android operating systems?
0 This powershell script works... save as .ps1 Function GET-SPLITFILENAME ($FullPathName) { $PIECES=$FullPathName.split(“\”) $NUMBEROFPIECES=$PIECES.Count $FILENAME=$PIECES[$NumberOfPieces-1] $DIRECTORYPATH=$FullPathName.Trim($FILENAME) $baseName = [System.IO.Path]::Get...
here's what I am trying to do. I have a few hundred users My Documents folders in which most(not all) have a file(key.shk for shortkeys program). I need to upgrade the software but doing so makes changes to the original file. I would like to run a batch file on the server to find the files in each My Docs folder an...
Batch file to find a file, create a copy in the same location and run this on multiple directories
None of the answers are explicitly statingwhythe--excludeparameter was seemingly not working.The Why--excludeuses a relative path to the directory being sync'd.For example, the following fails because thesynccommand is syncing the/data/directory, and the--excludeparameter is an absolute path.aws s3 sync /data/ s3://dat...
I am trying to sync data from ec2 to s3 bucket with exclude optionroot@ir:ls /data/ f1 f2 f3 root@ir:aws s3 sync /data/ s3://data/ --profile s3to --exclude "/data/f1/*" root@ir:aws s3 sync /data/ s3://data/ --profile s3to --exclude "/data/f1/" root@ir:aws s3 sync /data/ s3://data/ --profile s3to --exclude "/data/f1*" ...
aws cli s3 sync, exclude not working
, 2 and 3 are correct.Self-healingYou can deploy in master HA configuration. The recommended way is either 3 or 5 master with a quorum of(n + 1)/ 2Can Kubernetes reprovision/restart some the failed nova instances?Not really. That's after nova to manage all the server services. Kubernetes has anOpenStack modulethat allo...
I am trying to understand the relationship between Kubernetes and OpenStack. I am confused around the topic of deploying Kubernetes on OpenStack and doing my research I found there are too many tutorials. My understanding of the sequence is:Start severalnovainstances on OpenStack.Install Kubernetes master on one instan...
Deploy Kubernetes on OpenStack
Your configuration looks correct, and generally I'd say you're right that 1 request should not cause a scale-up. So there are a couple possible reasons I can think of:YourCPU requeston the deployment is set far too low. The HPA's scaling algorithm is based onCPU used / CPU request,notCPU limit, so if you set the CPU re...
I have a question regarding HPA in Kubernetes. Below is my HPA: it says always to keep a minimum replica of 3 and scale up when CPU utilization is over 50%.kubectl autoscale deployment superset -n superset --cpu-percent=50 --min=3 --max=20Though I have three pods up and running constantly, HPA brings up a new pod when ...
Kubernetes HPA bring new pod each time
Yes! The package size (including the layer) does affect theInit Duration"cold-start" timeThis is some benchmarking test on nodejs function with our without webpack (its a tool that helps to reduce the package size for nodejs). And we can see clearly how its effect the cold start init time:The picture was taken from our...
AWS Lambda Runtime: Python 3.7The Lambda Layer attached to my Lambda has a capacity of nearly 200MB when decompressed.I would like to know if this large capacity Lambda Layer will affect the uptime or cold start of Lambda machine.
Does the size of the AWS Lambda Layer affect my Lambda machine's uptime or Cold Start?
5 As-is, your code won't compile -- new returns a pointer, and you're trying to assign that to an object instead of a pointer (perhaps more accustomed to Java?) Assuming you fix that: A *obj = new A(); The entire object that obj points at is allocated on the heap. That in...
class A { public: int i; }; int main() { A *obj = new A(); return 0; } In the following code, when the object for A is created it is created in heap, but where is the memory of i in the obj created ?? Is it created on heap or stack or is there a different behavior??
memory allocation of object attributes in c++
You need to map the port of database to some host port on Server B then from Server A hosting Webapp you can access the db using Server-b-ip:mapped-port. Similarly for the web server you need to map port 80 and/or 443 on the Server A for it to be accessible. See references here
I'm trying to connect two docker containers with different servers. In Server A, I have my web application and for the Server B, My Database. How they can connect? Thank you
Connecting two containers with different servers
You can get anAtom feedof the repository's activity like this (using Node.JSmasteras an example):https://github.com/nodejs/node/commits/master.atomUse any Atom reader likeViennaRSSfor Mac orRSS Feed Readerfor Chrome, or a variety of other feed readers. Many web-based Atom/RSS readers can e-mail you when events occur.Sh...
Beginner here.I've starred a GitHub repository created and managed by someone else. Is there a way to monitor commits and send an email to the listener every time someone checks in to the master branch?Please note that I do not own the repo. I simply want to follow changes on a repository I like.
How to monitor GitHub commits of a third-party repository
You can add common issue tags as follows, note that this works only if you're in an organization account as in personal account you have to create each time (creating an org is free in GitHub)If you're in organization account (remember you must be the owner of that org in GitHub), follow these steps..Go to Settings->Re...
I want to apply my new issue labels to all of my repositories in Github.I can make a new issue label (e.g.important) in a repository. (Let's call this repositoryhello-repo)But I don't know how to useimportantinbye-repo.Do I need to add a new label each time I create new repositories?
How to use the same issue labels in another repository?
In c++ that is not dynamic memory allocation. No delete[] will be needed. Your example is basically a short-cut for this: char c[6]={'h','e','l','l','o','\0'};
Simple question: do I have to delete or delete [] c? Does the language matter? char c[] = "hello"
Is this dynamic memory allocation?
You can use the plain crontab syntax in theschedule.rbevery '0 0 * * 2,5', :roles => [:my_role] do runner "User.notify" end'0 0 * * 2,5'- “At 12:00 on Tuesday and Friday.”It's pretty simple:Minute Hour Day of Month Month Day of Week (0-59) (0-23) (1-31) (1-12 or Jan-Dec) (0-6 or Sun-Sat) ...
I am using whenever for scheduling in my rails app. I could not find how to run my cronjobs on every tuesday and friday of the week anywhere in the gem's documentation.This is what I am doing right now is this correct?every [:tuesday, :friday], at: '12:00am', :roles => [:my_role] do runner "User.notify" end
Rails-Schedule on tuesday and friday using whenever
You could use theprocess-exporterfor this. In this case, the metric you're searching for is the following:# HELP namedprocess_namegroup_oldest_start_time_seconds start time in seconds since 1970/01/01 of oldest process in group # TYPE namedprocess_namegroup_oldest_start_time_seconds gauge namedprocess_namegroup_oldest_...
I am trying to get the uptime i.e. how long a service has been running. I managed to get the service up/down status. I did some mappings in the stat panel for 0 and 1.This is my querynode_systemd_unit_state{instance="x.x.x.x:9100", name="service01.service", state="active"}So that does work but I can't figure out how to...
Trying to get the uptime of a serviced service, I have the state up/down
How easily normals can be computed on the GPU depends on the mesh topology.It is easy to compute normals for a mesh withtriangle-list topology: Use one GPU thread per triangle. This results in highly regular reads and writes and will work for any valid configuration of blocks and threads in CUDA. Unfortunately, triangl...
I am trying to access last and next indices coordinates inside the kernel.ex: int idx = blockIdx.x * blockDim.x + threadIdx.x;then pos[idx].x, pos[idx].y, pos[idx].z would give current coordinates of a point. but cannot access other two. I am trying to calculate the normals of the changing triangle in GPU level using C...
c opengl: can i calculate normals in gpu? (CUDA)
For this kind of info I usually filter out the output from describe command:kubectl describe pod myFailingPod | grep -E "(Image:|State:)"
If I have several containers in a pod, and 2/3 are OK, is there a nice way to obtain a summary of which are OK and which are not?
Quick way to summarise container status within a multi-container pod
I faced the same issue and for some weird reason AWS SDK did not recognize the associate array in PHP. Maybe because at the time of JSON encoding associative array in PHP becomes object and SDK expect it as Array. Changing $payload as follows will fix this issue. $payload = array('test1', 'test3', 'test3'); Also reme...
I'm trying to use the aws php sdk to invoke a aws-lambda function and get the return value like so: $client = LambdaClient::factory([ 'key' => 'mykey', 'secret' => 'mysecret', 'region' => 'us-west-2' ]); $payload = [ 'key1' => 'value1', 'key2' => 'value2', '...
Passing a 'Payload' to an AWS Lambda function with the AWS PHP SDK
32 Try the link https://github.com/pulls/review-requested, which shows you all Pull Requests that you've been tagged in as a reviewer. Note that you may have to click the Open or Closed buttons on that page to filter the Pull Requests you wish to see. The equivalent Github...
In github you can add a number of reviewers. What I'm looking for is, as a user, to list all the pull requests that I'm currently tagged as a reviewer in. I have looked at the views that's available in the "pull request" link but this don't meet my demands. For instance if I go to the "Review requested" i get a "No ...
How to see all Pull requests that im tagged as a reviewer
(.*) will match all characters. That's all you need.ShareFollowansweredJan 20, 2011 at 21:38sciritaisciritai3,71811 gold badge1717 silver badges2222 bronze badges1but im getting some troubles with the images on the site, they are not appearing if I put the ruke like: ^([a-z\-_0-9]+)/(.*)$ index.php?engine=$1&q=$2any cl...
What pattern do I need to rewrite all characters in the requested URL? I tried:RewriteRule ^([a-z\-_0-9]+)/([^/.]+)$ index.php?engine=$1&q=$2 [NC,L]But the dots.are not getting recognized.
.htaccess rewrite all characters
It's important that key names in dask graphs are unique (as you found above). Additionally, we'd like identical computations to have the same key so we can avoid computing them multiple times - this isn't necessary for dask to work though, it just provides some opportunities for optimization. In dask's internals we m...
Dask supports defining custom computational graphs as well as opportinistic caching. The question is how can they be used together. For instance, let's define a very simple computational graph, that computes x+1 operation, import dask def compute(x): graph = {'step1': (sum, [x, 1])} return dask.get(graph, 'st...
Opportunistic caching with reusable custom graphs in Dask
The source code of the CodeQL queries is available in the GitHub repository. The documentation also lists the existing queries: CodeQL query help CodeQL CWE coverage However, which queries (or rather query suites) are run as part of GitHub workflows depends on the configuration of the workflow.
I'd like to cross-check the vulnerabilities covered by GitHub's CodeQL service and OWASP Top Ten Web Application Security Risks so that I know where the gaps are. I can't find a list of vulnerabilities covered by CodeQL. Does GitHub publish the list of rules?
Does GitHub publish the CodeQL ruleset?
Incd /var/www/html/Running once to setup:gh repo forkwith a remote set to the originEach time to update:gh repo syncsetting the base repository each time to updateThen usinghttps://git-scm.com/book/en/v2/Customizing-Git-Git-Hooksa server-side post-receive hook to move the classes directory to the desired locationShareF...
Trying to deploy internal development from GitHub to a Centos 7 webserver, running into 2 issues.Firstly, I'm using PHP and need to deploy the class files to a folder that is not/var/www/html/, where I have cloned the repository to usingcd /var/www/gh repo clone linkto/repo html(this may be the wrong thing to do).Secon...
Using github CLI (cli/cli) to deploy repo
Themount.cifsfile is provided by thesamba-clientpackage. This can be installed from the standard CentOS yum repository by running the following command:yum install samba samba-client cifs-utilsOnce installed, you can mount a Windows SMB share on your CentOS server by running the following command:Syntax:mount.cifs //SE...
I am trying to setup a script that will:Connect to a windows shareUsing LOAD DATA LOCAL INFILE, upload the two files into their appropriate db tablesUmount shareSituation:I can currently vpnc into this remote machineProblem:I cannotmount -t cifs //ip.address/share /mnt/point -o username=u,password=p,port=445 mount erro...
mount remote windows share from centos
1 As the setting says this is the maximum buffer size not the buffer size used for all requests. Although maxBufferSize is only used for defining the buffer size for headers when the message is streamed Share Improve this answer Follow ...
I have a Duplex communication which sends Small as well as Large messages. In very rare cases I have more data to be send, because of this reason I am in a confusion to change the TransportMode to Streamed. So I set the buffer size to 1GB (or a Large Size). maxBufferSize="1073741824" Is this cause my Small message co...
Is there any concern in Setting "maxBufferSize" to 1GB of WCF NetTcpBinding
Most of the code in.htaccessshould be the same on local and production. (Otherwise, how do you test it?)However, one of the cleanest ways to separate directives between servers is toDefine(requires Apache 2.4) a variable in the server config of one of the servers (eg. the development machine):Define DEVELOPMENTThis can...
How to create an Htaccess which will have a specific and separate code for Localhost and Production. I mean when we work on localhost, it should be work localhost code and in Production it should be load only Production code. So that i can use one Htaccess for Local and Production and it will save lot of time. Followin...
Htaccess separate for Localhost and Production
<div class="s-prose js-post-body" itemprop="text"> <p>It seems that you have a problem with your rvm gemset not beeing used.</p> <p>First, <code>cd /yourprojectdir</code> and create a <code>.rvmrc</code> and matching gemset for your project</p> <pre><code>rvm --rvmrc --create 1.9.2@yourprojectname </code></pre> <p>Then...
<div class="s-prose js-post-body" itemprop="text"> <p>I am running Ubuntu 10.04 on a linode trying to get sinatra, passenger, nginx, and rvm to all play nice with Ruby 1.9.2. I'm running into an issue with gems though:</p> <p>~/.rvm/gems/ruby-1.9.2-p290/gems$ ls returns </p> <pre><code>daemon_controller-0.2.6 fastthre...
Sinatra Gem Won't Install
It looks like key condition expression does not supportNOT begins_with(x). This might be because the result set is not contiguous (it's the items beforex, aggregated with those afterx).Some possible solutions are:make thegameScoreIda non-key attribute (or replicate it into a new non-key attribute), then you can query o...
I am trying to run a DynamoDB query that says I want items that do not begin with a particular value. I cannot seem to find a way to do this.I have tried the following 4 ways of evaluating, and none of them work. Each one gives me an invalid operator error.MyKeyConditionExpression(s) that I have tried look like this:!b...
DynamoDB: Does not begin with
This is a frequent issue, specially in 32-bit platforms. I suggest to use some kind of fragmented or "chunked" array class, like this one:https://blogs.msdn.microsoft.com/joshwil/2005/08/10/bigarrayt-getting-around-the-2gb-array-size-limit/Of course there will be a loss of performance. However, it depends on your speci...
I am running into the following problem. Using C# (and XNA), I try to allocate an array of moderate size (~55 MB) of type Color[]. Color is 4 bytes. However, 90% percent of the memory allocation tries fail due to an "Out of memory" exception despite the system having 16 GB RAM (~12 GB free).I am already using the Memor...
How to deal with C# memory allocation (Address space fragmentation)
No, it is not possible, you need to redeploy AKSedit: 08.02.2023 - its actually possible to some extent now:https://learn.microsoft.com/en-us/azure/aks/configure-azure-cni-dynamic-ip-allocation#configure-networking-with-dynamic-allocation-of-ips-and-enhanced-subnet-support---azure-cliI'm not sure it can be updated on a...
I'd like to move an instance of Azure Kubernetes Service to another subnet in the same virtual network. Is it possible or the only way to do this is to recreate the AKS instance?
Is it possible to change subnet in Azure AKS deployment?
Adding an additional answer here as I feel the existing one capture the 'what' but not enough of the 'why'. The reason it's best to store individual entries separately in the cache have little to do with perf. Instead, it has to do with allowing the system to perform proper memory management. There is a lot of logic i...
I have a website where I cache a lot of info. I am seeing conflicting information around storing stuff in the asp.net cache. For example lets say I have this data structure: Dictionary<string, List<Car>> mydictionary; I could store the whole thing with a string key as "MyDictionary" and then drill down once i pull ...
What is the most efficient way to store / retrieve items in asp.net httpContext.Cache
-1This issue seemed to resolve itself. The error occured while I was using my laptop at work, but once I got home and tried again it worked fine.ShareFollowansweredDec 8, 2009 at 14:51Michael EdwardsMichael Edwards6,43888 gold badges4545 silver badges7575 bronze badges11@Rob move to a different location and try again–j...
I am trying to clone a project from GitHub but I keep getting this error:github.com[0: 207.97.227.239]: errno=Result too large fatal: unable to connect a socket (Result too large)This is the version of the Git client I am using:git-gui version 0.12.0.23.ga91be git version 1.6.4.msysgit.0Has anyone seen this before?
Git client throwing error
Your assumption is close, except that it is slightly more optimized in practice. Cache reads and writes are performed on the underlying hash table and appended to internal ring buffers. When the buffers reach thresholds then a task is submitted to Caffeine.executor to call Cache.cleanUp. When this maintenance cycle r...
I am using caffeine in the following configuration: Cache<String, String> cache = Caffeine.newBuilder() .executor(newWorkStealingPool(15)) .scheduler(createScheduler()) .expireAfterWrite(10, TimeUnit.SECONDS) .maximumSize(MAXIMUM_CACHE_SIZE) ...
Caffeine combining both scheduler and executor service
The OutOfMemoryError exception is thrown when you don't have enough memory on your heap to create a specific object. So if you were trying to create a massive object that required the maximum amount of space your heap had, you would get the error. That doesn't necessarily mean that there is no memory available though. ...
In Java, the following is allowed:public void handleError() { try { System.out.println("Out of memory error!!"); } catch(OutOfMemoryError outOfMemoryError) { System.out.println( "This is catch block of out of memory error Message:"+ outOfMemoryError.getMessage()); } ...
Why does Java allow catching Error types when applications aren't supposed to handle them?
df1 %>% filter(col1 %in% df2$col2) %>% remove_rownames %>% column_to_rownames('col1')
I am trying to subset/filter a data frame according to the corresponding column elements from another data frame. Here is what I used to do thisdf <- df1[df1$col1 %in% df2$col2,]And then I am going to set the column as row namesdf <- df %>% remove_rownames %>% column_to_rownames('col1')However I have no idea how to com...
How to subset a data frame with R pipeline
This is because the content type is missing so the browser doesn't know that your file should be interpreted as HTML.Please addContentType: 'text/html'in the parameters passed tos3.upload.See also the explanations and links given inUpload Image into S3 bucket using Api Gateway, Lambda funnction
I am looking to allow public users to view HTML files located on an AWS S3 bucket on their browser. These HTML files are created and uploaded to my S3 bucket via node.js, and a URL linking to the file is generated.I am using this method to upload the HTML files:s3.upload({ Bucket: bucket, Key: "HTMLFiles/file....
AWS S3 - Allow public to view HTML files
you can create new file with updated values and use command replace along wih createkubectl create configmap NAME --from-file file.name -o yaml --dry-run | kubectl replace -f -
How to patch "db.password" in the following cm with kustomize?comfigmap:apiVersion: v1 data: dbp.conf: |- { "db_properties": { "db.driver": "com.mysql.jdbc.Driver", "db.password": "123456", "db.user": "root" } } kind: ConfigMap metadata: labels: {} name: dbcm
How to patch configmap in json file with kustomize?
0 No, you do not need to run it with docker-compose. I would recommend creating the following files: Dockerfile - production image build Dockerfile.dev - development image build The Dockerfile.dev would look something like this: FROM node:10-alpine ENV PORT 8080 WORKDIR /...
How can I run NodeJs in Docker and Connect it with MongoDB and RabbitMQ? Do I need to run it in with docker-compose? And also I need 2 environment Run nodejs application in a development environment with nodemon. The Second environment is production experiment node app.js I found this docker-compose yml, but I can n...
How can I run NodeJS in Docker with MongoDB and RabbitMQ?
The GRADLE_USER_HOME environment variable isn't set by default. You'll need to explicitly set it, and then copy over the downloaded dependencies in the next stage. FROM gradle:jdk13 AS cache WORKDIR /app ENV GRADLE_USER_HOME /cache COPY build.gradle gradle.properties settings.gradle ./ RUN gradle --no-daemon build --s...
I have the following Dockerfile FROM gradle:jdk13 AS appbuild WORKDIR "/home/gradle/" COPY --chown=gradle:gradle "./build.gradle" "/home/gradle/" RUN gradle dependencies COPY --chown=gradle:gradle "./src/" "/home/gradle/src/" RUN gradle build --info FROM openjdk:13 ENV LANG en_US.UTF-8 COPY --from=appbuild "/home/gra...
Gradle dependency caching during docker multi stage build?
Using&means 'address of'.Using(int *)means 'treat value as an address'.So in this case, the(int *)treats the value "tiger" as anintpointer. *(sort of)*It is a little more complicated. It's not exactly the value"tiger", since it is casting from aconst char *rather than the string literal itself. Instead,catis a pointe...
I am new to programming and just tinkering on my own to get a better feel for concepts covered in book lessons; however, I cannot find an explanation for the following: In the code below, why are the addresses returned using (INT *), different from those using & ?char animal[20] = "doberman"; const char * cat = "tige...
In C++ what is the difference between (int *) and &?
<div class="s-prose js-post-body" itemprop="text"> <p>You can create a <code>docker volume</code> using EFS:</p> <pre><code>docker volume create \ --driver local \ --opt type=nfs \ --opt o=addr=10.0.0.50,rw,nfsvers=4.1,rsize=1048576,wsize=1048576,hard,timeo=600,retrans=2 \ --opt device=:/ efs </code></...
<div class="s-prose js-post-body" itemprop="text"> <p>I am trying to mount EFS inside a docker container running on EC2 server. EFS mount in EC2 is working fine with,</p> <pre><code>sudo mount -t nfs4 -o nfsvers=4.1,rsize=1048576,wsize=1048576,hard,timeo=600,retrans=2,noresvport &lt;efs-address&gt;:/ efs </code></pre> ...
How to mount EFS inside a docker container?
How to setup your local repo and add it to githubGithub clearly documents the steps but I'll repeat anyways.Create the repo ongithub.com.Copy its link which generally is of the formgithub.com/<user_name>/<repo_name>.gitThis link is also calledremote.Create your local directory.mkdir <my_dir>Name of your local directory...
i am new in git. I created a folder in my Github with an intention to save some of my works from local computer. So i connected my local folder to remote folder, but i didn't use "git push -u origin master" command. Because I feel like the name of remote folder (folder that i created in github) isn't appropriate, so de...
local pc's folder to github folder?
You can skip auditing at all by adding the--no-auditflag.npm install --no-auditIf you want this to apply to devDependencies only, you can run it this way:npm install --no-audit --only=devIf you want this to apply to production dependencies only, you can run it this way:npm install --no-audit --only=prod
Is it possible to configure npm to skip audit of vulnerabilities fordevDependencieswhen running commandnpm install?
NPM configuration to skip vulnerabilities audit for devDependencies on install
Your query is correct for what you described.If you want to only get "last" 10-minutes window where your metric is 0 for 10 minutes, and recent 10 minutes are not critical for you, you can add this condition to the query:and changes((sensor_status offset -10m)[10m:])!=0Or if recent 10 minutes are critical, you may use ...
I have a custom Gaugesensor_statuswith labelsjob, type, id.I want to get the time whensensor_status{job='myjob', type='streaming'} == 0for more than 10 minutes.The result should be data with timestamp when it occurs by each labelid.sensor_status{job='myjob', type='streaming'} == 0 and (changes(sensor_status{job='myjob'...
Prometheus - query data when is equal 0 for more than 10 minutes
Finally got it. Python has-argument after which it reads arguments from stdinhttps://docs.python.org/3/using/cmdline.htmljupyter nbconvert --to python --stdout .\some_nb.ipynb| python - --argument_one=1
The use case is the following:convert jupyter notebook to pythonrun converted notebook on-the fly with additional argumentsWhat I have tried so far:jupyter nbconvert --to python --stdout .\some_nb.ipynb | pythonsome_nb.ipynbawaits for arguments viaargparseso normally I would do something like:python some_nb.py --argume...
Run python code from stdin with additional script arguments in PowerShell
Try this: -(void)appendString{ //stringWithString returns an autorelease object //so you don't need to worry about its memory management NSString *locStr = [NSString stringWithString:@""]; for (int i=0;i<[result count]; i++) { //if your locstr is created by initWithString instead, //the following line is going to ...
I have declared a local string inside the method. I am releasing the string inside the same method. I found my code crashing if release that object. If I don't release the string, code runs successfully. I have called that method in viewdidappear so that method is called while pushing and poping. Nothing gets printed...
How to release a string created inside a method?
Unfortunately, the best answer we currently have is to create a new cluster and move your resources over, then delete the old one.We are very actively working on making cluster upgrades reliable (bothnodesand themaster), but upgrades are unlikely to work for the majority of currently existing clusters.
I began to try Google Container Engine recently. I would you like to upgrade the Kubernetes Cluster to the latest version available, if possible without downtime. Is there any way to do this?
How to update Kubernetes Cluster to the latest version available?
You could test against the "Referer" header, you can't really rely on that but it's the best possible.E.g.http://jsfiddle.net/QmnKR/one of the headers will beReferer: http:// fiddle.jshell.net/QmnKR/show/light/
When creating websites, I like to let my clients view the work in progress. At the moment I do this by uploading their website to a directory, and use .htaccess to password protect that directory. But keeping track of passwords and ensuring the directory is still protected after an update is becoming an issue.I have no...
.htaccess redirect unless in iframe
I ended up getting MAVProxy on host and dronekit-python in the docker flask container properly connected.Seemus790's answer in thisgitter threaddid the trick.Working solution: MAVProxy on host machine (Mac OS in my case)mavproxy.py --master=127.0.0.1:14550 --out udp:127.0.0.1:14551 --out udp:10.55.222.120:14550 --out=t...
I am using dronekit-python in a docker container and am attempting to connect to an instance of MAVProxy running on my host machine (Mac OSX) using the following command:vehicle = connect('udp:host.docker.internal:14551', wait_ready=True)but am getting the following error:File "/usr/local/lib/python3.7/site-packages/py...
Dronekit-python running in docker connecting to MAVProxy on host
0 This issue occurs if PHP is unable to log the session information in the file. Check the session_save_path in the php.ini file. Share Improve this answer Follow answered Jan 11, 2019 at 5:35 ...
My nginx.conf file looks like this: user www-data; worker_processes 4; events { worker_connections 1024; } http { include mime.types; server_tokens off; root /var/www/html; fastcgi_cache_path /tmp/nginx/cache levels=1:2 keys_zone=phpcache:100m inactive=60m; fastcgi_cache_key "$scheme$reques...
Nginx: 502 Bad Gateway
If you want to use nginx as a reverse proxy for express you can configure your server as follow: server { listen 80 default_server; server_name _; location / { proxy_set_header Host $host; proxy_set_header X-Real-IP $remote_addr; proxy_pass http://localhost:3000; #port where you are serving your e...
This question already has an answer here: Node+Express+NGINX application returning localhost instead of domain (1 answer) Closed 3 years ago. I want to run the express server using...
How to run a express server using nginx in localhost? [duplicate]
Containers in Fargate tasks share a network namespace, so you don't need to use links at all. You can simply communicate via localhost. For example, if you have container A running a web server on port 8000, container B could reach it with curl http://localhost:8000/
I try to setup a little example on AWS Fargate and try to have one application container with one database container there. The task definition starts both containers, but my app container fails with getaddrinfo ENOTFOUND db db:3306. Therefore, the database container could not be found by my app container. Since Farg...
Linking Container in AWS Fargate
3 You can allocate the memory in the heap with new. The heap is much larger than the stack. Code below. #include<iostream> #include<bitset> using namespace std; int main(){ bitset<400000000UL>* coord_ptr = new bitset<400000000UL>(); bitset<400000000UL> coord ...
This question already has answers here: Define a large bitset in C++ (4 answers) Closed 9 years ago. I need to use a bitset of size 400000000. Just like this: #include<iostream> ...
Create a large bitset [duplicate]
The lineoptions(tibble.print_min = 5)Sets a user-level option. This means that when a user runs any other code after running your snippet, theirtibble.print_minvalue may be different from the one they may have set previously and which they may prefer or even rely on for certain outputs. This is considered bad practice,...
this is CRAN feedback for package release.I tried to summary the question onr-package-devel. However, I do not quite understand the problem and wish to figure out what this problem means.Anyone know what this mean by ** reset to user's options()**? Which of my part cause this problem?feedbackPlease always make sure to ...
Please always make sure to reset to user's options()
You need use the ENV variable from your heroku app. If you do a heroku config you can have access to all of your ENV variable. You just add some and use it directly in your application. With this trick you don't need update your code to change your configuration and the configuration if not define in your code base. I...
I'm using RoR hosted by Heroku and I'd like to store files on s3 using paperclip. My source code is hosted on github and is world readable. What is the best practice to keep the keys a secret from the rest of the world? Paperclip suggests that the access keys are stored in a configuration file (or in code), so for exa...
Ruby on Rails, Paperclip, Heroku, GitHub and AWS - securing keys
delete does three different things: Runs the destructor of the object (or of all objects in the array in the case of delete[]). Marks the chunk of memory previously used by the object as free. If possible, informs the operating system that a chunk of memory is free for other programs to use. Your question is about #...
I directly started with managed languages and have barely any experience with C++, hence this question might be too basic. In a managed language like .net, GC frees the memory. From what I read, in C++ this is done by calling delete. But what does it do to free memory? Does it it set all the bits at a memory location ...
What does freeing memory mean? Does it mean setting all bits to zeros?
That's not a valid PromQL query (there are no quotes around the expected value ofportid) and[1-10]is not a valid regular expression (which is what=~expects).Try this instead:pkts_sent{portid=~"[1-9]|10"} < 1000
I have a metric (pkts_sent) and it has a couple of parameters (switchname and port). I need to fire an alert based on the condition that if number of pkts_sent is below 1000 but it could be fired only for ports 1 to 10. I am not sure how i could create a range of the ports in the query.pkts_sent {portid=~[1-10]} < 1000...
Range query in prometheus
There is no possibility for now to use PersistentVolumes in Daemon Set.There was a feature request for it onGithubbut unfortunately it was closed. Link to this request:volumeClaimTemplates available for Daemon SetsThere is acommentin the link above which is describing this topic a bit more.
I was curious if it is possible to move completely off hostpaths and use local volumes (persistent volumes) instead. However, there doesn't seem to be a way to include volumeClaimTemplates into a daemonset.Statefulsets provide volumeClaimTemplates, but they require the replicas to be manually defined, as opposed to aut...
Kubernetes one pod per node, one persistent volume per pod
Run each program separately first for a few iterations and checknvidia-smi dmonto see how much memory that program actually requires. Then setconfig.gpu_options.per_process_gpu_memory_fraction = ...in your session configuration based on the memory information you learned fromnvidia-smi dmon. If the memory required for ...
I am getting error when i run 2 tensorflow scripts on single GPU. I have tried growth and GPU memory allocations steps, still the first script executes without problem while 2nd script failes with ResourceExhaustedError , Graph session creation error.Kindly help.
how can i train two tensorflow scripts on single gpu parallelly?
It looks like for your case database is overhead. You just need some distribute lightweight key-value storage with shared key lock support. Here are some candidates: etcd (https://coreos.com/etcd) consul (https://www.consul.io, especially https://www.consul.io/docs/commands/lock.html) redis (http://redis.io)
I have a simple C++ service (API endpoint) that increases a counter every time the API is called. When the caller posts data to http://10.0.0.1/add the counter has to be incremented by 1 and return the value of the counter to the caller. Things get more complicated when the service is getting dockerized. When two inst...
How to atomically update a counter shared between Docker instances
No, You can't. Instead you can use the FileBeat system module which will monitor your OS layer, infohere
Is it possible to check if java application (standard "main function" app, not web app) is running with Heartbeat (https://www.elastic.co/products/beats/heartbeat) ?
Elastic search + Heartbeat - Is it possible to check if java app is running?
There are two things at play here:Thefile system pathTheURL pathIf you're running anAmazon Linuximage, your web content should be deployed inside/var/www/html-- as is the case with just about every reasonable Linux installation.If your index page is stored at/var/www/html/index.php, then your URL will behttp://123.45.6...
I have a web application which is currently working fine on my local machine and I am now trying to get it to work on EC2.I transferred the index.php file into the folder /var/www and I am able to access it by visiting my elastic IP (for example,http://123.45.678.910/).The trouble is that I also added the folder named ...
EC2 web application folder structure
1 Your token is wrong. You can get a worker token in the manager node: docker swarm join-token -q worker It works for me. https://docs.docker.com/engine/reference/commandline/swarm_join/ Share Improve this answer Follow ...
Start docker swarm : docker swarm init --advertise-addr Join docker swarm: docker swarm join --token :2377 I am using Windows 10, it is working fine with Windows container mode, but gives below error in Linux container mode. Error: Error response from daemon: remote CA does not match fingerprint. Expected: 91030413f...
Docker swarm join linux container Error - remote CA does not match fingerprint
This isn't really possible, since the Wiki is statically generated content. There's no option to my mind to add dynamic content in this fashion. This does force the developer to keep well on top of their documentation, however.
I am currently writing a small documentation on Github Wiki. I have a repo for my example codes and would like to cite some lines from examples codes in my wiki. Rather than writing a code block statically, I simply want to "get 1st to 5th lines of code from this file", so when ever I update the an example code, the c...
Get a part of a code from repo to Github Wiki
The main part of deamonizing a program is in disconnecting it from its controlling terminal. To do that, you call setsid(). setsid() requires that the caller is not a process group leader (a process run directly from a shell with job control). If you fork and then continue in the child, the child most definitely will ...
I'm reading Nginx Open Source and I wonder why would someone kill the parent process and let child process handle the rest of the program? Why not just let parent process handle it? Your help is very much appreciated. I use Eclipse CDT to debug the program and this causes my debug come to a dead end since it continues...
What is the purpose of killing parent process and leave child process run after fork()?
This: for i in ans: ans.append(curr+i) You are iterating through ans and adding stuff onto the end of ans at the same time. It is an infinite loop: you'll never get to the end of ans because you keep making it longer (until you run out of memory). If you want to add new elements to ans which are each of the exist...
I wrote a code to get all subsequences of a string using recursion in python. Below is the code. def solve(s): if (len(s)==0): return [""] curr = s[0] res = s[1:] ans = solve(res) for i in ans: ans.append(curr+i) return ans if __name__=="__main__": s=str(input()) print(...
Why is there Memory Error in recursion function in python?
Sure that's possible.According tothe docs, you need a section in your.travis.ymlfordeploy. Specifically, you need to switch it on for when you create a Tag in your repository, and you will need to tell Travis to throw nothing away prematurely. Quoting the docs for thedeploy worldsnippet:deploy: provider: releases ...
So I figured out how gradle etc works. I heard about travis CI and got it working. Now the last thing I want to achieve, is to make gradle create jar files of the source folder and to give the file names according to their build number, so that none of the previous files get overwritten. I heard about the GitHub deploy...
Deploy jar File using Travis CI and Gradle automatic build count
You cannot use theBACKUP DATABASEcommand to backup a single table, unless of course the table in question is allocated to its ownFILEGROUP.What you can do, as you have suggested, is export the table data to a CSV file. Now in order to get the definition of your table you can 'Script out' theCREATE TABLEscript.You can d...
How to take table-level backup (dump) in MS SQL Server 2005/2008?
Table-level backup
SonarQube Community Edition is licensed under the GNU Lesser GPL License v3, as you can read on theLicense page on sonarqube.org.If you want to know what you can legally do with this LGPL v3 license, you can chek the following page:https://tldrlegal.com/license/gnu-lesser-general-public-license-v3-(lgpl-3)
We want to build a commercial product using SonarSource Community Edition which we deployed on our own servers, I am not able to find relevant information regarding its licensing faqs, can we create the product using it and sell it?
Sonarsource Community Edition License
You can use GIT to solve the issue, You need to create a master branch and all for all client websites you can cut a branch from master branch.So if you need to need to make any changes to the core code and if it should get reflected in all the websites, you could make that change in master and then do agit merge maste...
I'm beginner with Git and I'm wondering, for a project, if there is a way to use the master branch as a root and at the same time use branch to customize websites per client.What I have now is a general code wich can be duplicate on differents hosts. Then I have 1 config file with informations like :Client NameLogoEmai...
GIT - Custom design per client
1 Make sure you pass in a "Classic" Personal Access Token. The new fine grained access tokens don't support the GraphQL API yet. Share Improve this answer Follow answered Mar 22, 2023 at 20:09 ...
I am trying to create repo using GitHub CLI on a GitHub Action and it stuck on GraphQL: Resource not accessible by integration (createRepository) Here is the action.yaml name: Create Repo on: push: branches: [ master ] env: GH_TOKEN: ${{ secrets.GITHUB_TOKEN }} TEMPLATE_REPO: template # this is the templ...
GitHub CLI command gh failing to create repository