Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
You skipped step 3 in the Getting Started guide:Create a Service Role.ShareFollowansweredJun 18, 2016 at 6:42Karen BKaren B2,71311 gold badge1818 silver badges1919 bronze badges9Actually it isn't working properly. I was successful in creating another role but I am still receiving the same "Cannot Assume Role" message. ...
I am trying to setup the CodeDeploy service with one of my ec2 instances. On the form to do this, everything works properly until I reach the "Service Role" section.This field does not display any possible ARNs.This is what I see:I created an IAM user and gave them Admin privileges and pasted their ARN into this field ...
Service Role ARN blank in CodeDeploy
Get the SemanticModel for your tree, then call GetDeclaredSymbol() with the ClassDeclarationSyntax. That'll give you the ITypeSymbol and you can look at BaseType from there. You don't want to try this with syntax only because of partial classes.
Good day, I've been trying to get the value of inheritance class name of other inheritance class in Roslyn API.like below picture,when I'm scanning the 'Face' class, I want to get the name of "Human". like this structure, I can get the first inheritance class name by using ClassDeclaration.BaseList. I can get the name ...
How can I get the value of inheritance class of other inheritance class in .NET Roslyn API?
It looks like you made a typo...you're missing.comin the remote's URL:git@github:oscargodson/storageLocker.gitShould be[email protected]:oscargodson/storageLocker.git
First time since a tutorial in July that I'm trying to do a remote push on Github. I added my SSH key, and i can successfully connect to Github, but when i do:$ git push origin masterI getssh: github: no address associated with name fatal: The remote end hung up unexpectedlyI tried:$ git remote set-url origin git@githu...
Github push origin master not working
Thanks to Steffen, who suggested looking at how the keys are created. With further investigation I think I've got a handle on whats happening here. My original suposition that it was linked to the bucket region was a red herring. It appears to be due to what the management console does when you manipulate keys. If you...
I've noticed a difference between the returns from boto's api depending on the bucket location. I have the following code: con = S3Connection(AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) bucket = con.get_bucket(S3_BUCKET_NAME) keys = bucket.list(path) for key in keys: print key which im running against two buckets, on...
S3 boto list keys sometimes returns directory key
What you can do is:initialize a git repo right within your existing VS project:git init .(no additional directory created, except for a.git)Ask "GitHub for Windows" toscan for existing repoand find the one you have created.(Although you should simply be able to drag&drop the VS project directory to the GitHub for Windo...
I have an existing C# project in VS 2012 express that I'd like to add to a repository. I'm very new to source control, so please excuse me if this is a novice question.I installed GitHub and created a repository, and I'd like this repository to contain the source files inVisual Studio\Projects\My Project, i.e. the defa...
Change directory of source files in VS 2012 express
2 The simplest will probably be to revert it. It will create a new commit at the tip of current branch, which the exact opposite changes your bad commit brought. git revert <commitHash> Share Follow ...
I committed some file 3 days before but now i want to remove that commit(or delete files included in that commit specifically if possible) without losing any data after or before that date.if that's possible please Help. Thanks!!
How can i delete a commit on github which i committed 3 days before without losing later data?
For reasons pointed out by @torek, a Git submodule is not the right tool for your task, but creating and committing a simple symbolic link within your repository will do what you want. Assume I have an existing repository file doc/README.md, and I want to add a README.md at the top level of the repository which should...
So I want to link a single README.md to a README.md deeper on in the file structure, I've looked into submodules but I haven't seen A way to link just one singular file that is in the same repo. So far it just creates another folder called README.md with the whole repo in it which is not what I want. Suggestions?
Sub module that points to a single file in the same repo
13 Did a work-around for this issue: Since the IAM user who created the EKS Cluster by default possess complete access over the cluster, inspite of the aws-auth configmap. Since the IAM user who created, had been deleted, we re-created the IAM user, as it would have the s...
Was trying to restrict IAM users with the rbac of AWS EKS cluster. Mistakenly updated the configmap "aws-auth" from kube-system namespace. This removed the complete access to the EKS cluster. Missed to add the groups: in the configmap for the user. Tried providing full admin access to the user/role that is lastly men...
Mistakenly updated configmap aws-auth with rbac & lost access to the cluster
Yes, this is possible with FxCop. A guide to creating custom rules canhttp://www.binarycoder.net/fxcop/index.html.FWIW, I'd recommend creating a "do not eat exceptions" rule instead, with throwing a new exception with the original as an inner exception being one of the "not eating" patterns recognized by the rule. Th...
Our (pretty large) code base has a few of these constructs:try { DoSomething(); } catch (Exception e) { CleanUp(); throw new MoreSpecificException(); }And I want to ensure that we throw a new exception with InnerException set in most of these cases.Is there a standard FxCop rule that detects throwing new ex...
Is there a standard FxCop rule that detects throwing new exceptions without setting InnerException?
*/30 0-1,8-23 * * * /usr/bin/php /var/www/update/inv.phpthe range is inclusive, so 0-1 will do 00:30, 01:30, then 8-23 will do 0830 to 2330ref:http://team.macnn.com/drafts/crontab_defs.htmlShareFolloweditedMar 21, 2013 at 3:13answeredMar 21, 2013 at 2:35uNopleuNople55455 silver badges1010 bronze badges1Yo...
I have a php script which crontab executes every 30 minutes, during off-peak hours around 2-7am I don't get much traffic and so I wish to not run the script during these hours.I'm not sure how to make a cronjob that will do this as I would find it hard to test.The cronjob I have at the moment looks like this*/30 * * *...
cronjob which backs off during off-peak hours
1.Create a new class inherits from NSObject; 2.Make sure it implements AWSCredentialsProvider; 3.Declare required properties such as Access,Secret and Token and initialiser methods; 4.Then in implementation file implement the declared methods. Code snippet: .h file @interface CustomCredentialsProvider : NSObject<AWSCr...
I have credentials generating from Server along with Token.I used AWSStaticCredentialsProvider whereas i need to pass the token for Authentication. Referred many links but no further help other than the idea of custom implementation of credentials Provider.Any code would be helpful. Refer: [AWSWebIdentityCredentialsPr...
How to implement custom AWSCredentialsProvider
In order to use Docker, you don't need to be a root user, you just need to be inside of the docker user group. On Linux: If there is not already a docker group, you can create one using the command sudo groupadd docker. Add yourself and any other users you would like to be able to access docker to this group using th...
New here, was wondering if someone had experience with building images as non root user? I am building Kotlin project, (2 step build) and my goal is now to build it as non root user. Here is what my Dockerfile looks like. Any help would be appreciated: # Build FROM openjdk:11-jdk-slim as builder # Compile application...
Building Docker image as non root user
I have found the answer. Its working. the code is as follows to rewrite and redirect the urlOptions +FollowSymLinks -MultiViews # Turn mod_rewrite on RewriteEngine On RewriteBase / ## hide .php extension # To externally redirect /dir/foo.php to /dir/foo RewriteCond %{THE_REQUEST} ^[A-Z]{3,}\s([^.]+)\.php [NC] RewriteR...
the following code is not working to remove the file extension .phpOptions +FollowSymLinks RewriteEngine On RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME}.php -f RewriteRule ^(.+)$ $1.php [L,QSA]my url not changing fromexample/about.phptoexample.com/about
remove file extension htaccess not working
I got some communication from BT regarding the deprecation of TLSv < 1.2 for their Sandbox, which would entail an update of one's SDK.https://github.com/paypal/TLS-update/tree/master/Braintree?utm_campaign=BT_NA_US_MerchantNewsletter_Q4_20161025&utm_medium=email&utm_source=Eloqua&elq_cid=5173769
My code is:require_once 'braintree/Braintree.php'; Braintree_Configuration::environment('sandbox'); Braintree_Configuration::merchantId('6mh3jxjv5smdpyh3'); Braintree_Configuration::publicKey('j4hgps5mp4bqtm2m'); Braintree_Configuration::privateKey('97a526828d7402c09d4d5fdfff197ab5'); $arguments=array(); echo $token = ...
i got a ssl certificate error when i want to generate a token for every click checkout on braintree
Aha. The key storage flags should be Exportableandpersisted.X509KeyStorageFlags.Exportable | X509KeyStorageFlags.PersistKeySet
On Win Server 2008, I can execute the code below, and when I try to export the cert via the MMC GUI, I do not have the option to export the private key as well. However, if I use the GUI to import the cert, I am allowed to export the private key also. What am I missing in my code?string certfile = @"mycert.p12"; Secure...
X509Certificate2 has private key not exportable?
Hi I have got way how to handle Cookies in Ngnix Server. location / { proxy_pass http://localhost:8080/Url/; proxy_pass_header Set-Cookie; proxy_pass_header X-Forwarded-For; proxy_set_header X-Real-IP $remote_addr; proxy_cookie_domain localhost test.com; proxy_cookie...
I have implemented the spring security with in spring mvc application and it is working fine for me But I am try to access this application with Nginx server configuration now Spring Security stop working at Last I have found that spring security is unable to create JSESSIONID cookies with Nginx server. Can any one...
Spring Security unable to create JSESSION id with Nginx Server
The last assumption is incorrect: memory.Span does not need to pin the memory, as the garbage collector is aware of its underlying reference. Pinning is independently available in case you would like to pass the pointer to a native API.
I believe that following two pieces of code should be equivalent: // first example string s = "Hello memmory"; ReadOnlyMemory<char> memory = s.AsMemory(); using (MemoryHandle pin = memory.Pin()) { Span<char> span = new Span<char>(pin.Pointer, 1); Console.WriteLine(span[0]); } // second example ReadOnlySpan<ch...
How is the memory unpinned in Memory<T>.Span?
Actually, this will take only about 44MB of memory, mostly taken by the output.Your input is20x128x128x3The convolution kernel is4x4x3x32The output is20x128x128x32When you sum up the total, you get(20*128*128*3 + 4*4*3*32 + 20*128*128*32) * 4 / 1024**2 ≈ 44MB(In the above,4is for the size in bytes offloat32and1024**2is...
I'm trying to run a hyperparameter optimization script, for a convNN using Tensorflow. As you may know, TF handling of the GPU-Memory isn't that fancy(don't think it will ever be, thanks to the TPU). So my question is how do I know to choose the filter dimensions and the batchsize, so that the GPU-memory don't get ex...
Avoiding exhausting GPU resources in convNN Tensorflow
Take a look at your/etc/hostsinside thebackendcontainer. You will see192.0.18.1 dir_db_1or something like that. The IP will be different anddirwill represent the dir you're in. Therefore, you must changeTYPEORM_HOST=localhosttoTYPEORM_HOST=dir_db_1.Although, I suggest you set static names to your containers.services...
I'm using nestjs for my backend and using typeorm as ORM. I tried to define my database and my application in an docker-compose file.If I'm running my database as a container and my application from my local machine it works well. My program connects and creates the tables etc.But if I try to connect the database from ...
Docker Compose cannot connect to database
3 Actually when I faced Laravel excel problem on php-fpm 7.4 on docker, I find your question. And this answer RUN apk update \ && apk upgrade \ && apk add zlib-dev \ && docker-php-ext-configure zip --with-zlib-dir=/usr \ && docker-php-ext-install zip don...
This question already has answers here: Fatal error: Class 'ZipArchive' not found in (31 answers) Closed 1 year ago. I have the following error on my docker machine/instance: Clas...
Class ZipArchive not found using Docker [duplicate]
mDNSResponder(multicast DNS daemon) SIGINFO signal can dump a snapshot summary of the internal state to/var/log/system.log, including the cache details. To do this:Keep system log opened in one terminal:tail -f /private/var/log/system.logSend a SIGINFO signal to mDNSResponder from another terminal:sudo killall -INFO mD...
Closed.This question isnot about programming or software development. It is not currently accepting answers.This question does not appear to be abouta specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic onanother Stack Exchan...
How to view DNS cache in OSX? [closed]
3 First off, you'll need a lot of memory for this. Several GBs of RAM, not including GC room. If you're reading a "single line" of 600M digits, in the end, after the line is actually read (you're likely not even getting this far), the string of 600M digits will require 1.2G...
I am trying to read a extremely large file to an array. it is obvious that i am not reading it to an array at the moment but that is not the problem i am having. it is about 600 million digits and this seems to cause it to get an error. Exception in thread "main" java.lang.OutOfMemoryError: Java heap space BufferedRea...
Reading a Large File to an array java.lang.OutOfMemoryError: Java heap space
Something very easy to setup is the new Docker swarm mode (if you have Docker 1.12.2) https://docs.docker.com/engine/swarm/ With this all you have to do is connect your two servers by following the doc. You can then create an overlay network. Then to create your containers you will have to use the command docker servi...
I have two servers on the same LAN. Their IP addresses are 10.0.0.1 (Server A) and 10.0.0.2 (Server B). The MySQL server runs on Server B. The docker container runs on Server A. It's IP address is 172.17.0.2, and the eth0 of the host is 172.17.0.1. My question is, how to connect to Server B in the docker container ins...
How to visit another host inside docker?
9 We are no longer using Vagrant for our backend, and as such we aren't maintaining the Vagrant images. We're looking into a way of doing this, but for now you can email us at support AT travis-ci.org and we can spin up a debug VM for you if you need to debug an issue. ...
These two blog posts describe a way to debug failing regression tests using the same VM image that Travis-CI uses. It's a great idea, but the download link given there is out-of-date: the .box files they link to are 32-bit images, and Travis-CI now uses 64-bit images. Where can I download the 64-bit images that Travi...
Where can I download the 64-bit Travis-CI VM images?
OK, the answer to this question, fundamentally, is: there is no definitive way to tell. Mercurial actually tags commits with the name of the branch they were checked in to, but git simply doesn't; apparently, it isn't considered important what the name of the branch was. This was a design decision and it doesn't loo...
On github, when I view a commit, it shows me the commit message and the changes, along with any comments at the bottom. However, it doesn't tell me what branch the commit was checked in to. Even if I 'Browse Code', it is browsing the code for a particular 'tree' (presumably the state of the code when the commit was ...
How to tell which branch a github commit was for?
In a normal 32 bit windows app, the process only has 2GB of addressable memory. This is irrelevant to the amount of physical memory that is available. So 2GB available but 1.5 is the max you can allocate. The key is that your code is not the only code running in the process. The other .5 GB is probably the CLR pl...
My memory is 4G physical, but why I got out of memory exception even if I create just 1.5G memory object. Any ideas why? (I saw at the same time, in the performance tab of task manager the memory is not full occupied, and I could also type here -- so memory is not actually low, so I think I hit some other memory limit...
Why am I getting an Out Of Memory Exception in my C# application?
I released bindfs 1.13.10 with a workaround for this.Explanation for why it didn't work:https://github.com/mpartel/bindfs/issues/66#issuecomment-428323548
Bindfs doesn't work for folder inside "/proc"...[root@some_host some_folder]# bindfs --map=root/ "/proc//" "/home//" Failed to resolve source directory `/proc//': No such file or directory [root@some_host some_folder]# ls "/proc//" some_fileWhy?Thanks!UPDATE:Example with Docker container...I ended up finding out that f...
bindfs - Doesn't work for folder inside "/proc"
You can useCloudWatch Metric MathAVGfunction to get the average of two metrics:The AVG of a single time series returns a scalar representing the average of all the data points in the metric.The AVG of an array of time series returns a single time series. Missing values are treated as 0.
I want to get average of network input and network output metric as a single metric from Amazon CloudWatch. Is there any way to achieve this or any script or opensource tools available?
How can I get the aggregate of two CloudWatch metrics in a single metric?
1 1). Can I tmp = (char **)malloc(sizeof(char *) * MAX_SIZE) malloc it again without free it? Yes , you can again allocate memory again . But tmp will now point to a new allocated memory and to previously allocated memory . 2). Does the ptr still has the values and also ...
If I have a pointer: char ** tmp = (char **)malloc(sizeof(char *) * MAX_SIZE), after assigning values to each block, I have a new pointer char ** ptr = tmp. 1). Can I tmp = (char **)malloc(sizeof(char *) * MAX_SIZE) malloc it again without free it? 2). Does the ptr still has the values and also tmp points to a new blo...
C malloc a pointer multiple times
It all comes down to how secureor paranoidyou'd like your implementation to be. It may also depend on the type of data you're playing with. For instance: I'd definitely do this for credit card numbers or other sensitive information.Asthecommentshave already stated, you would typically terminate SSL connections at the f...
I have an API server running behind an nginx reverse proxy. It is important to have all requests to my API server be secured via TLS since it handles sensitive data.I've setup nginx to work with TLS (LetsEncrypt) so that seems to be okay. However, requests from nginx to my API server are still insecurehttprequests (thi...
Is HTTPS behind reverse proxy needed?
0 I'm not sure if this is exactly what you are looking for but you could try this: git diff --shortstat "@{1 month ago}" Share Improve this answer Follow answered Oct 26, 2017 at 22:10 sc...
In GitHub (or, using just git), is there a way to get the total number of additions/deletions to a repository over the past month inclusive of all branches (not just master)?
Getting Total Lines of Code Altered Over Past Month Including All Branches?
-1Docker way is "one container for one process". You need one container for apache and one for mysql. You can useofficial php imageandofficial mysql imagefor linking the containers i recommend use docker-composeShareFolloweditedJul 13, 2016 at 19:09answeredJul 13, 2016 at 16:26Bukharov SergeyBukharov Sergey9,97555 gol...
I have the next dockerfile:FROM ubuntu:16.04 RUN apt-get update && apt-get upgrade -y && apt-get install -y apache2 mysql-server mysql-clientAfter, Docker build asking me the password root:While not mandatory, it is highly recommended that you set a password for the MySQL administrative "root" user. If this field is l...
MySQL in Docker frozen at root password config
2 Let's say you have the following log 2021-09-29 15:51:18,624 [main] DEBUG com.company.app.SparkResources - AUDIT : {"user":"Raspoutine","method":"GET","pathInfo":"/analysis/123"} you can parse it like this to be able to handle the part after "AUDIT : " as a JSON fields...
I am trying to parse log entries which are a mix of text and JSON. The first line is text representation and the next lines are JSON payload of the event. One of the possible examples are: 2016-07-24T21:08:07.888Z [INFO] Command completed lessonrecords-create { "key": "lessonrecords-create", "correlationId": "c1c0...
How to parse mixed text and JSON log entries in AWS CloudWatch for Log Metric Filter
Mmm - seems there is a problem with DockerFile support in VS. However, when I use the Orchestration Support, using docker-compose, the functionality works as expected, so I'm answering the question myself :-)
I've got an ASP.NET core 2.2 web app that I have enabled docker support for. I have created a test app here for review here. I am running it in VS with Docker locally. I want to add environment variables/secrets to the app settings secrets in order to override the values in the appsettings.json file. In order to ...
aspnet core 2.2 web app environment variables not changing in docker
Please see similar Github Issuehttps://github.com/aws-amplify/aws-sdk-ios/issues/1671The comments talk aboutthe file is non-sensitive data, so resources that should be accessed by authenticated users should be configured with the approiate controls. Amplify CLI helps you with this, depending on the resources you are pr...
I'm using awsconfiguration.json for AWS Cognito for my iOS Application written in swift. But I'm afraid of security that awsconfiguration.json is stored in my local directory. How can I protect this json file against a third man attack?
How to protect awsconfiguration.json data details in iOS app?
You could link to the GitHub permalink for this branch:https://github.com/$group/$project/tree/$branchnamefor instance, for the upcoming Git version:https://github.com/git/git/tree/nextBut it is easier to just write "this affects $branchname". Anybody working on the project, has it locally anyway and can simply switch ...
I'm wondering if it's possible totarget Issues to specific branches of a repo.Use case:I've a repo which have 1 main branch,and 2 development branches, A and B.From branches A and B other branches grow, depending of the issues.I'm already aware that is possible to reference a commit to issues via '#' keyword and the nu...
How to target specific branch on GitHub issue
No, Its not possible to link multiple git account to one single heroku accounts, if you have a situation , you can switch between heroku accounts or you can maintain repos in a single github account that is linked with your heroku account. Read the link for more info :https://help.heroku.com/5S1JZYSQ/can-i-use-multipl...
I have read the documentationhttps://devcenter.heroku.com/articles/github-integration#enabling-github-integrationbut it does not state how to link more than one github account to a single heroku account.We can see clearly that there is adropdown listnext to the github username so, I am assuming this is possible.Otherwi...
How to link multiple github account to one heroku account?
Is this DeploymentConfig consist of 5 replicas. Then it is not possible to create separate routes for different pods. You have to create separate DeploymentConfigs to be able to create separate routes for different pods.
We want to spinup 5 pods for an application deployed in Openshift cluster. We want to configure 2 routes for the same application.For e.g. Pod1, Pod2 & Pod3 can be reached using route1 Pod4 & Pod5 can be reached using route2..Is it possible to configure two different routes for multiple pods
Openshift Multiple routers for an application
I came across this while searching for an answer to a similar problem and wanted to post my findings for anyone else in the same position.As a team of three, we found that edits to my /etc/hosts file appeared to work, while editing the other two's host files seemingly did not. Upon further digging, we realized I was on...
I'm attempting to redirect an existing web address to a remote machine on my subnet. To do that, I put the following into /etc/hosts192.168.1.249 holub.comand flush the DNS cache withsudo dscacheutil -flushcache;sudo killall -HUP mDNSResponderHowever, when I ping holub.com (the most reliable way that I know to see how ...
/etc/hosts ignored in mac El Capitan (10.11.5)
Edit to address your numbered questions.There are no data types you must allocate with malloc. Only if you want a pointer type to point to valid memory must you use the unary&(address-of) operator ormalloc()or some related function.There is nothing wrong with your code - the line:foo foo2;Allocates a structure on the ...
1)For which datatypes must I allocate memory with malloc?For types like structs, pointers, except basic datatypes, like intFor all types?2)Why can I run this code? Why does it not crash? I assumed that I need to allocate memory for the struct first.#include <stdio.h> #include <stdlib.h> typedef unsigned int uint32; ty...
When must I use malloc to allocate memory?
You can useDaemonsetsin combination withnodeSelectororaffinity. Alternatively you could configurepodAntiAffinityon yourPods, for example:apiVersion: apps/v1 kind: Deployment metadata: name: rest-server spec: selector: matchLabels: app: rest-server replicas: 3 template: metadata: labels: ...
In my application, I have a rest server which locally interacts with a database via the command line (it's a long story). Anyway, the database is mounted in a local ssd on the node. I can guarantee that only pods of that type will by scheduled in the node pool, as I have tainted the nodes and added tolerances to my p...
How to ensure only one pod runs on my node in GKE?
+200Why do you want to do this? It is a bad idea to train users to indiscriminately install root CA certificates just because a web site tells them to. You are undermining the entire chain of trust. A security conscious user would ignore your advice to install the certificate, and might conclude that you are not taking...
I have a desktop product which uses an embedded webserver which will use self-signed certs.Is there something that I can put in a web page that would detect that they haven't added the root CA to their trusted list, and display a link or DIV or something directing them how to do it?I'm thinking maybe a DIV that has ins...
Self-signed certs -- helping users know they need to add root CA to trusted cert store
You can usefindto accomplish this.find -mtime 1 -regex [your_pattern_here] -exec rm -f {} \;mtimelooks for any files older than N days old, and the[your_pattern_here]in this case would be the pattern of files you want tokeep. It'd be best to do this without theexecportion at the end first to make sure it's finding the...
Closed.This question isoff-topic. It is not currently accepting answers.Want to improve this question?Update the questionso it'son-topicfor Stack Overflow.Closed12 years ago.Improve this questionI need to figure out how to automatically remove most files from my /tmp directory on a centos server. The directory keeps fi...
How do I delete old[er] files from my centos /tmp directory EXCEPT certain files still in use? [closed]
First, add a function node at the end of each sensor node and save the output as variable. The code will vary greatly depending how you are getting your sensor data, but here is how I do it:msg.payload = Number(msg.payload); flow.set("presion_agua_psi",msg.payload); flow.set("sensor_presion_agua","Wemos D1"); return {p...
It is quite a while since I coded something and it is the first time I am dealing with Influxdb and NodeRED. I am acquiring four sets of measurements from a sensor connected to a Pi. This is a screenshot taken during the debug, the measurements are coming trough.I managed to get the data from the sensors into NodeRED:T...
InfluxDB and NodeRED
This can be done by having some sort of standard metadata file and then having a bot crawl look in each repo and aggregate the data. This bot essentially builds a catalog of every service.We use this approach at my current employer, where we include data about the service, its name, the team that owns it, its importan...
With the growing number of microservices, and people moving around projects. It is getting difficult to find someone who supports a service.Is it possible to put into GitHub enterprise some metadata and who the owners are and the people supporting it?AI and other knowledge management tools such as a bot can find or ind...
How can we store additional meta data about a Git repo?
I add something like this to my php - be sure to send headers before you write out anything else....header("Cache-Control: no-cache, must-revalidate"); header("Expires: Sat, 26 Jul 1997 05:00:00 GMT");
The title might be kind of bad or misleading but here is my question:Is there a good way to make sure that if someone has visited my site in the past and then revisits it after some time, the old content won't be showing and the new content is there instead?
How to clear the cache if site appears wrong
1 Git is completely self-contained locally, so there's no way of knowing what people are doing with your repository on their local machines. So no, there's nothing built-in that would answer your questions. You could tell people to push all work to their forks or "personal ...
I'm working on an OS project with unmergable source files that is living on github. As a rule, we want to avoid merging files, and discourage parallel development of the same source. So far the team has been co-located, and coordination has not been hard. However, I am concerned that as the project grows in the numb...
How can I extend git to see the files other people are editing with git or github?
Everything looks correctYes, pfx is just a container that wraps up the cer file and the pvk file in to a single package that is easy to deploy. Only the cer portion ever leaves the server in any request.Yes, the cer is the certificate. When you view the certificate in the browser you are looking at all the details insi...
I have used a mkcert.exe tool to generate self signed .cer file and then pvk tool to generate .pfx file.I hosted a website in iis and then deployed a .pfx file in IIS server.When a client system access the site ,first it will prompt like certificate is not installed, after forcefully opening site, we will install certi...
what is the difference between .cer and .pfx files
So I found that there's a function that is called in every pipelines when the spider closes after it's finished crawling and everything is through the pipeline, which isdef close_spider(self, spider): passThere's also a function called on startup, which isdef open_spider(self, spider): pass
I'm working on writing a scrapy pipeline that will call a function to clear our cdn's edge servers of the scraped urls. I figured out how to store the list of visited urls easily enough, but the issue is knowing when the crawler is done.The cdn's api takes urls in batches of 100, so I can easily call it's clear functio...
Store scrapy items to process after spider completes
I found the problem and the solution.From Sonar website I see this:By default, when no coverage report is found, the JaCoCo plugin will not set any value for coverage metric. This behaviour can be overriden to force coverage to 0% in case of a lack of report by setting the following property :sonar.jacoco.report...
I'm running Jacoco and Sonar on multimodule Java8 project. I have unit tests in each of the modules and to save resources I collect all 'integration tests' into one "integration-tests-runner" and run them all there (wrapping them with before and after tests).When measuring coverage UT generates exec file per module tar...
Sonar false overall coverage (jacoco)
1 Ideally, you would: connect as nginx directly to make the installation don't manage the private/public key through GitLab, but through a deployment tool like Ansible (see "How to use GitLab and Ansible to create infrastructure as code") That way, no chown to do, and t...
I am trying to find some guide or documentation that discusses best practices for setting up gitlab CI/CD to auto deploy a web server (nginx) / Centos or any Linux. Setting up the CI/CD as user root is easy, but i don't like the idea of having a root key in gitlab. If i create a 'gitlab' user and assign it to the same...
Gitlab CI/CD - user and permissions best practices
Not directly with GitHub page, which only support GFM (GitHub Flavored Markdown Spec) What you can do is use your GitHub repo with those mardown pages as sources for generating a static website (where you can add any additional information to the HTML code you are generating). A static website generator like Hugo can ...
I have a markdown (MD) page on GitHub in a public repository that I want to add some additional information when someone clicks on a hyperlink. Ideally something similair to Bootstrap's Popover. Is this possible and if so how?
Is it possible to create a tool tip, info tip or hint in GitHub markdown?
A Github-centric way to do this would be to create a private repository for every one of your clients that is a fork of your main repository. The client repositories would have two branches:masterwhich tracks the main development branch in the original repository andprojectwhich shows the code that you're developing fo...
I work in a product development company. We have a CMS application which was developed on MEAN stack and we use Jenkins CI for automating deployments. We use github for hosting our code.We receive requests to implement new features from our clients and these features should only we available to the clients who request ...
What would be the Git Workflow - Single Product, Many clients?
It seemsKEP-2307: Job tracking without lingering Podsis now GA, since Kubernetes 1.25 (Aug. 2022)It includes the ability, for a Job Controller, todelete podsIn the case where a user or another controller removes a Pod, which sets a deletion timestamp, the Job controller treats it the same as any other Pod.Since deleted...
Pods inerrorstate are not getting cleaned up.I can manually remove these pods using:kubectl delete pod --field-selector status.phase=FailedIs there anyannotationorlabelthat I can use for auto-cleanup? What's the recommended way to do this?
How to delete failed Kubernetes pods automatically?
You can set a static nodePort in your Service definition that will not change on each deploy.For example:{ "kind": "Service", "apiVersion": "v1", "metadata": { "name": "my-service" }, "spec": { "selector": { "app": "MyApp" }, "ports": [ { ...
I have an app deployed on kubernetes cluster. The kubernetes node port is 30010 which will be redirect to 41018.[root@kubernetes-slave ~]# iptables -L -n -t nat Chain KUBE-NODEPORT-HOST (1 references) target prot opt source destination DNAT tcp -- 0.0.0.0/0 0.0.0.0/0 /* d...
Should I add an iptables rule on the random kubernetes node port if I want to access from outside world
This is probably because your cluster is running in a different timezone then the one used by you.You can check what timezone will be set in aPODusing:kubectl run -i --tty busybox --image=busybox --restart=Never -- date.As for youryamlit looks good, there is no need to change anything with thespec.schedulevalue.A small...
When I set the Cronjob Schedule as*/1 * * * *,it would work.When I set any number which is in0-59to thecrontab minute,such as30 * * * *,it would work as well.However when I set the Cronjob Schedule as30 11 * * *,it even doesn`t create a job at 11:30.All the config is followed:apiVersion: batch/v1beta1 kind: CronJob m...
How Can I set Kubernetes Cronjob to run at a specific time
0 You can set your origin as usual: git remote set-url origin git://usr/to/mirrorr/<user>/<repo>.git But change the push url to your GitHub repo git remote set-url --push origin git://github.com/<user>/<repo>.git Share Improve this answer ...
In git 2.3, we can git clone codes fast using --reference and --dissociate by borrowing objects from existing clones. My question is can we make git fetch like that? Because under some poor network condition, git fetch from github will be very very slow. I want to make a mirror on my own server, so I can fast fetch f...
Can I fetch changes from more than one repo in git?
If you want to changeindex.php?p=foo&idnews=barto/foo/bar, try theses rules :RewriteEngine on ##externally redirect "/index.php?p=foo&idnews=bar" to "/foo/bar"## RewriteCond %{THE_REQUEST} /index\.php\?p=([^&]+)&idnews=([^&\s]+) [NC] RewriteRule ^ /%1/%2? [L,R,NC] ##internally redirect "/foo/bar" to "/index.php?p=foo&...
I want to create .htaccess mod rewrite but still have problem.Suppose i created friendly URL like this :mydomain.com/mypage/12345/title-of-newsI want that friendly URL process the below URL in hidden process :mydomain.com/index.php?p=mypage&idnews=12345Values of "p" and "idnews" are dynamic, so they will have different...
Simple mod rewrite query string but still failed
6 This has several problems. void *memset(char *s, char c, size_t n) { size_t i; for (i = 0; i < n; i++, s++) /* incrementing s means you cannot return */ { /* the original value */ *s = c; /* consider using s[i] = ...
Folks, here is an implementation of memset(), however I have been told that there is one logical mistake in the code. Could you help me find it. I feel that a double pointer for the target string should be passed to this function, which will be like passing the address of the pointer variable and not the pointer itse...
what is wrong with this code snippet
Assuming that you upload from the top level directory as listed below:- docker-compose.yml - app/ - package.jsonyou are mounting in your directory contents as defined here:volumes: - ./:/appThis in turn results in the/appdirectory containing theappdirectory, so it will look like/app/app/package.json.you...
So I am trying to use dokcer compose utility to spin up containers for web app and database.I havedocker-compose.ymlfile in the same directory, as theappdirectory, which containspackage.jsonfile, so it looks like this (for brevity, rest of files ommited):- docker-compose.yml - app/ - package.jsonAnd the docker comp...
Docker compose: failing to start container, as it fails to find file, which exists locally
st way is change the url as suggested in comment.2nd is :$url = $_SERVER['HTTP_REFERER']; if (strpos($url,'yelp') !== false) { header("Location : YOUR_PAGE_URL"); }
I would like visitor from my yelp listing to land on a different page of my website.For example: On yelp.com my website URL is domain.com. But when the visitor clicks on that URL, they should be redirected to domain.com/newpageI am trying to do this only for the visitors who come to my website through yelp. How is this...
Redirect visitors only from Yelp.com
The URL part of an ErrorDocument directive should either start with a leading slash, which indicates a path that's relative to your DocumentRoot, or it should be a full URL (if you want to use an external document).You shouldn't need the RewriteEngine and RewriteRule directives at all here.So, assuming your notfound.ht...
I want to redirect any 403 using .htaccess, but it does not seem to work.Options +FollowSymlinks RewriteEngine on ErrorDocument 403 notfound.html RewriteRule notfound.htmlAll help appreciated.Thanks Jean
redirect 403 error using .htaccess
I'm not sure whichbackend host headeryou removed from the default pool, but I'm thinking that wasn't necessary. Afteradding your custom domain to AFDyou can use anAzure managed certificate. I would simply go that route. But if you have any issues, comment down below.
I have an app service with front door setup on Azure and have added a custom domain to the front door.To ensure the request url is recorded correctly on server, I have cleared the "Backend host header" property on the defaultPool on the front door.This initially resulted in a 404 when using my custom domain, but adding...
Azure app service with unsecure custom domain and front door
3 I think using branches would be a better option to go with it has many advantages over committing the code in the main branch Share Improve this answer Follow answered Jun 3, 2019 at 12:57 ...
Suppose I have a local repository with folder structure like this: ProjectName └── src └── main └── java └── projectname └── subfoldername ├── File1User1.java ├── File2User1.java ├── ... └── FileNUse...
Github pushing commits to different folder structure
If you know the query you want to run, you can use the-eargument for mysql at the command line for your script. So the "Command to Run" in your cron management tool would be:mysql -u <username> -p<password> -h <name-of-mysql-server> <databasename> -e "<YOUR-QUERY-HERE>"The general structure of a query to delete recor...
I'm trying to set up a Cron job for deleting MySQL records where a date field is older than three weeks, but I can't figure out what the string is that goes in the box. Here's a pic of the Cron management screen. Can anyone help please?http://i46.tinypic.com/id4nsj.jpg
Require Cron syntax for MySQL maintenance
To run pg_dump you can usedocker execcommand:To backup:docker exec -u <your_postgres_user> <postgres_container_name> pg_dump -Fc <database_name_here> > db.dumpTo drop db (Don't do it on production, for test purpose only!!!):docker exec -u <your_postgres_user> <postgres_container_name> psql -c 'DROP DATABASE <your_db_na...
i'm running my app using docker-compose with the below yml filepostgres: container_name: postgres image: postgres:${POSTGRES_VERSION} volumes: - postgresdata:/var/lib/postgresql/data expose: - "5432" environment: - POSTGRES_DB=42EXP - POSTGRES_USER=${POSTGRES_USER} - PO...
how do i backup a database in docker
Prometheus doesn't provide a function, which could return the percentage of raw samples with values smaller than the given threshold on the given lookbehind window. This functionality can be emulated viasubquery feature. For example, the following query returns the percentage ofprobe_duration_secondssamples with the va...
I'm using theblackbox exporterto gather metrics from various endpoints, and I want to set a SLI to determine the number of GET requests that are slower than300msand1sper service.The exporter provides agaugemetric calledprobe_duration_seconds.I'm trying to run a PromQL query to calculate the percentage of probe_duration...
Prometheus - Percentage of gauge values below a certain threshold
GitLab has a super robust API. It's probably the way to go here.TermsI do want to standardize on some terms so I can make sure my recommendation is making sense. You say that yourstagesare running in parallel- but it isjobsthat run in parallel within agivenstage. For my response, I'm going to assume that you meant that...
I have defined 3 stages in gitlab-ci.yml. When there is a new commit, pipeline runs and these 3 stages run in parallel, which is expected and needed. (These stages run pre-requisite steps like security checks on code and other linting functions). I also have Scalr (another provider) inject external stages into the sam...
How to make external stages in gitlab pipeline "wait" until all the gitlab internal stages are done?
Based on @M.Deinum comment... I used ApplicationListener but with ApplicationReadyEvent! So, my example becames:@EventListener(ApplicationReadyEvent.class) @Scheduled(cron="0 0 5 * * *") public void somethingToDoOnRebootTime() { // code here, to run every day at 5a.m., AND at boot first time... }
I made many searches over internet about an option mentioned by Baeldunghere, but I can't find any example. I would like to use something like this:@Scheduled(cron="@reboot") @Scheduled(cron="0 0 5 * * *") public void somethingToDoOnRebootTime() { // code here, to run every day at 5a.m., AND at boot first time... }B...
Can we make a @Scheduled execution on Spring, mixed with cron parameter, forcing a first execution at boot time?
I have very large databases with tens of thousands of tables some of which have up to 5GB of data in 10's of millions of entries. (I run a popular service)... I've always had headaches when backing up these databases. Using default mysqldump it quickly spirals the server load out of control and locks up everything... ...
While doing a MySQL dump is easy enough, I have a live dedicated MySQL server that I am wanting to setup replication on. To do this, I need dumps of the databases to import to my replication slave.The issue comes when I do the dumps, MySQL goes full force at it and ties up resources to the sites that connecting to it. ...
How can I slow down a MySQL dump as to not affect current load on the server?
includeis meant to be used with templates and not to simply output values.According to thedocumentationTo make it possible to include a template, and then perform an operation on that template’s output, Helm has a special include function:{{ include "toYaml" $value | indent 2 }}The above includes a template calledtoYam...
I searched a lot but didnt get a right answer, what is the difference between include and {{}} in helm.ex :metadata: name: {{ include "cluster-dev.fullname" . }}here can't it be like thismetadata: name: {{ Values.nameOverride }}
What is the usage of include in helm
You could put an invisible character between the # and the number. e.g. #&#8203;5 will be rendered as #5 on GitHub. &#8203; being the unicode "zero width space" character.
As opposed to reference an issue, I want to just output the text #5 without any link. How can I do that? For example, a shader error message from Unity could contain # and a followed number: Shader error in 'Nature/Terrain/Standard': actual parameter #3 must be same type as formal out parameter ("weight") actual...
How to escape a '#' so it will NOT reference an issue in a GitHub comment?
Need to translate the YAML file to JSON file:cat > nginx-pod.json <<EOF { "apiVersion": "v1", "kind": "Pod", "metadata": { "name": "nginx1" }, "spec": { "containers": [ { "name": "nginx", "image": "nginx:1.7.9", "ports": [ ...
How can I create a Pod using REST API ?I checked the Kubernetes API documentation:https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.10/#-strong-write-operations-strong--54They are writing that need to use POST request:POST /api/v1/namespaces/{namespace}/podsI have this YAML of simple nginx pod:cat > ngin...
How to create Pod from REST API
Since the problem is not with Internet Explorer, but the generic validation of certificates, I was able to debug this with Chrome's "Console" interface (Developer Tools via F12). The URLs which were causing the SSL error were listed with the console error "Failed to load resource", and the URL was given.Internet Explo...
I have a website that IE9 says "Internet explorer blocked this website from displaying content with security certificate errors."I have no idea what errors it's talking about however. Firefox correctly shows the site is secure and that all content is correctly served over HTTPS. The certificate is not expired. The cert...
IE9 how to trace "Internet explorer blocked this website from displaying content with security certificate errors."
I found the solution thanks to a StackOverflow user's comment. I don't know why when I created the VPC Endpoint Gateway to connect to Dynamodb, AWS didn't update automatically my route tables. It's a VPC Endpoint Gateway so you have to be sure that the traffic to Dynamodb is redirected to the gateway created (with a V...
I'm setting up a new AWS Fargate task to query Dynamodb. The task is not public, so it doesn't have a public IP address and the subnet in which the task is placed doesn't have the access to the Internet. The Dynamodb table I'm trying to query uses the KMS encryption type. I created: a VPC endpoint gateway to allow t...
I can't query dynamodb table from AWS Fargate task
The static analyser will attempt to find errors in your code by inspecting it without running it. It's really good at spotting times when you accidentally make an allocation error. It'll tell you about things that are unarguably going to cause a problem and occasions when you've stepped outside of normal Objective-C c...
I'm somewhat new to Objective-C and while I think I've got a decent grasp of memory management, I'd like to become familiar with the best ways to test this (if possible). I know of two options: 1) Build and Analyze (from within Xcode) and 2) Instruments. Are these the best tools for the job? How accurate is the Build ...
Testing memory management in Objective-C (iOS)
Short answer: No.Long answer: Yes?The whole philosophy of CouchDB is to think about everything that might be queried later, compute it now (when it's convenient), and store the result in an index. All data access from CouchdB comes from scanning one of these indexes, which is very fast and tends to take roughly the sam...
Does CouchDB supports caching out of the box or do I have to manually have a caching proxy in front of it like Squid?
CouchDB supports caching?
That error is generated when App Engine needs to shut your backend down but the backend fails to exit within 30 seconds. Some reasons why this might be happening are listedhere. Depending on the type of error, App Engine may be sending your backend a notification of the impending shutdown, so it's a good idea toregiste...
My backend job is working on the basis of cron job(every 4 hour).But it is terminated with out processing the data. The server log displays as following :500 15377121ms 0kb instance=0 AppEngine-Google; (+http://code.google.com/appengine) E 2012-10-05 01:50:18.044 Process terminated because the backend took too long to ...
GAE :Process terminated because the backend took too long to shut down in backends job
0 The error `"Exiting with failure status due to previous errors" means exactly that. There was an earlier problem which, while not fatal to the running of the program, is reason enough to exit with a failure code. Given that you're backing up from the root level, this is a...
I've been trying to backup my ubuntu11.04 with the following tar command sudo tar -cvpzf /media/TOSHIBA\ EXT/backup.tar.gz --exclude=/backup.tar.gz --exclude=/lost+found --exclude=/proc --exclude=/sys --exclude=/mnt --exclude=/media --exclude=/dev --exclude=/home/manuzhang/Music --exclude=/home/manuzhang/Videos --excl...
errors when backuping ubuntu11.04 with tar
tibble::lst()creates self-named lists (analogous todata.frame).map(lst(A, G, M_1, M_2), \(x) x[[1]] %*% x[[2]])or slightly more inscrutablymap(lst(A, G, M_1, M_2), lift(`%*%`))(or, of course)lst(A, G, M_1, M_2) %>% map(lift(`%*%`))
I have a list of 4 matricesA = list(a = matrix(1:4, 2), b = matrix(2:5, 2)) G = list(a = matrix(10:13, 2), b = matrix(5:8, 2)) M_1 = list(a = matrix(10:13, 2), b = matrix(5:8, 2)) M_2 = list(a = matrix(2:5, 2), b = matrix(5:8, 2)) dlist <- pmap(list(A, G, M_1, M_2), list) %>% simplify() %>% map(., ~{names(...
Multiply all the matrices in a nested list together
You need to adjust the *a pointer, this means you need to pass a pointer to the *a. You do that like this: int *a; void initArray( int **arr ) { *arr = malloc( sizeof( int ) * SIZE ); } int main() { initArray( &a ); return 0; }
I need to do a simple thing, which I used to do many times in Java, but I'm stuck in C (pure C, not C++). The situation looks like this: int *a; void initArray( int *arr ) { arr = malloc( sizeof( int ) * SIZE ); } int main() { initArray( a ); // a is NULL here! what to do?! return 0; } I have some ...
Dynamically allocating arrays of structs in c++. Why is the array getting overwritten? [duplicate]
If you look closely at the screenshot, you'll see that (and maybe infer that) you're in a shared hosting service. The hosting provider puts all domain names in a separate directory inside thedomainsdirectory.Thus, yourDOCUMENT_ROOTshould be/domains/spss-tutorials.com/public_html
EDIT: I found it. It turned out to be/home/spss01/domains/spss-tutorials.com/and I don't think I could have figured that out without the help of my hosting provider.END EDIT.For example, I've an image in adownloadsfolder whose url ishttp://spss-tutorials.com/downloads/decimals_histogram.pngHowever, in FileZilla it look...
What's my documentroot?
I could fix this issue by forced an update of CSF, so the configuration includes the newer binary command locations.ShareFollowansweredFeb 1, 2016 at 13:18TechiescornerTechiescorner80322 gold badges1313 silver badges2626 bronze badgesAdd a comment|
I could see the following warning on top myCSF WHMplugin.WARNING: RESTRICT_SYSLOG is disabled. See SECURITY WARNING in Firewall ConfigurationAlso I am getting the following error while checking thewatch system logExecutable [] invalidCan anybody help?
RESTRICT_SYSLOG warning in csf cpanel plugin
I found the following roles to work. Make sure you attach these policies to your group or user in IAM:Private registry:AmazonEC2ContainerRegistryFullAccessPublic registry:AmazonElasticContainerRegistryPublicFullAccessShareFollowansweredJan 14, 2021 at 21:34dataviewsdataviews2,72188 gold badges3434 silver badges7676 bro...
I am facing the following issue when attempting to retrieve an authentication token and authenticate your Docker client to your registry:user is not authorized to perform: ecr-public:GetAuthorizationToken on resource: *I was able to push an image to my private repository just fine without issue. I am using an EC2 ubunt...
AWS ECR user is not authorized to perform: ecr-public:GetAuthorizationToken on resource: *
-1 This was indeed answered in Change Basic HTTP Authentication realm and login dialog message. Short explanation: You were actually defining realms with auth_basic directives of Nginx on the server side. But "whether to prompt this message or not" is basically a design cho...
I am using nginx proxy to server my web-page. For login user need to provide his 2 factor authentication code and his password, to let users know that they need to enter their password+2 factor code to login, I send them a message "Login required, username, password+VIP token"' And this what I get now This has stopp...
Chrome basic authentication custom message stopped working
Try using requests to download file as blockfilename="something.mp4" r = requests.get(LINK, stream=True) with open(local_filename, 'wb') as f: for block in r.iter_content(block_size=1024): if block: f.write(block)Referhttp://docs.python-requests.org/en/latest/user/advanced/#body-content-workflow
I have a simple program that reads data from a website then writes it to a file usingurllib. But the file isseveral hundred megabytesin size and Python keeps raisingOut of Memory Errorimport urllib content = urllib.urlopen ( LINK ) f = file ("Something.mp4", "wb") f.write ( content.read () )Then when I download and wri...
Efficiently read and write large file
Support for IMDSv2 was added in:1.11.678of aws-java-sdk (commit06a2180e)2.10.20of aws-java-sdk-v2 (commit53451414)I believe the default chain for both versions of the SDK will attempt IMDSv2, but have not verified that.
If an instance is configured forIMDSv2then after upgrading AWS-CLI we are able to connect to IMDSv2 , but we can not connect to it from java code.Also, we could not find any documentation regarding whether we need any specific version of AWS java SDK to support IMDSv2, OR is it supported implicitly, OR is it not suppor...
How to use IMDSv2 from AWS Java SDK
I have found the Solution In my admin folder I have a php.ini file I have changed my memory_limit = 64M; to memory_limit = 128M;. Now it works fine.. I didn't get Fatal Error.
I'm using opencart. In my admin page when I access CATALOG>PRODUCTS (I have 73 products - Totally I have four pages). When I access my second page it shows this following error Fatal error: Allowed memory size of 67108864 bytes exhausted (tried to allocate 4000 bytes) in /home/justborn/public_html/system/library/image...
Fatal error: Allowed Memory size of 67108864 in opencart
At the moment, this is a recurring issue with no simple answer.There are two common approaches I hear of.First involveschowning the directory before using it.RUN mkdir -p /home/jboss/myhub/logs ; chown -R jboss:jboss /home/jboss/myhub/logs USER jbossIn case you need to access the files from your host system with a diff...
trying to mount a volume to my container from thedocker runcommand. It seems like the folder is always created as root instead of the container user. This makes it so that I'm lacking rights on the folder(cant create or write files for logging).Doing some testing using this command:docker run -it --entrypoint /bin/bash...
Mounted folder created as root instead of current user in Docker
From the clientCache documentation The value for the httpExpires attribute must be a fully-formatted date and time that follows the specification in RFC 1123. For example: Fri, 01 Jan 2010 12:00:00 GMT So, if you want to use the http expires headers for your static content, set it like this: <staticContent> <cli...
I'm trying to get YSlow to give me an A on the "Add Expires header" section by setting the web.config file. I've been looking around and this is what I put in based on what's out there: <staticContent> <clientCache httpExpires="15.00:00:00" cacheControlMode="UseExpires"/> </staticContent> </system.webServer> ...
http expire headers in asp.net with web.config
Please note that HSTSPolicyBypassListonly applies to static HSTS-preloaded entries (for example, "app", "new", "search", "play"). This policy does not prevent HSTS upgrades for servers that have dynamically requested HSTS upgrades using a Strict-Transport-Security response header.You may check whether you have met thes...
This has been written about many times, but I was not able to solve itthis is a small company with serveral (about 50) devices solely with internal IP with web server (like ESXI web, printer, ...)upon opening the sitehttps://myesxiserver.domain.localorhttps://myesxiserverI getNET::ERR_CERT_AUTHORITY_INVALIDcurrently I ...
NET::ERR_CERT_AUTHORITY_INVALID in EDGE for INTERNAL site
The "traditional" way to allocate memory is to have it in a contiguous address space (the "heap", growing upwards by calls to sbrk()). Each time you hit a page boundary, there will be a page fault and you get mapped a new page. There are two consequences of this strategy: pages can only be freed when all allocations ...
I realize that most CPUs are better at reading data at an aligned memory address, that is at memory address that is a multiple of CPU word. However, in many places I read about allocating a page-aligned memory. Why might someone want to get a page-aligned memory address? Is it only for even bigger performance?
What are benefits of allocating a page-aligned memory chunk?
This should have a similar effect to the scriptmentioned in Yu-Lin Chen's commentbut makes it explicit on the command line. You might find it easier to avoid escaping/resolution problems with it:git filter-branch -f --env-filter "if [[ \$GIT_AUTHOR_EMAIL = '[email protected]' ]]; then[email protected]; fi" HEADYou'll p...
So, our team has been working for several months on a project on Github. It turns out that one of our team members misconfigured his local git bash shell, and was commiting with an email address that had a typo in it.The commits have already been pushed to the remote, and are fully ingrained into the history of the rep...
Change email associated with past commits in git
2 The problem is here: git push You must be using old version of git <2. In older version whenever you use git push/pull without any branch name after the command git push all you branches (modified) to the server. In your case you must have cloned the pull request as well...
I am trying to clean up git repository. There are some identified big files whose size is reduced and recommitted. The issue is history still has old files. So I used bfg cleaner job to prune the git repo. For this I first create a mirror clone of repo and then do some filtering to reduce the size. Clone command:- gi...
Git push error while trying to push mirror to remote repository
Isn't this the job of the load balancer? To make sure requests are routed appropriately?Yes, you are right. But...replicas: 1You only use a single replica, so the load balancer has no options to route to otherinstancesof your application. Give it multiple instances.I've come to find that if multiple requests are made a...
I have 2-3 machine learning models I am trying to host via Kubernetes. I don't get much usage on the models right now, but they are critical and need to be available when called upon.I am providing access to the models via a flask app and am using a load balancer to route traffic to the flask app.Everything typically w...
Multiple Requests Causing Pod to Crash due to OOM - isn't this the job of a load balancer?
Your last line (RewriteRule) should look like thisRewriteRule ^([^/]*)$ index.php?p=$1 [L]Don'tadd a leading slash before.Otherwise it will be like anabsolutepath (/index.phpfrom root).With the rule above and yourRewriteBase, this rule will act this way:If not existing file/folder then rewrite it to /jellyfish_rework/i...
I am working on a website to make the URL's more SEO friendly with thehtaccess, however, i bump into a problem when I set up the htaccess file. If I use the URLs I prefer to setup, i get refered back to the localhost and not towards the development map I am working in, so in this case, i get refered back tolocalhost/bl...
htaccess on Localhost refers to root, not folder
So it seems that is in fact an error with gits SSH gist server. While this works: streamgraph git:(master) ssh -T [email protected] Hi adrianturcato! You've successfully authenticated, but GitHub does not provide shell access. This times out: streamgraph git:(master) ssh -T [email protected] How I fixed this: I cha...
I am trying to push some code to a gist. I know there are a lot of other people with similar issues, and I have looked at all of the other posts I can find and they don't seem to address my problem. ➜ streamgraph git:(master) git push ssh: connect to host gist.github.com port 22: Operation timed out fatal: Could not ...
ssh: connect to host gist.github.com port 22: Operation timed out fatal: Could not read from remote repository
You should be able to useLimitRangeresource for defaults. Ref:https://kubernetes.io/blog/2022/09/19/local-storage-capacity-isolation-ga/apiVersion: v1 kind: LimitRange metadata: name: storage-limit-range spec: limits: - default: ephemeral-storage: 10Gi defaultRequest: ephemeral-storage: 5Gi ty...
What happens if I don't set ephemeral-storage limits in any of my pods? How does kubernetes sets the default limits?I'm asking because I an error like this:the node was low on resource ephemeral storage xxxxKi which exceeds its request of 0What I don't understand is how the limit is being set if I don't set it?(btw, I ...
Default setting of ephemeral storage
The issue is with github, the notebook is there but the rendering fails. You need to refresh it multiple times until it loads. There is an issue here: (still unresolved in 2021)https://github.com/jupyter/notebook/issues/3555#issuecomment-748173157
Google's colab notebook allows saving to a github repo. It's awesome! But after updating my notebook 30+ times. I'm now getting an exception when I save to the repo.Despite the ominous message, the repo seems to be updated. Any clues on what is going on here?
Google Colab Notebook save to github error
Thehostnamedirective simply sets the hostname inside the container (that is, the name you get back in response to thehostnameoruname -ncommands). It does not result in a DNS alias for the service. For that, you want thealiasesdirective. Since that directive is per-network, you need to be explicit about networks rather...
I'm unable to get the Docker Composehostnamecommand to work.I'm running a simpledocker-compose.yml:version: '3' services: redis1: image: "redis:alpine" hostname: redis1host redis2: image: "redis:alpine" hostname: redis2hostOnce I run this withdocker-compose up, I should be able to rundocker-compose ...
Docker Compose hostname command not working