Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
One suggestion could be to forget its Kubernetes worker nodes, and think about how you will give some domain pointing to a set of instances? Imagine you are running copy of your static website in 10 servers and you want to have same domain to all the nodes? You have either an external load balancer or a reverse proxy. ...
Let's suppose I have bare-metal servers forming a Kubernetes cluster where I am deploying my application. How can I point one domain name to all of the worker nodes of the cluster without a Load Balancing Service or Ingress Controller ?
How to point domain to Kubernetes cluster without Load Balancer and Ingress controller
Don't test for a pointer's "sign", it's irrelevant. The only invalid pointer that malloc will return is the null pointer. If it returned something else, you can use it.
I am trying to allocate ~ 1GB , but when I check the pointer address it's a negative number. Could this be a configuration limit I am hitting somewhere. goal to allocate ~1GB limit of ~0.7GB ([struct]48 * 16890) 1.5GB free memory no large amounts of memory were allocated before this. using fedora linux. no compiler...
C: allocation limits
I missunderstood this. It's pretty obvious now: Client: 1.13.0, Server: 1.7.16. The server must be the clients version (+/-1 is supported, seeherefor more). So I edited the server version usingkops edit clusterand updated it withkops update cluster.
According to this:https://github.com/kubernetes/kops#compatibility-matrixthe versions should be fine. When I runkubectl get nodeI get the following output:Error from server (NotAcceptable): unknown (get nodes)kubectl version:Client Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.0", GitCommit:"ddf47ac13c...
Kubectl returns Error from server (NotAcceptable): unknown (get nodes)
Could you please create a bug report athttps://support.mibexsoftware.com? We can then analyze the issue in detail. It would also help if you could send us the debug logs of the plug-in. I can give you more detailed instructions after we have your bug report.Thanks,Michael from Mibex Software
I have SonarQube for Bamboo plugged in and working fine with MSBuild. I also have SonarQube or bit bucket server, which, as far as I can tell, is configured correctly, but pull request analysis is not working. I have a repository in bit bucket configured to allow Sonar Qube analysis and have min. severity set to INFO. ...
Pull request analysis not working in Sonar Qube for bit bucket server
1 As per your description, there could be a possibility as stated. You need to check your package.json file for updated dependencies. Also, it is likely that node_modules/ path in your repository source is added to .gitignore file and is hence not tracked by Git, due to whi...
I'm creating a next app for work so I've been working on it on my home and work computers. Last night I was working on the project on my home computer that already had the Supabase package installed. I pulled my changes to my work computer. Supabase requests were working as expected. I committed and pushed my changes ...
Missing Supabase modules after pulling from github
First of all, secrets is not what you should use for deploying. Secrets exists for safe storage during development by helping prevent sensitive data from being storing in code / checked into source control. See related question How do UserSecrets work in the Cloud? And you are right, that you should use Environment Va...
I've started a new .net core project and I configured it to use the Google+ Api for authentication. My Client Id and Secret are stored using dotnet user-secrets. When I build my docker image I get Unhandled Exception: System.ArgumentException: The 'ClientId' option must be provided. I know this is because my secre...
Error The 'ClientId' option must be provided. When building a docker image
I see this question as "how do I persist Docker data on a USB device?".On your machines, you need to mount your USB device into/var/docker. And then restart your Docker service.However, with this solution, when you unplug the USB device, all of the containers have to be stopped. Otherwise,data will be lost.
Suppose we have the Linux OS installed on two identical machines that supports the version of latest Docker. Then suppose we build a container image based on this OS. We can assume this image will now run on either machine. We now put this image onto a USB drive and plug it in the other identical machine.Now, the hard ...
How do I run a Docker container on a USB drive?
http://cert.startcom.org/offers free certificates and they work with all modern browsers.
Closed.This question is seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. It does not meetStack Overflow guidelines. It is not currently accepting answers.We don’t allow questions seeking recommendations for software libraries, tutorials, tools, books, or other off-si...
Where to find a Free Open Source SSL for IIS, and how to install it [closed]
It will not do you any harm, put it that way. But a_retained_property = newValue; is not equivalent to: [a_retained_property release]; a_retained_property = [newValue retain]; Instead you should use: self.a_retained_property = newValue;
I have many viewControllers. and all of them have many retained properties. so, I would add these code to every viewControllers: - (void)viewDidUnload { self.every_retained_property = nil; [super viewDidUnload]; } - (void)dealloc { [every_retained_property release]; [super dealloc]; } Till now things ...
A little bit wondering of memory management on ios
Ifbrowser is not caching the page there's no much you can do than using one of them:<meta http-equiv="Cache-control" content="public">OR<meta http-equiv="Cache-control" content="private">Youcan'tprogramatically override browser settings. It would be a security issue!ShareFolloweditedSep 27, 2013 at 20:41answeredSep 27,...
Someone know how to force a page cache in the browser? I need to reload a page several times, and the browser isn't caching the page!
How to force the browser to cache a page .html?
After contacting 123-Reg (my hosting provider), they submitted this solution, which works perfectly:RewriteEngine on RewriteCond %{HTTP_HOST} ^example.com [NC] RewriteRule ^(.*)$ http://www.example.com/$1 [L,R=301,NC] RewriteCond %{ENV:HTTPS} !=on RewriteRule ^.*$ https://%{SERVER_NAME}%{REQUEST_URI} [R,L]Basically t...
I have made a .htaccess file to redirect all website traffic tohttps://www..This is my complete .htaccess file:RewriteEngine On RewriteCond %{HTTP_HOST} !^www\.example\.com [NC] RewriteRule ^(.*)$ https://www.example.com/$1 [R=301,L]The below redirects work exactly as expected:http://example.com -> https://www.example....
.htaccess file not redirecting http://www. to https://www
Below rule will rewrite to the location of pdf, showing of pdf will depend upon your sethandler rule in apache config.I am assuming root is the location.RewriteEngine On RewriteRule ^test$ /file.pdf [L]
I want a rewrite rule in my htaccess file that if someone callswww.domain.de/testa specified PDF is openedIs that possible?
specify a .htaccess rewrite rule to open a pdf file
Varnish is a Caching Reverse Proxy while CloudFront is a Content Delivery Network. My suggestion: If you are looking at serving content all over the world, go with CloudFront. If you're just serving to a local user base, simply fire up a high memory instance and install Varnish and serve content off Varnish. This take...
Recently came across an article on using AWS CloudFront to deliver the whole site, instead of using Varnish as reverse proxy. Is it really better than Varnish, especially in case of Magento?
Replacing varnish with aws cloudfront
In your Compose file, you have avolumes:block that overwrites the image's code with content from the host. Delete this.services: backend: volumes: # <-- delete - ./backend:/app # <-- deleteWhen this block is present, the/appdirectory in the container is the./backenddirectory in the host system....
I am implementing a application using a NestJS server, working with a PostgreSQL database and the Prisma service to handle data.I have an issue when trying to run a Prisma migration when launching my service in my Dockerfile. Here is my docker-compose.yml file :version: '3.8' services: # POSTGRES postgres: con...
Prisma migration in a Docker - NestJS server
2 You can keep using require "github.com/y" in your go.mod require, and add below a replace statement like: require ( ... github.com/y version ... ) replace github.com/y version => github.com/x your-fork-version Share Follow ...
This question already has an answer here: go get on forked github repo got "unexpected module path" error (1 answer) Closed 3 years ago. I forked a go project and used go get to i...
How to fork a go project? [duplicate]
It is not possible to deploy 2 different Docker images to Cloud Run. The Cloud Run works in the following way: You build a container image and upload to Google Container Registry Deploy to Cloud Run with the container image. Your service is automatically scaled up and down to a specific number of container instance...
I've built an app that uses two home made micro services, each of the micro service having its own Dockerfile. When I build it locally I use docker-compose for practical reasons. Currently, when I deploy to Cloud Run I use commands like docker tag xxx docker push xxx Then I select the image I want to deploy on Cloud ...
Is it possible to deploy two different docker images within the same Cloud Run service
There are different ways to troubleshoot this problem. This is something addressed in the Developer Forum of AWS. Please go through the following steps to troubleshoot this problem. These could be basic steps, but I am pointing out the most general steps required. Try sending an SMS from the AWS Console. If this work...
I have been using Amazon SNS to send SMS for a while, but two days ago it stopped working. I didn't change anything on my side. Even if I try to send SMS from the console, it still doesn't work. I only have 35% delivery rate for last couple of days, but since yesterday it is not working at all.
AWS SNS is not sending SMS anymore
http://gitref.org is great for teaching you the basics of git. As for workflow, if you're coming from SVN all I can say is branches. Branch your ass off. Every time you start a new bug, feature, whatever, start a new branch off of the current master. Don't work in master directly. It's much easier to do merges tha...
I've worked almost a year as a developer, and during that time I've worked only with Subversion as version control. Now I'm in a 2-man project where we just set up a GitHub repository, and I'm looking for an easy workflow just to get started - until I have had the time to learn all the small nuances of Git that I beli...
Simple GitHub workflow to get started with
I had in fact two problems :1) My cron was executed with php.ini for php-cli. I used the -c flag to load the good one.2) I was relying on $_SERVER to declare important constants using variables that do not exist in cli-mode. Now if these variables are not set I parse commande line vars withgetopt()
I'm using crontab to call a php script.In this script there iserror_log('test');When the script is executed from http or direct command line likephp -f script.phpEverything is fine, my error is log.But when called from cron it's not working.Here is my cron* * * * * -u www-data /full_path_to/php -f /full_path_to/script....
PHP not working with CRON
This can happen if you are getting a Root CA from the peer side and that Root CA is not loaded on the your side.ShareFollowansweredMay 18, 2012 at 12:52JayJay24.5k2525 gold badges9797 silver badges141141 bronze badgesAdd a comment|
I am using this code:void check_cert(SSL* ssl) { //ssl initiation SSL_library_init(); SSL_load_error_strings(); const SSL_METHOD *meth; meth = SSLv3_method(); SSL_CTX *ctx; SSL *_ssl; ctx = SSL_CTX_new(meth); int result = 0; //getting the CA certificate result = SSL_CTX_load_verify_locations(ctx1, "cacert.pem", NULL)...
SSL Certificate verification fails: Using OpenSSL
For the java client you specify the endpoint when setting up the servlet, seehttps://github.com/RobustPerception/java_examples/blob/master/java_simple/src/main/java/io/robustperception/java_examples/JavaSimple.java#L39for example. You can change the endpoint to whatever you like.
I am trying to expose some metrics from a spring-boot service to prometheus. Unfortunately, both spring-boot actuator and the prometheus simple-client expose their metrics through the/metricsendpoint.How can the endpoint of the simple-client be changed?Thanks
How to change the endpoint of prometheus metrics
Seems like you need to modify File.Copy line: var targetPath = Path.Combine(TargetDir, file.Name) File.Copy(file.FullName, targetPath , true); I changed first argument from file.Name to file.FullName - this should fix the exception
namespace Backup public partial class Form1 : Form { public Form1() { InitializeComponent(); } private void btn_Backup_Click(object sender, EventArgs e) { List<DirectoryInfo> SourceDir = this.lbox_Sources.Items.Cast<DirectoryInfo>().ToList(); string TargetDir = this.tbox_T...
FileNotFoundException but file exists C#
Yes, The DependsOn attribute can take a single string or list of strings. http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-attribute-dependson.html Syntax: "DependsOn" : [ String, ... ]
Given i have this example template: { "AWSTemplateFormatVersion" : "2010-09-09", "Mappings" : { "RegionMap" : { "us-west-1" : { "AMI" : "ami-655a0a20" }, ... } }, "Resources" : { "Ec2Instance" : { "Type" : "AWS::EC2::Instance", "Pr...
AWS: Cloud Formation: Is it possible to use multiple "DependsOn"?
You need to add startup command, you can try it. npx serve -s Or pm2 serve /home/site/wwwroot --no-daemon --spa Related Posts. 1. Default Documents (custom 404) on Azure AppService Linux website 2. pm2 not found on Azure App service with Node 14 runtime 3. linux azure web app not showing my reactjs app
I'm a newbie to Azure I created a new Vue project using vue create which runs locally and even serving the dist folder too it run successfully. (serve -s dist) And then I deployed the application using GitActions to Azure(Web App Service) which Azure DevOps services indicated that the deployment has been successful: ...
Application is not loading when deployed via azure app service
There are 2 relevant options in/etc/dd-agent/conf.d/docker_daemon.yaml:collect_disk_statsIf you use devicemapper-backed storage (which is default in ECS but not in vanilla Docker or Kubernetes), docker.data.* and docker.metadata.* statistics should do what you are looking for.collect_container_sizeA generic way, using ...
Is there any way to monitor disk usage of docker containers in DataDog?I can see in DataDog web all the CPU, RAM and IO metrics for my containers.But I can't see any of disk space related metrics.Their pagehttps://docs.datadoghq.com/integrations/docker/says about:docker.disk.used (now reported as docker.data.used) dock...
Container disk usage in DataDog
As said previously, one solution is to create a jre directory and copy a JDK in it.
I have download Eclipse MAT and try to launch it on Windows. Executing MemoryAnalyzer.exe display a popup Incompatible JVM with this message: Version 1.8.0_242 of the JVM is not suitable for this product. Version: 11 or greater is required. As said in this doc, I have created a jre directory and copied an OpenJDK 16 w...
Eclipse Memory Analyzer - Incompatible JVM
I suppose your aim is to make it look like all those extracted commits were only in the feature/FT001 branch and the master branch was only initialized. If you have Git 2.23, You can do this by creating an orphan branch: git switch --orphan master This will stage all your files, so then you will have to unstage them ...
I used the git subtree command on an existing repository, in order to extract a subfolder into its own repository. Then, I pushed the extracted commits to its new repository, into the master-branch. However, I should have pushed it into the "feature/FT001"-branch all along and wanted to change everything accordingly: ...
Rename master-branch and create a new, empty master
Do I need to add IDisposable to all these poco classes so they are removed right away? Anything that implements IDisposable should have IDisposable.Dispose() called as soon as the object is no longer needed. Typically the best way do accomplish that is with a using statement. IDisposable is there to allow for expen...
I have a schedule system that generates games into a certain schedule that could be a 1000 games or more. During this alot of processing occurs and object creation. When I run this auto schedule feature my memory footprint for this website goes up significantly. The objects created are plain poco objects and I am u...
.NET Object creation causing memory spike and doesn't go away
git pull or git pull origin feat/feature1.0 should bring the changes in from the remote repo into your local repo. Both of these will pull changes to your local repo, but they would also update your current branch ( assuming you are on the local branch feat/feature1.0) with the changes made to the remote repo version ...
I created a branch. Let’s call it feat/feature1.0. My friend created a branch off of this branch. Let’s call it feat/feature1.1. My friend has committed and merged their changes into the original branch (feat/feature1.0). I now want to pull the changes back to my local repository. How do I do this? I’ve tried a git pu...
How do I pull merged changes from a sub-branch back onto the original branch locally?
You are not going to get accurate values (in the sense you expect) from Prometheus.The immediate problem is that PromQL'sincrease()andrate()use extrapolation instead of actual available data. Seethis Prometheus issuefor details. There are a number of workarounds for that, none of which are ideal. (E.g. you can take the...
I'm trying to query the custom counter promethues metric but can't seem to find the right query. I would like to query the number of occurances in the last minute. For this I triedsum(increase(my_counter[1m]))as well assum(rate(my_counter[1m])). Neither seemed to be giving the accurate values. How can I query them inst...
Can't query accurate details with promethues custom counter
Make sure you have configured theGitHub page publishing source.If it is still set to None, you won't see anything:
Whenever I publish my website on github, only the README thing pops up, not the actual HTML file that I had uploadedHere take a look at my repo:https://github.com/arib0903/aditto.github.io
why does only READ me pop up on published github page
You could just use anotcondition to exclude port 80# Log TCP traffic to x.x.x.x for all destination ports except 80 iptables -A OUTPUT -p tcp -d x.x.x.x ! --dport 80 -j LOG
I'm new to iptables and Linux-firewall in general. Can somebody help me with it? I want to write a table using iptables that will log outgoing TCP connections to a specific IP address, except these that go through port 80 (HTTP).Her is what I have so far:iptables -N LOGGINGiptables -A OUTPUT -j LOGGINGiptables -A LOGGI...
How can I log outgoing TCP to IP and not (HTTP) iptables
As we can see, your environment has a custom Docker as a cp layer, that's why UI Firewall is not available for your cp node group. More details are in Container Firewall Rules Management article. Nevertheless, even if your cp layer is not a custom Docker but Jelastic certified dockerized template, UI Firewall will be ...
I have defined the following Jelastic configuration for my environment: env: topology: nodes: - nodeGroup: bl nodeType: nginx-dockerized tag: 1.14.2 displayName: Node balancing count: 1 fixedCloudlets: 1 cloudlets: 4 env: DOCKER_EXPOSED_POR...
How to close docker environment from the outside world?
3 1, 2, 3: don't refer to memory as registers; although I've seen that done before it is confusing. a 64-bit processor has 8-byte-wide registers and a 32-bit processor has 4-byte-wide registers, but when those registers are copied into RAM it's just bytes in memory. what pa...
I have come under the impression that if a program reserves memory for a variable type integer (which is 8 bytes in size) the memory that is reserved will be 8 different registers that appear sequentially in memory. My questions are as follows: Assuming that answers will vary is my statement above a good generalizati...
How are memory registers used to hold different types?
The only difference is that one specifies the full path to the PHP binary, and the other doesn't.PHP can be installed in different locations. That's why it works on one server, and not the other.
I am very new to php, if anyone can help me by telling what the difference is between these two cron commands, I would be very thankful:/usr/local/bin/php -q /home/delightl/public_html/myfolder/run.phpandphp -q /home/delightl/public_html/myfolder/run.phpI am totally confused.. On my GoDaddy server, the second command w...
What is difference between these two cron jobs?
Assume you have the name of the repo owner as name, and the remote-tracking name of the local branch as branch, you can make a request to the GitHub API: GET https://api.github.com/repos/$name/$repo/pulls Then you go through each object in the returned JSON, check if item["head"]["label"] == "${name}:${branch}". If y...
I want to find the latest pull request associated with a local branch so that I can dynamically build a url to that pull request. My end goal is to build a url like this from a bash script: https://github.com/organization/repo/pull/871 This questions is about how to get pull requests associated with a specific branch....
Git - Is it possible to get pull requests associated with a local branch?
According to the AWS documentation about hosted Direct connect, you can only have sub 1G connection through your ISP.
I need to have a 5 gbps Direct Connect connection to my Amazon VPC from my servers residing at a ISP data-centre.I can't wait for more than 1 week to set it up. Is it possible through a hosted Direct Connect connection to get the 5 gbps?
AWS hosted Direct Connect connect through ISP
I had the same issue. I was finally able to get a container to sit "idle" with the following command inside the Task Definition: "tail", "-F", "/dev/null" Then I could connect in with an interactive execute-command.
I would like to launch an interactive shell into a public Docker image on my AWS ECS/Fargate cluster to run network/connectivity tests from inside the cluster. It seems the official way to do this is with aws ecs run-task followed by aws ecs execute-command [1][2] I'd like to use existing, public Docker Hub images rat...
Interactive shell in Docker image with Amazon ECS with `aws ecs run-task` followed by `aws ecs execute-command`
remove the file from tracking: git rm --cached config-dev.php && git commit -m "config-dev.php" add it to the .gitignore echo config-dev.php >> .gitignore git add .gitignore git commit -m "adding config-dev.php to .gitignore" Publish these changes git push On your colleagues machine fetch the new configuration git...
This question already has answers here: How do I make Git forget about a file that was tracked, but is now in .gitignore? (35 answers) Closed 8 years ago. I have a file config-dev....
How to remove a file from being tracked by git? [duplicate]
Directly login into GitHub using your browser and upload the files through the browser, this will completely eliminate the use of git, now make git ignore all the media files using.gitignorefile. Now only the binaries will be managed by git.Hope this helps!
I have a repository with about 55GB of contents, with binary files that are less than 100MB each (so no LFS mode) from a project which has almost filled up an entire hard drive. I am trying to add all of the contents to a git repo and push it to GitHub but every time I dogit add .in the folder with my contents after in...
Adding large folder to git repository without caching to repository?
Depending on what you're doing with NLTK, I may have found a solution. The base nltk module imports a lot of dependencies, many of which are not used by substantial portions of its feature set. In my use case, I'm only using the nltk.sent_tokenize, which carries no functional dependency on sqlite3 even though sqlite3...
I am building a python 3.6 AWS Lambda deploy package and was facing an issue with SQLite. In my code I am using nltk which has a import sqlite3 in one of the files. Steps taken till now: Deployment package has only python modules that I am using in the root. I get the error: Unable to import module 'my_program': No ...
sqlite3 error on AWS lambda with Python 3
I came up with a solution that leveraged storing a secret in secrets manager as plain text. Store the secret in secrets manager as plain text. They console will have JSON brackets but I removed those. Use the cli to get the secret output as plain text. Now the \n and \s in the text will be converted to the line bre...
EDIT: AS OF Feb 2020, AWS SEEMS TO have FIXED THIS BUG. THE BASE64ing and other wise is no longer needed. I have my secret stored as a string but of course when aws stores the secret it removes white space and line breaks. On top of it it wraps the value in json. When I run aws secretsmanager get-secret-value --secre...
How to normalize a private key stored on AWS secrets manager
There is two option to expose the service outside, you can use the ingress controller to connect to the server.The simple method is change your service type to NodePort, then you should be able access server using NodeIP and service external port number.here is the more info.https://kubernetes.io/docs/concepts/services...
I created an headless service:apiVersion: v1 kind: Service metadata: name: myapp-service-headless spec: ports: - port: 80 selector: app: myapp clusterIP: NoneFrom Kubernetes dashboard I can see itsInternal endpoints:myapp-service-headless:80 TCP myapp-service-headless:0 TCPIn this application, I also se...
How to connect to an headless service in Kubernetes from outside?
1 You can use MemoryFailPoint to check before a large allocation. This isn't exactly what you're asking for, but if your failure is at a particular known large allocation point, this at least allows you to cleanly handle the allocation failure at the point that it occurs. ...
The OutOfMemoryException for my applications is thrown way before the system itself runs out of memory, as explained by: MSDN Blogs - Out of memory? Easy ways to increase the memory available to your program. When you run your C# application, you might get an OutOfMemoryException thrown, even if your machine has lo...
Determine available memory that can be used before OutOfMemoryException is thrown
Is there any easy way of recompiling the code for GPU and running on "many more" GPU cores?NoWhat is the easiest way to port existing boost C++ code for GPU assuming I want to run the threads as-is?There isn't one. As noted in comments, C++17 includesparallel algorithms, and NVIDIA ship aC++17 compiler with CUDA suppor...
I have a very compute intensive single-threaded POSIX+boost C++ program that I just launch in many copies on a 48-core CPU (it's kind-a brute-force A* search algorithm)Is there any easy way of recompiling the code for GPU and running on "many more" GPU cores?What is the easiest way to port existing boost C++ code for G...
Launching many copies of POSIX C++ code on GPU
I am not sure about thedocker-compose.ymlbut thedockerfilethat you have seems to be working for me.TheDockerfilelooks like this,FROM python:3.6-slim RUN mkdir /app WORKDIR / RUN touch /app/modbus.dbBuild the dockerfile,docker build -t test . Sending build context to Docker daemon 2.048kB Step 1/4 : FROM python:3.6-sl...
I want to create an empty DB file usingtouchbyDockerfileordocker-compose.ymland volume it. Actually, I'm able to create it manually within the container as follows:docker exec -it <container-name> bash # touch /app/model/modbus.dbWhereas, when I use the following procedure it throwsexited with code 0and stops:version: ...
How to create a file using touch in Dockerfile or docker-compose?
9 Please take a look at the docker file I just uploaded: https://github.com/bdruemen/jenkins-docker-uid-from-volume/blob/master/Dockerfile . Here the UID is extracted from a mounted volume (host directory), with stat -c '%u' <VOLUME-PATH> Then the UID of the container us...
I am new to docker, so if this is a fairly obvious process that I am missing, I do apologize for the dumb question up front. I am setting up a continuous integration server using the jenkins docker image. I did a docker pull jenkins, and created a user jenkins to allow me to mount the /var/jenkins_home in the contain...
Changing the user's uid in a pre-build docker container (jenkins)
It seems your are using a not supported filesystem for the OverlayFS storage driver. Please, have a look of thesupport filesystem for each storage driverSo, first, retrieve your the filesystem you're using withdf -h.Then, you have 2 options:change the Docker storage-driver from the file/etc/docker/daemon.jsonand use a ...
I've tried to move my Docker's directory from/var/lib/dockerto an external hard drive, which is formatted with NTFS. I've followedthis guide. However, when I dosystemctl start dockerI get an error, and in the journal I find these:Jun 15 11:38:32 lampo.sial kernel: overlayfs: upper fs does not support tmpfile. Jun 15 11...
Error after moving Docker's dir to NTFS: overlayfs: upper fs does not support <xxx>
You need to supply the maximum number of messages to return. Amazon SQS never returns more messages than this value (however, fewer messages might be returned). Valid values are 1 to 10. Default is 1.SeesetMaxNumberOfMessages(Integer maxNumberOfMessages).
ReceiveMessageRequest receiveMessageRequest = new ReceiveMessageRequest(queueURL); List<com.amazonaws.services.sqs.model.Message> messages = sqs.receiveMessage(receiveMessageRequest).getMessages();Whenever I try and pull all of the messages from my SQS queue, the size of the 'messages' list is always 1. How do I ensur...
Why am I only getting one message from my SQS Queue?
DestinationRuleis a resource that adds additional routing policies after routing happens to aService, for example say that you have the following service:apiVersion: v1 kind: Service metadata: name: my-service namespace: default spec: selector: app: my-service ports: - name: http protocol: TCP ...
Can any body explain istio Subsets and destination rules in a a simple manner and explain the problem they are trying to solve by introducing the subsets.
What's Istio Subsets all about
Yes, all variables increase memory usage. This is because references need to store the memory address of what they refer to. However, references use a negligible about of memory (4 bytes on a 32-bit machine, 8 bytes on a 64-bit machine) compared to copying/creating a new list. Please see Ned Batchelder 's article on h...
We have variable in our script which has values like this: result = {abc: [], xyz: []} Now we create another variable like: abc = result['abc'] So what will be the memory usage by this script. Will all the data from result will be copied in abc variable or just reference of that data? Will this increase my memory us...
python copies variable by reference or by value
The IP addresses and hosts that you need to configure/allow are officially documented here:https://learn.microsoft.com/en-us/azure/application-insights/app-insights-ip-addressesI'd copy and paste the "relevant" portions, butthere's a huge number of them depending on what you want/need to do andthen they'd be wrong here...
I'm having issues getting application insights to report data to Visual Studio Online from behind our firewall. I opened the firewall rules noted inthis articlebut it didn't make a difference. I've uninstalled and reinstalled several times. The only thing that is showing up in the Operations Logs is that it's periodica...
Application Insights Behind a Firewall
Have you tried adding the project folder to Atom? "File > Add Project Folder..."That might give reference to the .git folder, anyhow it worked for me :)
I'm using Github desktop for a Unity project that I'm working on with some class mates. I've been using Atom to merge conflicts because it provides a helpful GUI that highlights each conflict, especially because I am a total beginner at using git. However, for some reason Atom is no longer highlighting the marked areas...
Atom's Git integration not working: merge conflict markers not highlighting
Grep for it:grep -lr yourscript /etc /opt/anotherlikleydirfailing that, search the whole system : grep -lr yourscript /Edit:failing that, search in binaries too: grep -lar yourscript /failing that, the script is either executed by a logged in user or a scripted remote login... if that's the case, try peachykeen's appro...
On one of my linux servers I have a script that performs some controls. Is there a way of finding out where this script is called? This can be in another script, cobol program, crontab, ... Opening every one of them will take a very long time.
Is a script called somewhere
1 ARG BUILD_DIR=/build # build # одноразовый контейнер FROM node:14 as build ARG BUILD_DIR There is global and local scope Global can be translated into local of each container Share Improve this answer Follow ...
ARG BUILD_DIR missng after second FROM. What I doing wrong? Dockerfile: # build # одноразовый контейнер FROM node:latest as build ARG BUILD_DIR=/build # создание директории приложения WORKDIR $BUILD_DIR # установка зависимостей # символ астериск ("*") используется для того чтобы по возможности # скопировать оба фай...
docker arg misssing in multi-stage builds
0 Run the command with Administrator rights. 'Sudo' (Linux) or 'Run as Administrator' (Windows). Share Improve this answer Follow answered Oct 18, 2017 at 19:07 Vagner GonVagner Gon 59511 ...
I try to start a new Ionic 3 project with the new dashboard. I do this : Start my project with sudo ionic start MyApp I link my project to the dashboard sudo ionic link ✔ Looking up your apps - done! ? Which app would you like to link Create a new app ? Please enter a name for your new app: MyApp ionic config set a...
Ionic 3 : Configure project fail with github
If I'd need to choose one of your approaches I would probably choose the first approach.But I'd probably change the path here.Instead of using the applications' metrics to decide how many jobs/pods you need I would probably use the queue's metrics.For that, I usedKEDAand I recommend it. Using KEDA can help you scaling ...
I'm completely new to Prometheus and trying to implement an HPA for a use case.Our use-case is application/pod will be processing jobs from a queue asynchronously. Each pod will pull as many jobs as it can and once they have reached a limit it must start to autoscale.To achieve this there are two approaches,Pod will ex...
Gauge or Histogram - Which is a more suitable metric for determining load of a pod?
0 If you copy your git repositories to somewhere else you will have all remote branches and pull requests in there. basically that is all stored in the hidden .git directory which is inside every repository. The reason why you only see the main branch is because that is the...
I have 2 repositories where 1 repository named Web has 4 branches. The issue when is taking backup i see only main branch data is been backed up. How to backup all the branches data of that repository?
Why backhub taking only master branch of repositories?
Since you have private files to clean before being able to change branch, I would recommend fixing the situation in a new clone first:git clone https://[email protected]/username/reponame anewclone cd anewclone git checkout -b development --track origin/development git merge master git checkout master git merge develop...
I have a project with a master branch and a development branch. All changes should be committed in the development branch. I accidentally committed to the master branch and pushed it to Github. This commit is needed in the development branch. Since it is hard to undo the last push I think it might be a good idea to mer...
Git merge accidental commit to development branch
This is actually possible programmatically, without relying on workers to report their identity - I ran into the same problem, and found the following:Sagemaker Ground Truth does automatic logging of worker actions. Among the things it logs is theworkerId, with which you're familiar, thecognito_user_pool_id, and thecog...
I've set up an AWS Sagemaker Ground Truth labeling project and am using a private team for the work. I want to track which member of my team gives each answer.The only user specific information is aworkerIdas seen, for example,here.The sagemaker documentation does not have any information about this ID, nor is it anywh...
AWS Sagemaker Ground Truth WorkerID for private team
The problem is not with AFNetworking but with NSURLCache. By default NSURLCache will not cache files bigger then 10% (not sure what the exact percentage is) of the cache size.But increase the cache size will help:[[NSURLCache sharedURLCache] setMemoryCapacity:(20*1024*1024)]; [[NSURLCache sharedURLCache] setDiskCapacit...
I'm using AFNetworking to download some images from the internet to my app. I'm using this code to download those images,AFHTTPRequestOperation *requestOperation = [[AFHTTPRequestOperation alloc] initWithRequest:[NSURLRequest requestWithURL:[NSURL URLWithString:_linkString[indexPath.item]]]]; requestOperation.responseS...
Large Image Caching from Afnetworking
I assume crontab's cwd (Current Working Directory) is not same as where the script is stored.this would solve your problem:import os script_dir = os.path.dirname(os.path.realpath(__file__)) os.chdir(script_dir)You can get the directory where you script is by calling "os.path.dirname(os.path.realpath(file))"if you chang...
I'm totally lost here. I'm trying to create a scheduler to run python script on my Mac, but I'm getting the following error:Traceback (most recent call last): File "/Users/Root/Desktop/Project/Data/script.py", line 148, in <module> run( File "/Users/Root/Desktop/Project/Data/script.py", line 121, in run con...
FileNotFoundError while using crontab to run Python script
1 Yes, it's theoretically possible for fragmentation to cause out-of-memory exceptions. Suppose you do lots of allocations of small objects that mostly fill your memory, then you delete every other object. This will produce a large total amount of free memory, but they'll a...
Closed. This question needs to be more focused. It is not currently accepting answers. Want to improve this question? Update the question so it focuses on one problem only by editing this post. Closed 5 years ago. ...
Does memory fragmentation leads to out of memory exception? [closed]
Put this code in yourDOCUMENT_ROOT/.htaccessfile:RewriteEngine On RewriteCond %{THE_REQUEST} \s/+([^?]*)\?lang=[^\s&]+ [NC] RewriteRule ^ /%1? [R=301,NE,L]
Sorry, but I couldn't find any help for my problem. Hoping you guys can help.I have the url:http://example.com/many-pages?lang=xxI would like to turn that into just:http://example.com/many-pagesStripping away the "?lang=xx" where "xx" is the language code. Also make it a 301 redirect.
RegEx htaccess remove variable
It depends what you mean by "generated". If you really meant "read", then the first step would be to be either to look into the TLB if the address has already been translated or, if your cache supports virtual addresses, to look in the cache itself to see if there is an entry corresponding to that virtual address (an...
I am having a doubt regarding memory management in operating systems.I know that cache is a temporary storage location used to speed up memory accesses whereas TLB is used to speed up translation from virtual address into physical address. Now if a virtual memory address is generated,what is the first step taken? If ...
TLB translation vs cache
Use backreferences for this one:RewriteRule ^foo/(.*)$ bar/$1ShareFollowansweredJan 5, 2010 at 20:02Ignacio Vazquez-AbramsIgnacio Vazquez-Abrams786k155155 gold badges1.4k1.4k silver badges1.4k1.4k bronze badgesAdd a comment|
We simply have to move every page inside of a directory called "Music" to a directory called "Information"... That's the only change. There's lots of links to pages music and we don't want to break them all. I'm not great at rewrite conditions....Basically, all I want to do is when a user types in music/index.php, for ...
.htaccess redirect all files in one folder to exact same in another folder
0 Based on InfluxDB Hardware sizing guidelines, in moderate load situation with a single node InfluxDB deployed on a server with these specifications: CPU:6 cores and RAM:8-32 GB; you can have 250k writes per second and about 25 queries per second. These numbers will defini...
I develop my metrics based on influxdb. I want to keep the data forever therefore my retention policy is set to inf and my shard retention policy is set to 100 years (the max I could set). My main concern has to do with degrading performance by keeping this data. My series will not be more than 100000 (as adviced for ...
InfluxDB max available expiration and performance concerns
One of the benefits of using a WAF like ModSecurity which is integrated with your web server is that it doesn’t require any special handling for HTTPS. Just configure HTTPS in Apache (they are many tutorials online on how to do this) and ModSecurity can still read the traffic the same as it does for HTTP traffic.
Mod-security_SSL_CertificationHello evryone I'm setting up a new web server for attacks testing (dvwa) in my Centos machine and i want to set up an application security layer usingMod_Securitymodule on Apache and its working great, now i want to change security fromhttptohttpsusingSSLcertification and I don't know how ...
modsecurity ssl certification project
To do the job and allow from IP without password prompt, and also allow from any address with password prompt, it work perfect like this :Order deny,allow Deny from all AuthName "password please" AuthUserFile /home/accountpath/.htpasswd AuthType Basic Require valid-user Allow from 192.168.1.1 Satisfy Any
Using the following.htaccess, I am able to successfully limit access to users with the correct username/password and users from a specific IP address:AuthType Basic AuthName "Please enter your password" AuthUserFile /path/to/.htpasswd Require valid-user Order deny,allow Deny from all Allow from 123.123.123.123 Satisf...
Whitelisting IPs with .htaccess / .htpasswd
IP Firewall is the right way to do this. Is there a reason why you don't want to use IP Firewall?ShareFollowansweredJun 18, 2020 at 14:43Mark BrownMark Brown8,31722 gold badges1717 silver badges2121 bronze badges11Thanks for your answer. I was trying to avoid an extra step to mention ips in firewall. I have already use...
I want to access cosmos db through specific ip address only. The only way I can see to achieve this is, specifying that ip address in 'Firewall and Virtual Network' section of azure cosmos db. Is there any other way to achieve this without specifying ip address in firewall ?
Access to azure cosmos db through specific public ip address only
I have fixed the problem using the solution availablehereRE_XML_ILLEGAL = u'([\u0000-\u0008\u000b-\u000c\u000e-\u001f\ufffe-\uffff])' + \ u'|' + \ u'([%s-%s][^%s-%s])|([^%s-%s][%s-%s])|([%s-%s]$)|(^[%s-%s])' % \ (unichr(0xd800),unichr(0xdbff),unichr(0xdc00),unichr(0xd...
While trying to post the data extracted from a pdf file to a amazon cloud search domain for indexing, the indexing failed due to invalid chars in the data.How can i remove these invalid charecters before posting to the search end point?I tried escaping and replacing the chars, but didn't work.
Removing invalid characters from amazon cloud search sdf
The math functions in Data Factory are add, mul, div, etc. You cannot use the *,+,/ literals. Here is the converted expression@concat( 'RANGE:', string(add(1,mul(sub(1,1),div(int(variables('totalcount')),20)))), ':', variables('totalcount'), ':50' )which gives the result "RANGE:1:18000:50". You d...
Trying to pass sum or multiply number to subtract from date dynamically in blue pipeline as below:@concat( 'RANGE:', 1+((1-1)*(variables('totalcount')/20)), ':', variables('totalcount'), ':50' )The above expression says Unrecognized expression: 1+((1-1)*(variables('totalcount')/20))
Pass number function in azure pipeline dinamically
Unfortunately, GitHub does not provide a way to see overall activity across branches in a repository, other than the default and thegh-pagesbranch - more information onhttps://docs.github.com/en/account-and-profile/setting-up-and-managing-your-github-profile/managing-contribution-graphs-on-your-profile/why-are-my-contr...
I would like to see an overview of my activities (e.g., commits in several branches, issues) on GitHub to see in which periods I was busy. I tried the following ways:URLhttps://github.com/MYNAME/REPOSITORY/commits/?author=MYNAME, the problem is it returns only themasterbranch, while I have commits in other branches as ...
Get an overview of my activities on GitHub
I've found a solution, what I did to work around it is instead of running the action on closing the PR I'm running it on a new commit on master, this has to be triggered by someone with 'write rights' to the repo, therefore, it has access to the repo secrets. It's a bit harder to check if the commit is a merge commit ...
I've started playing with GitHub actions, but I'm struggling with accessing repository secrets which I pass as env's. My workflow file: name: Invite on: pull_request: branches: [master] types: [closed] jobs: invite: runs-on: ubuntu-latest steps: - name: Hello world action uses: lekt...
GitHub Actions - empty env secrets
3 Because DMS doesn't support Snowflake as destination yet so I think you could use S3 as target then use Snowflake bulkload to load data from S3 https://docs.snowflake.com/en/user-guide/data-load-s3-create-stage.html Snowpipe to do continuous loading. Share ...
I am trying to use AWS DMS Database Migration Service along with Snowflake as a source database. Is there any way I can achieve this ? All I could see options for IBMDB2, MySQL, SQL Server, Amazon Aurora, Oracle, SAP Sybase etc. But not for Snowflake. Can ODBC string for SnowFlake be put in as a source endpoint ? Or a...
Can SnowFlake be used as a source endpoint in Data Migration Service of AWS?
1 As noted in the comment by Inertia, git reset --hard would return your index and working tree to the last committed state. You could also do a git reset HEAD to get your index back and then git checkout [files or folders] if you want to keep any other uncommitted changes....
IS there a way to retrieve back my laravel app files after I typed git rm -r * I just want to delete the files that I just deleted in the directory with git rm data.txt But all of them. Then I mistakenly typed git rm -r * Now most of my app directories are all gone. these are now what's left
Files are all gone. I deleted all the files with git rm -r *
One way or another you must tell boto3 in which region you wish the kms client to be created. This could be done explicitly using the region_name parameter as in: kms = boto3.client('kms', region_name='us-west-2') or you can have a default region associated with your profile in your ~/.aws/config file as in: [defaul...
I have a boto3 client : boto3.client('kms') But it happens on new machines, They open and close dynamically. if endpoint is None: if region_name is None: # Raise a more specific error message that will give # better guidance to the user what needs to happen. raise NoRe...
boto3 client NoRegionError: You must specify a region error only sometimes
Mounts are always directory-based. So having a mount in your yaml file tells k8s to mount the contents of the configMap (which could be one or more files) into the directory.Whatever has been inside of the directory before the mount is gone then.See the official documentation here:https://kubernetes.io/docs/tasks/confi...
I have created a configmap and a pod yaml file .I have tried multiple solutions but none worked for me .kubectl describe cm cf3 Name: cf3 Namespace: default Labels: <none> Annotations: <none> Data ==== index.html: ---- hii im marimmo Events: <nonepod yaml fileapiVersion: v1 kind: Pod metadata: na...
Kubernetes configmap removes all the contents of existing directory
I know this is pretty old but wanted to share my view anyway.For your issue, you might want to try the following annotation: nginx.ingress.kubernetes.io/use-regexPlease remove "session-cookie-path" from annotations as Session Cookie Paths doesn't support regexFor more information, please visit below links:https://kuber...
I have created an ingress controller configuration with following path definitions:paths: - path: (USA)/my-app/(.*)$ ............... - path: (UK)/my-app/(.*)$The problem happening here is when I don't set the following annotation;nginx.ingress.kubernetes.io/session-cookie-pathI get regular expression in INGR...
Ingress session-cookie-path setting regular express in Ingress Session Cookie; resulting in user logout
I would say kind of, but it could still be done better. Specifically, you could do: @property (nonatomic, copy) NSDate* coreWeStart; ...and then: self.coreWeStart = self.start; ...to get the same thing with less code. Also be sure to do self.coreWeStart = nil in dealloc (and self.start = nil too).
is this line of code correct re memory management (re NSDate copy)? I have a class with properties: @property (nonatomic, retain) NSDate* start; @property (nonatomic, retain) NSDate* coreWeStart; Now in the init method, assuming self.start is already set, is this correct re setting the coreWeStart to the same date: ...
is this line of code correct re memory management (re NSDate copy)?
Try using below codeRewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_URI} !=/favicon.ico AddType video/ogg .ogv AddType video/ogg .ogg AddType video/mp4 .mp4 AddType video/webm .webm AddType application/x-shockwave-flash swf
This question already has answers here:mp4 in safari fails with htaccess authentication(6 answers)Closed11 years ago.i m just working on a private page with some videos, which should only be accessible through htaccess. all videos are html5 encoded, and the site works fine for every browser without the htaccess file.a...
Safari html5 video mp4 not playing in htaccess folder, [duplicate]
I thinkAnts Memory Profiler can help you so much here, however it is not free, but they have a trial version "for 14 days or so", give it a shot.
I need to trace objects that are allocated in a request. Is there any tool or command that can help me know how many bytes are allocated and when?
.NET how to measure how many bytes allocated in a request?
It is important to Android that when you generate your self-signed certificate, you mark it as a Certificate Authority in order to empower it to certify certificates — even if only to sign itself and so certify that it is itself.This is done in thebasicConstraintsextension, declaringCA:TRUEinstead of the defaultCA:FALS...
I have generated self signed certificate for my server. Then added it to Android with Settings -> Security -> Install.When I'm trying to connect to my server from the application I'm getting error:java.security.cert.CertPathValidatorException: Trust anchor for certification path not found.As I understand after I've add...
How to trust self signed certificate on Android?
Note: This option if you would like to use the same container withmultiple serviceInstallSupervisordwhich will makes you able to runcrondandpostgresql. TheDockerfilewill be as the following:FROM postgres:alpine RUN apk add --no-cache supervisor RUN mkdir /etc/supervisor.d COPY postgres_cron.ini /etc/supervisor.d/postgr...
I am using the "plain" postgresql:alpine docker image, but have to schedule a database backup daily. I think this is a pretty common task.I created a scriptbackupand stored in the container in/etc/periodic/15min, and made it executable:bash-4.4# ls -l /etc/periodic/15min/ total 4 -rwxr-xr-x 1 root root ...
Cron in postgresql:alpine docker container
As suggested by @Jan Garaj, the solution is notification policies. I used a regex to not send the alerts to my slack channel.ShareFollowansweredJul 27, 2023 at 13:40ClGideClGide4355 bronze badgesAdd a comment|
I created an alert that notifies me when I have unhealhy AWS load balancers. What I want is to use a regex so that I do not get notified when a preproduction load balancer is unhealthy. One idea is to filter on the LoadBalancerName variable.I can do just that when creating a silence.Do you know how I can do that when c...
Grafana Alert conditions
With PhpStorm now havingbetter integration with Docker engine(including Docker for Mac), you can now just do the following (orread this article):Command line:Pull thephpunitDocker image:docker pull phpunit/phpunit(Mac and maybe Windows)Bridge the Docker socket to theAPI_URL:socat -d TCP-LISTEN:2376,range=127.0.0.1/32,r...
I want to configure my PhpStorm IDE to run PHPUnit tests within my Docker container.It seems like I'm restricted to either using a local PHP executable, or one through SSH, as the interpreter for the tests.I could install an SSH service on my PHP container, but it seems like a bit of a hacky solution, and articles onli...
Running PHPUnit within a Docker container with PhpStorm
1 Grab a book on operating systems like "UNIX Internals: The new frontiers". Are you sure you don't mean mm_struct, which is the thing actually describing the address space? In short, all kernel threads see the same memory. Further, entire kernel memory and userspace memory...
One of my friend name Raj, told me that there is only one global page table for kernel processes and they don't use vm_area_struct for there memory. vm_area_struct is used only to keep track of memory of user processes not kernel processes. So I want to know what is the real story? Can someone shed light on which memo...
Is there a vm_area_struct pointer for kernel processes?
Just for someone running in this problem. The problem was that the plugin for importing jacoco xml files has been removed. If you use sonarqube its normally included. So if you can't see the coverage report check that the jacoco plugin has been installed.
My sonarqube server version 8.3.1 has the cobertura plugin enabled for showing cobertura coverage reports. Now I want to report jacoco XML-coverage data (generated with gradle jacoco plugin) and its not working. Are there any known issues between cobertura and jacoco? Or should both mechanisms work on one sonarqube?
Reporting coverage to sonarqube configured with cobertura from gradle
In this answer, I'm assuming by your decisionthis development need to be a separate productthe fork is intended to never find its way back to the source repository.You want toimporta repository into the new projectYou've created aFork. Forks are really meant to supportInner Source workflowfor when you don't have access...
I work in project born under in another repository on DevOps. After some weeks of works we decide that this development need to be a separate product.So, I create a new project in DevOps and from first repository I create a fork to my new project.Now, every time I create a pull request DevOps ask me to push my code not...
Set default Repository for git on devOps after fork
You have 2 choices:Either those projects are really linked together, and you can configure an Ant build script to have a multi-module project that wraps all your projects =>http://docs.codehaus.org/display/SONAR/Analyse+with+Ant+Task#AnalysewithAntTask-AnalysemultimodulesprojectOr those projects are different (=> they ...
I have some projects and i wrote an ant script to run sonar with this projects.Its okay so far. But i need to show these projects under a top level project like sub projects.I am using just Ant to run sonar and i just working on pure code not on binaries.(I just need to analysis)I could not find how i can solve this.
SONAR - Ant Script for sub Project
There is a difference between -Xmx and -Xms. Xmx reserves virtual memory (which you have plenty of). So for example some code like this (same numbers on Mac OSX and a few versions of linux): public class DeleteMe { public static void main(String[] args) { LockSupport.parkNanos(TimeUnit.SECONDS.toNanos(1));...
This question already has answers here: Do the -Xms and -Xmx flags reserve the machine's resources? (3 answers) Closed 2 years ago. Beside the discussion about huge heap size GC ch...
Is JVM immediately grab whole the Xms space for heap in huge cases like -Xms512G? [duplicate]
As explained inthis page of the doc, you can use the--targetoption ofdocker buildto create an image corresponding to an intermediate stage of the build.Using yourDockerfile, this would amount to the following commands:docker build --target build -t my-binary-test . docker build -t my-binary . docker run -ti my-binary-t...
ScenarioMultistage builds combine multiple Dockerfile sections into a single oneIntermediary and final stages can copy files from the upper stagesFinal stage is suggested to have only the binaries needed.With those in mind, I'd like to build the followFROM fat-base as build RUN setup unit test frameworks RUN setup inte...
Docker: How to use multistage images after build finishes
Do this instead:ID=$(docker run -itd image) && docker exec -it $ID bashBecausexargsexecutes it arguments without allocating a new tty.
In my development, I find myself issuing adocker runcommand followed by adocker execcommand on the resulting container ID quite frequently. It's a little annoying to have to copy/paste the container ID between commands, so I was trying to pipe the container ID into mydocker execcommand.Here's my example command.docker ...
Piping docker run container ID to docker exec
I would like to give you another approach to achieve what you want.You can usenginx-ingressto expose your applications and also restrict the access to your services using awhitelistannotation with IP CIDR you want.But, to redirect non-whitelist ips to a default service, you should use another annotation nameddefault-ba...
to be more specific I need help understanding how to create a Google Cloud firewall rule in order to allow a specific environment to be accessed from a specific IP address and everything else from public internet access on the same port.In my case, I have a GKE within there I have many pods which are accessed through s...
Google cloud filtered application on port 443 (firewall priority problem)
Using Docker Desktop and Kubernetes on Linux Ubuntu, I got the same error, but also with Docker Desktop being unable to start normally becauseI already had a Docker Installationon my machine, resulting in thedocker contextbeing set to the default Docker environmentinstead of the required Docker Desktop.Confirm the foll...
I want to run Kubernetes locally so I can try it out on a small Java program I have written.I installed WSL2 on my Windows 11 laptop, Docker Desktop, and enabled Kubernetes in Docker Settings.There are a number of SO questions with the same error but I do not see any of them regarding Windows 11 and Docker Desktop.I op...
Docker Desktop on Windows11: kubectl get pods gives error on my laptop: The connection to the server 127.0.0.1:49994 was refused
Its because system files are hidden on apache servers... Either select the option to see hidden files if you're using a GUI or type the command "ls -a" if you're on terminal and you should see the files. Any file that starts with a dot is going to be hidden by default. Your .htaccess files fall within the same catego...
Whenever I create an.htaccessfile in a directory it disappears. I am running a VPS at Digital Ocean (I have full control over the server). So I can't see why my.htaccessfiles are automatically deleted upon creation. I even tried to make the file on my computer and just transfer it to the directory via FTP but as soon a...
Creating .htaccess files (Ubuntu Server 12.04 w/ Apache2)
You can undo the changes of just the merge by doinggit reset --merge
I was working in a branch (SRC) and had a bunch of uncommitted changes I was going to move over to another branch (DEST). I was in SRC and did "git merge DEST." Forgot to commit my changes before. It would appear as if days of work are lost. Any help would be greatly appreciated. Any way to recover my changes?
Recover uncommitted changes after a mistaken merge
To verify that the problem isn't with the server, try running awget http://localhost/from your instance and see if you get the page you're expecting. If you do, it's probably firewall related.When you created your instance, you had to provide a security key (this is what you use to ssh into it) and a security group.Yo...
I'm trying to figure out how to set up Nginx on a micro-instance on AWS.The micro-instance is set up and running. I've associated an elastic IP to it. I SSH into it, make a new user runapt-get upgrade, thenapt-get install nginx. Nginx gets installed. Then I runsudo service nginx start. It looks like nginx is started. I...
Setting up Nginx on a new micro-instance on AWS
First, I should switch to the master branch. cd otherrepo/ git checkout master After making changes to the submodule repository, I should push these changes first. cd otherrepo/ ** making changes – Editing README.md file ** git add README.md git commit -m 'editing README.md' git push origin master Then, push the mai...
I have a repository that has a submodule repo, I want to run the workflow of the main repo in case of any submodule repo changes. Any suggestions ?
Automatically run the workflow on the main repo when a submodule repo changes
1 My usual rule of thumb is, for persisted data use disk cache & for quick, dirty & light data, use memory cache (or your lru cache). Be careful with storing bitmaps in memory cache, at least for android devices < 2.3.3, I believe. The pixel data of the bitmap is actual...
I have an LRUCache that I use to cache thumbnails and other such small bitmaps. Currently, I'm using it in my main activity, however, I'd like to use it my other activities too. This makes me wonder if it's a good idea to simply store this LRUCache object in my custom application singleton (which extends Application) ...
Is it a good idea to store LRUCache object in an Application singleton?