Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
It seems like latest ecs-cli version does support load balancer configuration with service up. Have you tried providing --target-group-arn option?. Assuming you have already created ALB and Target Group to associate ECS service. Here is sample command I just tested. ecs-cli compose --file docker-compose.yaml --project...
I am trying to use ecs-cli compose to manage my services and tasks on Amazon ECS. I'm unable to find a way using the service up command to create a new service with an application load balancer (even when that load balancer already exists). This seems possible with service create, but the API is different from the s...
ecs-cli compose service up with a load balancer
%{REQUEST_URI}includes the leading slash, so this should work:RewriteCond %{HTTPS} on RewriteCond %{REQUEST_URI} ^/car-sharing.html RewriteRule ^ http://www.example.com%{REQUEST_URI} [R=301,L,QSA]
I want to remove secure http only for a certain url:https://www.example.com/car-sharing.htmlsupposed to redirect tohttp://www.example.com/car-sharing.htmlI tried several .htaccess directives, for exampleRewriteCond %{HTTPS} on RewriteCond $1 ^(car-sharing\.html) RewriteRule (.*) http://%{HTTP_HOST}%$1 [R=301,L]orRewrit...
Remove https (ssl) from certain url only
I was able to resolve the issue by combining the echo example fromhttps://github.com/tokio-rs/tlsandhttps://github.com/snapview/tokio-tungstenite. This addressed both the http and tls requirements.ShareFollowansweredAug 25, 2020 at 13:15hopkinsjj9hopkinsjj91133 bronze badgesAdd a comment|
I upgraded my react front end to use https. That broke the connection to my backend tokio socket server. I tried to run a test against the tokio-native-tls echo.rs example which uses an identity.p12 file after I ran these commands:openssl pkcs12 -in identity.p12 -nocerts -out .conf/server.key openssl pkcs12 -in identit...
Connecting to tokio-native-tls echo.rs from react
CR only runs single containers. Pods can hold more than one container. A pod is a shared-namespace and administrative domain that is intended to hold a single 'main' service in a single container with a number of supporting containers running alongside it. The pod is a little more than a virtual construct though, as it...
I think both CR and Pod run in a Kubernetes Node which can be a physical machine or a VM. Knowing that pod is a shared namespace for container and Container Runtime runs container does this means that Container Runtime runs pod which then runs a container or?I'm just trying to picture these two and understand how they ...
Does Kubernetes pods run inside container runtime?
You can actually simplify it further by using this negative condition rule:RewriteRule !^domains/ domains/%{HTTP_HOST}%{REQUEST_URI} [L,NC]
I have problems with choosing a directory by HTTP hostname.I have a file structure like this:. +-- .htaccess +-- domains +-- 123456.a12.whatever.net | +-- .htaccess | +-- ... +-- example.org | +-- .htaccess | +-- ... +-- subdomain.example.org +-- index.htmlThe.htaccessfile in...
Apache RewriteRule is not working with single input
-1 Let me try that. Even if it doesn't work, we will solve this in further steps. On the contrary, try allowing all traffic within the internal subnet. Do a docker network ls and see the list of networks docker is using and the one we care about specifically in this proble...
I am struggling with opening a port, 2377 for docker swarm connectivity. Mac OSX specification: System Version: macOS 10.15.2 (19C57) Kernel Version: Darwin 19.2.0 There are lot of ways that I tried to open the port: Enabled the firewall using System Preferences->Security and Privacy-> Firewall. Then edited /etc/pf....
How to open/enable port in Mac OSX, port 2377 for docker swarm
Your main problem is that your result is not a time series. All Grafana panels/datasources are designed for time series. You may convert it to time series (in container, lambda function, proxy, ...). Or create text panel inhtmlmode and code all your business logic (REST call, parsing, ...) injavascript. Seeexample. Jus...
I know about the answers to JSON data source questionslike this one.It requires an implementation where I would need to run a container which I can not do in my current environment.I have a jira query like/rest/api/latest/search?jql=project%20%3D%20MyProject%20AND%20status%20in%20(Closed)&maxResults=0where a typical re...
Grafana parse HTTP get JSON result as source
If I've understood your task clearly, it'sParallel Processing with Kubernetesyou're looking for. With this feature of K8S, you can run a certain job with multiple pods running parallelly and those pods are terminated when the job is done.You can read more from the following documentation links -https://kubernetes.io/do...
I am looking for some advice, and I would be eternally grateful if anyone would be able to point me in the right direction.I have a docker container that I use to do machine learning based object detection/tracking across sets of video frames. Currently, I start up an ec2 instance with this docker container, and then s...
Kubernetes (GKE/AWS/Azure) Scaling for Large Jobs
For access check you can useifdirective and ssl module variables:$ssl_client_s_dn,$ssl_client_serial. Examplelocation /not/for/jhon/ { if ($ssl_client_s_dn ~ Jhon) { return 403; } }Good way to maintain list of allowed certificated themapdirective. Examplemap $ssl_client_s_dn $ssl_access { default 0; 01 1; 0...
I have SSL enabled in nginx with the client certificate enabled in my browser. With this I'm able to hit my site via HTTPS through port 443.What I'm looking for now is to use this information about the client to allow access to different parts of the API (URLs) but deny access to other parts. I can do this using IP a...
nginx authorization based on client certificates
0 Use docker image prune to remove unused images. https://docs.docker.com/engine/reference/commandline/image_prune/ Share Improve this answer Follow answered May 5, 2017 at 12:38 kstromeira...
I have just installed Docker for Windows and I am running on Windows Container mode. I downloaded this image microsoft/mssql-server-windows-express, which is pretty sizable, and the download gets interrupted halfway. Where should I look for half-downloaded image and what is the best way to clean it. docker image --all...
Cleaning half-downloaded docker image
For that you need to use Image loader library for loading images. You can use : Universal Image loader Picasso Glide This library uses for bitmap memory management and after using this library you didn't face OutOfMemoryError exeption.
In some phones (especially Samsungs), the application gives an error of OutOfMemory. I use a ViewFlipper with 4 ImageViews which they are shown in a tutorial. I have found no way to remedy the error or recycle imageviews. This is the method that manages the ViewFlipper: private void changeBackground(int position) { ...
Recycle ImageViews to avoid OutOfMemoryError
That is an invalid configuration. Remove the global section and try:receivers: - name: "test_api" webhook_configs: send_resolved: true url: 'http://localhost:9000/test/test_host'
I am using the following configuration for alertmanager for sending API post call for prometheus but somehow I am not able to get it, below is my configurationwith my flask server running on localhost at port 9000, that accepts a POST API callglobal: http_config: 'http://localhost:9000' route: group_by: ['alertnam...
how to use web_hook for doing POST Rest API to a server from promrtheus alertmanager
There is no property by the namesonar.exclusion.sonar.exclusionsisa valid property, but it applies tosourcefiles. You're trying to excludetestfiles - and yes the scannersdomake the distinction, especially for Maven projects.You should use insteadsonar.test.exclusionsIf you want to omit only certain rules, you have two ...
I'm trying not to analyse test files with Sonarqube.I have several Maven subprojects and the test files are under these paths:subproject1/src/test/,subproject2/src/test/and so onI'm passing the following option from Maven:-Dsonar.exclusion=src/test/**However, the test files are still analysed.I also tried:-Dsonar.exclu...
Excluding files from Sonarqube
I currently don't have IntelliJ installed, but you might want do push it from your terminal as a second option. You can use the steps below: After creating your new repository on git, in the directory you're working on, run git init. Which should "Initialized empty git repository in ....git". Run git add . This shoul...
I created 2 repos by github MacOS desktop in my project, using my new github account. But when I am trying to push commits through IntelliJ, there is a 403 error, but I can push successfully in github Desktop. remote: Permission to Auird/OJ.git denied to auicheng. fatal: unable to access 'https://github.com/Auird/O...
Github push failed in IntelliJ
You can set this here:https://aws.amazon.com/blogs/aws/s3-lifecycle-management-update-support-for-multipart-uploads-and-delete-markers/and say afer 7 days or so, just delete them.
Usingmultipart uploads, Amazon S3 retains all the parts until the upload is either completed or aborted. In an anonymous drop situation, it would be good for abandoned uploads to be automatically aborted after a timeout to reclaim the space and avoid the cost of holding any parts that made it.It would be possible to cr...
Does AWS S3 automatically abort multipart uploads after a timeout?
1 You can use git config command to that $ git config --global user.name "P.Mihiran" From the above command, it sets the user name as you mentioned. If you want to check the username you can use the following command to check user name git config --global user.name For mo...
This question already has answers here: Git for beginners: The definitive practical guide (37 answers) Closed 4 years ago. I am a new user to git hub. I just wanted to configure my...
How to do git hub configuration? [duplicate]
Rewriting my comment as an answer as it seems to be a valid workaround for this issue:You can add this annotation to the pods to disable probes on the istio-proxy containers:status.sidecar.istio.io/port: "0"The issue is being discussed on github, with more information:https://github.com/istio/istio/issues/9504#issuecom...
I want to disable istio readiness probe from doing health check.. on running service. Is there a way to disable http readiness probe and later enable it
How to disable istio readiness probe
38 in cron jobs, here is how you should pass the argument /usr/bin/php -q /home/**/public_html/scores.php date=12/05/2009 *take note there is no "?" Share Improve this answer Follow answered D...
Is it correct to use the following command in a cron job: /usr/bin/php -q /home/**/public_html/scores.php?date=12/05/2009 I haven't found any supportive article / material to answer it, hence i am putting forth this question to the community. So the question is is there a way for me to include a variable in a cron j...
Cron Jobs calling a PHP script with variables
The hash of an object is completely determined by the object's content. Find the root commit(s) of the two repositories and inspect their data (git cat-file -p). Something will be different. All other differences stem from there. Otherwise, the histories match up to some point, at which point there is something dif...
I have a clone of CPython on my own gitlab server from a few years ago. Total I just tried to pull/merge the latest from the the official github repository into my gitlab's repo, but it failed due to no common history. All of the commits are there, some authors, commit messages dates etc. However all commit hashes are...
Same git repository, same history, different commit hashes?
Adduse kartik\editable\Editable;to the top of the view. It appearse theEditableview widget extends the base helper widget. Since Yii2 does not treat views as classes, and thus no namespace, use of classes have to be declared at the beginning / inline of the view's logic.
I'm new in Yii 2! I install the kartik grid in my basic yii and create a grid with no problem . now I want to usekartik\grid\EditableColumnbut it needs :\kartik\editable\Editablei insatll the Editable via composer and now i cas see that directory in my vendor file :my vendor filenow the system gives me this Error again...
install kartik-v editable in Yii2
To answer your questions:Yes, Sonar Runner will work on your local network as it does not need to download things from the Internet - only from the Sonar Web server instance (which seems to be on your intranet).If your code is compiled before launching the Sonar analysis with the Sonar Runner, you'll be able to get all...
I need to run Sonar behind a firewall in a company that provides me with Sonar itself but does not allow downloads from the Internet (automated or not). This makes use of Sonar on maven problematic, because maven needs to contact its online repository. (There is an effort to create a corporate mirror of the repositor...
When running Sonar, does Sonar Runner "phone home" or sacrifice functionality?
Your syntax is incorrect. Please use the following code#every minute * * * * * wget -O - -q "http://example.com/cron/test1.php">/dev/null 2>&1 #every 15 minutes */15 * * * * wget -O - -q "http://example.com/cron/test2.php">/dev/null 2>&1You can use online crontab generators likehttp://www.crontab-generator.org/
In my cron job file I have two cronjobs defined:#Yo1 MAILTO="[email protected]" *1****wget -O - -q "http://example.com/cron/test1.php">/dev/null 2>&1 #Yo1 MAILTO="[email protected]" *15****wget -O - -q "http://example.com/cron/test2.php">/dev/null 2>&1The PHP files are simple just sending mails with different subjects....
Cron Jobs Run At Same Time
I don't know a better way. You can create your own attribute ForeverOutputCache, which will set Duration = Int32.MaxValue in constructor, but this is almost same approach.
What would be the best way to cache an actionresult forever, that is untill the application is restarted. // Cache this result forever, it will reset on app restart [OutputCache(Duration=999999)] [ChildActionOnly] public ActionResult Footer() { if (HttpContext.Application != null && HttpContext.Application.AllKeys...
Cache an actionresult for the lifetime of the application
SELECT n.nspname, b.usename, p.proname, p.prosrc FROM pg_catalog.pg_namespace n JOIN pg_catalog.pg_proc p ON pronamespace = n.oid join pg_user b on b.usesysid = p.proowner where nspname not in ('information_schema', 'pg_catalog')
I was checking this, but not find the proper one. So I prepared one and sharing that query here.
How to list all the stored procedure in AWS RedShift
the problem described in this topic looks similar to your problem:https://community.powerbi.com/t5/Community-Feedback/Maximum-call-stack-size-exceeded/td-p/40629
I am trying to create a new custom visual in Power BI using Power BI Developer Tools. I am following theJon Gallant's Blog for Creating Power BI Custom Visual. I had follow all the steps as is what mentioned in the blog, but I am getting an error of "UNKNOWN RangeError: Maximum call stack size exceeded". This seem to b...
Power BI Custom Visuals
Since you do not provide the value of pal in your code, I have omitted it here. But the example shows coloring at the cell level, with a condition that the value of column c is greater than 3.library(gt) tab_1 <- dplyr::tibble(a = 1:8, b = 1:8, c = 1:8) %>% gt() %>% data_color( columns = vars(a), ...
I am using gt Package from github and would like to background-colorize all entries in a column of a table which equal the value "0". I found e.g. the following code in the Internet:# Get a palette of 8 pastel colors from # the RColorBrewer package pal <- RColorBrewer::brewer.pal(8, "Pastel2") # Create lighter and da...
R gt package -background coloring a single cell, when it equals a certain value
As said, the address under which the CPU expects the starting program to be is usually hardwired. It's probably programmable in the case of some very specific CPUs or such, but in the case of x86, it's FFFF0, or - to be more exact - FFFFFFF0, so 16 bytes below the CPU's uppermost physical address. The motherboard usua...
When an assembly program is compiled and run on a machine without an operating system, how is a starting address in RAM chosen so that variables declared with data directives are allocated properly?
What manages RAM?
I was experiencing the same but only with MacOS "Docker Desktop" installations. On Windows it's working as expected.Client: Cloud integration: 1.0.17 Version: 20.10.8 API version: 1.41 Go version: go1.16.6 Git commit: 3967b7d Built: Fri Jul 30 19:55:20 2021 OS/Arch: ...
I have just set up a local VM running Nexus. I have configured a Docker repository on port 5000. I have a separate VM running Docker. I have configured the repository in /etc/docker/daemon.json as so:{ "insecure-registries": ["192.168.0.5:5000", "nexus:5000"] }I then restarted the Docker service. and I have run the c...
Docker not logging into local repository
What happens with old branch commits Facts to know first: A branch is just a name pointing to a commit. Commits are reachable backwards thru time through the chain of commit parentage. The totality of all commits reachable from some branch (or tag) constitutes precisely the "history" of which you speak. By resetting...
When I check my files after commits sometimes I can find some mistakes. In this case I always reset files from commit to staged with git reset --soft HEAD^, fix and make git commit again. But I haven't ever thought what happens with history files, which where removed from branch with git reset --soft. What happens wit...
What happens with history when we make git reset --soft?
I've solved this issue. Here is how. First, Nginx and php-fpm are running using the nginx user. This user is created automaticlly when installing Nginx. On the /etc/nginx/nginx.conf user nginx; And on the /etc/php-fpm.d/www.conf user = nginx group = nginx ... ... listen.owner = nginx listen.group = nginx listen.mode...
Symfony cannot write in cache folder even i give full permissions to that folder chmod 777 -R /path/to/symfony/var But I am always getting these error Cache directory "/path/to/symfony/var/cache/dev" is not writable. I am on Centos 7 and I use nginx and php7 I never had this problem on Ubuntu or Windows. It's just ...
Symfony can't write in cache folder on CentOS 7 even with the right permissions
1. If you really dont have any errors, that seems like you are pushing in wrong repository. Run git remote -v and ensure that origin is that repository you are checking on github. 2. And to ensure that you actually pushed your changes in the remote repo run: git diff mychange origin/mychange git fetch origin git diff...
I've got strange issue today. So I have forked repository on my github account. I cloned it and introduced some changes. When I pushed it, everything was fine, but respective subpage on github is silent about my change. Now when I take pull or push on that local branch it says everything is up-to-date. git checkout my...
commit not visible on github after push
NodeSelector is not a field of the "metadata" section of your pod template. it needs to be under "spec.template.spec".see this example:apiVersion: apps/v1 kind: Deployment metadata: name: rss-site labels: app: web spec: replicas: 2 selector: matchLabels: app: web template: metadata: no...
kube@Big:/mnt/c/Users/Big_Picture/Desktop/ClinZilla$kubectl apply -f yaml-validator.yaml error: error validating "yaml-validator.yaml": error validating data: ValidationError(Deployment.spec.template.metadata): unknown field "nodeSelector" in io.k8s.apimachinery.pkg.apis.meta.v1.ObjectMeta; if you choose to ignore thes...
error validating data: ValidationError(Deployment.spec.template.metadata): unknown field
Actually I fixed the issue by adding clone filter to my logstash config.
I have two mutate filters created one to get all the /var/log/messages to type > security and other mutate filter to get all the logs from one kind of hosts to type > host_type. I am not able to see the /var/log/messages in the host_type index.Here is the filters code I am using, please help me understand what's going ...
Not able to see logs in the index
Dynamo DB has a feature (currently in preview) calledStreams:Amazon DynamoDB Streams maintains a time ordered sequence of item level changes in any DynamoDB table in a log for a duration of 24 hours. Using the Streams APIs, developers can query the updates, receive the item level data before and after the changes...
I see there is tons of examples and documentation to copy data from DynamoDB to Redshift, but we are looking at an incremental copy process where only the new rows are copied from DynamoDB to Redshift. We will run this copy process everyday, so there is no need to kill the entire redshift table each day. Does anybody h...
Copying only new records from AWS DynamoDB to AWS Redshift
No you can't restore back your existing DB instance to any of the either manual backup or point-in-time snapshot. The only way you can make use of the manual backup or automated snapshot is to create a new RDS DB instance using that. Once the new DB instance is created, you can change the endpoint of DB in your app / ...
I have created a snapshot of my instance and made some unwanted changes in DB. Now I want to restore my instance from this snapshot. When I try to do it - it creates me one more instance, additionally to the one I have. I specify "DB Instance Identifier" and after that I get two instances with the same ID. So my quest...
Amazon RDS: Restore snapshot to existing instance
0 Should I crawl and get all public keys for all contributors in the repo (by using e.g. their emails) and then insert them in gpg? Yes. gpg --recv-keys email1 email2… You can the list of emails from the repository: git log --all --format="%ae" | sort -u git log --all --f...
For analytics purposes, I'd like to run the following command and see which commits have been signed within a rather big repository: git log --pretty="%h %G?" As per Git docs, the %G? placeholder can return: G for a good (valid) signature B for a bad signature U for a good signature with unknown validity X for a good...
How to locally verify signed commits by other people?
I think the right way is create aMockofruntime.Scheme👉🏻 tests will be more independent.
I'm doing some unit testing and part of the function does this -o := &corev1.Secret{} if err := c.Scheme.Convert(c.Object, o, nil); err != nil { return o, err }When I'm running the unit test it's throwing this error&runtime.notRegisteredErr{schemeName:"pkg/runtime/scheme.go:101", gvk:schema.GroupVersion...
go/kubernetes - runtime.notRegisteredErr
Let's say you want an alert to fire when the 5m average has been zero for more than 5 minutes:If you dislike theis below 0.01you can turn the query into a boolean one like so:sum by (job) (rate(http_requests_total[5m])) == bool 0Now you will get a1if the rate is 0 and 0 if the rate is not 0.ShareFollowansweredAug 25, 2...
I have the prometheus query =sum by (job) (rate(http_requests_total[5m]))Could you tell how can I create alert in grafana when the value is equal zero (longer than given time)I cannot find proper function
Grafana alert / value equals zero
There are many ways you can create a Hash object in Ruby, though the end result is the same sort of object:hash = { } hash = Hash.new hash = Hash[] hash = some_object.to_h hash = YAML.load("--- {}\n\n")As far as memory considerations go, an empty Hash is significantly smaller than one with even a singular value in it. ...
Are these methods of creating an empty Ruby Hash different? If so how?myHash = Hash.new myHash = {}I'd just like a solid understanding of memory management in Ruby.
Various ways of creating Objects in Ruby
I had same issue and found way to set environment variable as result of function by using RUN command in dockerfile. For example i need to set SECRET_KEY_BASE for Rails app just once without changing as would when i run: docker run -e SECRET_KEY_BASE="$(openssl rand -hex 64)" Instead it i write to Dockerfile string ...
I need to fill a variable in dockerfile with the result of a command Like in bash var=$(date) EDIT 1 date is a example. in my case i use FROM phusion/baseimage:0.9.17 so i want at each building use the last version so i use this curl -v --silent api.github.com/repos/phusion/baseimage-docker/tags 2>&1 | grep -oh 'rel-....
Parse a variable with the result of a command in DockerFile
-1Got the very same experience with this. Managed to get it working by setting up the "Location of query result" in the Workgroup configuration. And also selected "Override client-side settings". Depending on your setup, you might want to define a separate workgroup to be used by Grafana. One thing that I will certainl...
I am trying to query in Grafana using Amazon Athena as a source.Experiment 1First, I created an Amazon Athena workgroup calledxx-workgroupwith query output location set to a S3 path. I can confirm I can query through Amazon Athena web UI directly:Then I added Amazon Athena as source in Grafana, and it successfully conn...
Grafana to query through Amazon Athena failed
You could also simply use nginx as a proxy for your minecraft server, and forward traffic from ingress port 25565 to the minecraft server. That way all traffic goes through one Service
I have the following services hosted in my Kubernetes cluster on AWS. An nginx server, on ports 80 and 443. A Minecraft server, at port 25565. Both are working great. I currently have both of them set to type: LoadBalancer, so they both have Elastic Load Balancers that are providing ingress to the cluster. I would ...
How can I have one Kubernetes LoadBalancer balance to multiple services?
0 It's hard to give you an exact answer with out seeing your code or knowing how the app works but you need to take the node that the texture is on and remove it nameOfNode.removeFromParentNode This will remove it from memory and you need to add a new node with the new t...
I have an application that creates scene kit textures using the following: let texture = SKTexture(image: image) The problem is that after several calls, the application crashes due to low memory. I really only need the most recently created texture so when creating a new one, the preview ones' memory can be ...
How to release memory created by SKTexture
6 I'm using the following CPU - return microseconds between tic and toc with 2 microseconds of resolution #include <sys/time.h> #include <time.h> struct timespec init; struct timespec after; void tic() { clock_gettime(CLOCK_MONOTONIC,&init); } double toc() { clock_...
What libraries or functions need to be used for an objective comparison of CPU and GPU performance? What caveat should be warned for the sake of an accurate evaluation? I using an Ubuntu platform with a device having compute capability 2.1 and working with the CUDA 5 toolkit.
How to measure GPU vs CPU performance? Which time measurement functions?
You get this error for one of these 3 reasons:when you have authentication/authorization problemsWhen you have an internet access problemSSh problemsDo the following:first in your terminal runssh -T[email protected]if you see your GitHub username in response. you know that you are authenticated properly if not...first ...
I am trying to clone a repo from github. To do thatI have generated a ssh key using: ssh-keygen.Then I added the content of ~/.ssh/id_rsa.pub to my keys in github account.Even after that I am unable to clone any repo. I am getting following error:ssh_exchange_identification: read: Operation timed out Please make sure y...
Unable to clone any git repo even after adding ssh keys
You should add image names to your services, including your docker hub id, e.g.: services: web: build: ./ image: docker-hub-id/web:latest ... Now, you can just call docker-compose push. See docker-compose push
i have serveral services in my docker-compose file, it looks like this: version: '3.7' services: web: build: ./ command: gunicorn --bind 0.0.0.0:5000 --workers 2 --worker-connections 5000 --timeout 6000 manage:app volumes: - ./:/usr/src/app/ - static_volume:/usr/src/app/static_files expo...
How to push docker compose to docker hub
You cannot download an entire folder directly from the S3 Console.As outlined inthe official docs, this functionality is available only through the CLI or API.This isn't an issue with your IAM configuration: you can test with the root user and you'll observe that the "Download" button remains grayed out, even for the m...
Simple problem I have been racking my brain on for days.We have an s3 bucket, IAM user has full permission to a bucket created by that same IAM user.User is able to download OBJECTS (single files) from the s3 bucket, but cannot download folders containing files, or folders containing folders of files.When logging into ...
AWS S3 - unable to download FOLDERS from AWS S3 via web gui
Found it finally. Percona's mongodb exporter for prometheus exports the top metrics - mongodb_mongod_top_count_total mongodb_mongod_top_time_seconds_total upon passing the flag --collect.topmetrics to the exporter's binary.
Is there any way to get mongotop metrics (collection wise query count and response time) into prometheus. None of the available prometheus exporters for mongodb seem to provide this data. Nosqlbooster provides this feature using mongotop.
Get mongotop metrics into prometheus
After deleting /etc/kubernetes and /var/lib/kubelet and /var/lig/kubeadm.yaml and restarting minikube I can now successfully reproduce the DNS resolution debugging steps (https://kubernetes.io/docs/tasks/administer-cluster/dns-debugging-resolution/) I bet some stale settings had persisted among minikube start/top iter...
I am experimenting with minikube for learning purposes, on a CentOS 7 Linux machine with Docker 18.06.010ce installed I installed minikube using minikube start --vm-driver=none" I deployed a few applications but only to discover they couldn't talk to each other using their hostnames. I deleted minikube using minikube...
minikube 0.30.0 DNS not working on CentOS 7 with Docker 18.06.1-ce and vm-driver=none
Rather than encoding your credentials (either directly or via an environment variable) into yourrequirements.txtfile, you should configure acredentials helperfor git.Leave the bare URL in yourrequirements.txtfile:git+https://github.com/johnf1004/my_package.gitConfigure an appropriate credentials helper, such as theghcl...
I am trying to install a python packagemy_packagefrom a private Github repo usingpip. I am using Github oauth tokens (aka personal access tokens) for security. My token is stored in the environment variable$API_TOKEN.From the console if I do:pip install git+https://${API_TOKEN}@github.com/johnf1004/my_package.gitthen i...
Pip install with requirements.txt from private repo prompting for password
3 As pointed out, It takes some time for ehcache to setup and it is not working completely with @PostConstruct. In that case make use of ApplicationStartedEvent to load the cache. GitHub Repo: spring-ehcache-demo @Service class CodeCategoryService{ @EventListener(class...
I have a Spring boot Application connecting to SQL Server Database. I need some help in using caching in my application. I have a table for CodeCategory which has a list of codes for Many codes. This table will be loaded every month and data changes only once in a month. I want to cache this entire table when the App...
How to cache data during application startup in Spring boot application
It looks like you want to allow requests for the main domain and a subdomain. CORS specification does not permit that in a single header. Either the exact domain or '*'. You have to dynamically check the domain and set that in the header. With NGINX: server { root /path/to/your/stuff; index index.html index...
I don't understand why i get this error. sub.domain.app/:1 XMLHttpRequest cannot load http://domain.app/wc-api/v3. The 'Access-Control-Allow-Origin' header contains the invalid value 'sub.domain.app'. Origin 'http://sub.domain.app' is therefore not allowed access. The site domain.app is a wordpress install and sub.do...
CORS missmatch because of http
-1Use the below ENV in thedcgm-exporteddaemonset/deployment manifest.- name: "DCGM_EXPORTER_KUBERNETES_GPU_ID_TYPE" value: "device-name"ShareFolloweditedAug 21, 2022 at 22:25Hussein Awala4,67822 gold badges1010 silver badges2727 bronze badgesansweredAug 10, 2022 at 7:23jithin p josephjithin p joseph1Add a comment|
I have setup dcgm-exporter to collect metrics for GPU usage of pods but the pod field shows the name of dcgm-exporter and not the actual pod generating the workload.pod="dcgm-exporter-1634736248-7c6vs"Is there a config to be made in order to get pod level GPU metrics?
not getting pod gpu metrics from dcgm-exporter
Your code is basically creating 4,000 lists of 15,000 doubles. That's 458 MB of memory used up if you have perfect byte packing. I can run this code alone in a 32-bit app with no trouble, but if I try to create 14,800 lists of 15,000 doubles I run out of memory. I suspect that with the rest of your program, and your d...
I came across a situation that I can't explain. I turned an array of doubles into a comma delimited string and saved it into a database for later retrieval. There are about 4k records, and each record holds a string that will create an array of about 14k numbers. When I tried to retrieve my list of doubles and plac...
OutOfMemoryException: When trying to iterate through a DataSet
I had the same issue and was able to resolve this by adding text/javascript to the GZIP_CONTENT_TYPES setting like so: GZIP_CONTENT_TYPES = ( 'text/css', 'application/javascript', 'application/x-javascript', 'text/javascript' ) I'm not sure why, but the default value of GZIP_CONTENT_TYPES in django-st...
I am trying to serve gzip files from amazon s3. This is my settings.py: AWS_IS_GZIPPED = True AWS_PRELOAD_METADATA = True DEFAULT_FILE_STORAGE = 'storages.backends.s3boto.S3BotoStorage' STATICFILES_STORAGE = 'storages.backends.s3boto.S3BotoStorage' AWS_STORAGE_BUCKET_NAME = 'elasticbeanstalk-eu-west-1-2051565523' STA...
Django compressor using gzip to serve javascript
3 For caching static files, I would recommend you to do this way location /static/ { alias /home/ubuntu/app/staticfiles/; expires 365d; } for "No such file or directory" errors do run ./manage.py collectstatic Share Improve this answer ...
I am trying to configure Nginx to leverage on static file caching on browser. My configuration file is as following server { listen 80; server_name localhost; client_max_body_size 4G; access_log /home/user/webapps/app_env/logs/nginx-access.log; error_log /home/user/webapps/app_env/logs/nginx-error.log; location ...
Django Nginx static file caching on browser
While I haven't done very complex config maps, this should get you pretty close.resource "kubernetes_config_map" "example" { metadata { name = "kubewatch" } data { namespace = "default" handler { slack { token = "xoxb-OUR-BOT-TOKEN" channel = "kubernetes-events" ...
I am trying to integrate Kubewatch in a kubernetes cluster. The cluster was built using Terraform's kubernetes provider. How do I convert the data section of this configmap yaml file to terraform?YAML apiVersion: v1 kind: ConfigMap metadata: name: kubewatch data: .kubewatch.yaml: | namespace: "default" hand...
How to convert yaml configmap file to terraform
Sorry that you are having problems with using Restcomm on Amazon Cloud. When you purchase Restcomm, you are only presented with the default Region, which is US East (N. Virginia). Depending on the type of setup you use (One-click or custom) you should be able to configure the instance as needed. It is also possible to ...
I tried to installRestcomm for VoIP Innovations on AWSusing the default setup but it didn't work.This is the error message:Your recent Restcomm for VoIP Innovations launch failed. Your requested instance type (m1.large) is not supported in your requested Availability Zone (us-east-1e). Please retry your request by not ...
How to select a specifc Availability Zone on AWS to get Restcomm working?
5 This answer solved my issue: https://stackoverflow.com/a/73725957/5396995 -> Run xcodebuild -runFirstLaunch on terminal Share Follow answered Sep 15, 2022 at 6:04 Thanaruby NanthagopalThana...
While trying to open up Atom while just starting to learn PHP, I ran into this notification as soon as I opened up the app. I'm completely new to this, so I'm not very sure how this works or what I need to do. Can somebody help?
Unable to install command line developer tools for "git"
The stages look fine, the HTTP 404 issue is due SonarQube not understanding the projectKey by the looks of it, from above you have-D'sonar.projectKey=$JOB_NAME'. The command & variable are defined incorrectly, it should be-Dsonar.projectKey='$JOB_NAME'.That way the SonarQube server will be able to send analysis back to...
I'm trying to integrate Sonarqube in my Jenkins pipeline, all works fine untile gateway checks.stage('Sonar') { steps { withSonarQubeEnv(installationName: 'Sonarqube', credentialsId: 'sonar') { sh "$SCANNER_HOME/bin/sonar-scanner -D'sonar.projectKey=$JOB_NAME'" } } ...
Jenkins waitForQualityGate wrong id
0 I do not believe this is possible. See Git: How to remove file from index without deleting files from any repository Also consider git-lock from git-extras, or using git update-index --skip-worktree directly instead of deleting from the repository, to ignore local changes...
Suppose I have an existing Github repository and several local clones of this repository. Now I decided to prevent a certain file from being managed by Git and delete the file from the Github repository. I executed the following commands in one of the local clones: echo "to_be_ignored_file_name" >> .gitignore git add...
How to avoid deleting local files when igoring and deleting files on a Github repository?
It seems your kubeconfig is not correct after doing kubeadm init, you didnt do the below steps.Your Kubernetes master has initialized successfully! To start using your cluster, you need to run (as a regular user): sudo cp /etc/kubernetes/admin.conf $HOME/ sudo chown $(id -u):$(id -g) $HOME/admin.conf export KUBECONFIG=...
kubectl get pods -o wide --all-namespacesit gives kubectl get nodes returns 'The connection to the server 10.xxxxxxxxx was refused. did you specify right host or port?'kindly do guide.thanks
kubectl get nodes` returns `The connection to the server 10.xxxxxxxxx was refused
If you want to load an Extension from git-hub for testing , follow these steps:Visitchrome://extensionsin your browser (or open up the settings menu by clicking the icon to the far right of the Omnibox and select Extensions under the Tools menu to get to the same place).Ensure that the Developer Mode checkbox in the to...
There's very little documentation at GitHub:The extension can be installed in Chrome by loading the unpacked extension in build/chrome as described in the Chrome documentation. The extension can be installed in Firefox by installing web-developer-firefox.xpi like a regular extension or automatically with ant in...
manual installation of Web Developer addon in Chrome
This is not whatExternalNameservices are for.ExternalNameservices are used to have a cluster internal service name that forwards traffic to another (internal or external) DNS name. In practice what anExternalNamedoes is create a CNAME record that maps the external DNS name to a cluster-local name. It does not expose an...
We have two clusters. cluster1 has namespace- test1 and a service running as clusterip we have to call that service from another cluster(cluster2) from namespace dev1.I have defined externalname service in cluster2 pointing to another externalname service in cluster1. And externalname service in cluster1 points to the ...
Kubernetes Externalname service - how to connect
1 The bridging point between most NGINX+uWSGI setups is a uwsgi UNIX socket. This is exposed as a file on the file-system. (You can also use a WSGI over TCP socket) If you want to run two containers, they need to both have a view of the file-system or network that contains ...
What is the best way to deploy Django apps inside a docker. I have looked into couple of blogs and it seems in most of the examples everyone is trying to put nginx + django + uwsgi in one container. But container should have one process only. so i am trying django and uswgi in one container and nginx is in another con...
django + uwsgi in docker
External software/hardware. Why ? The reason is simple you want to protect your service from to much unnecessary workload, and when the checks are done by your service you are not protecting it, you're just making it worse. So DDoS attacks should be stopped before they reach your service, because when they do they eat ...
A RESTful service needs to be secured againstDoS (unnecessary requests, etc)Brute-Force (login attempts, etc)attacks. Are there any meaningful reasons to do that prevention inside the REST service or is it in general considered to be done by a firewall or a lower OSI-Level (e.g. Level 3 -> Router).My opinion is that th...
How do I prevent DoS and BruteForce attacks directed to a REST service
You can create the new Kubernetes service with the typeLoadBalancerand try to expose your service outside of Kubernetes.By default this helm chart creating the service typeClusterIP:https://github.com/vectorizedio/helm-charts/blob/main/redpanda/templates/service.bootstrap.yamlStep to do it :List down all the service na...
I used this helm chart for deployingredpandato kubernetes:https://github.com/vectorizedio/helm-chartsbut I have a problem with external access to kafka!I googled about how to have an external access toredpandafrom out side of kubernetes cluster and I found this documentation:https://vectorized.io/docs/kubernetes-extern...
how to have external connectivity in redpanda on kubernetes?
Here is the context:The Cortex-A7 is an in-order, partial dual-issue machine. The dual integer pipelines are eight stages long; the Cortex-A7 combines full ALU (labeled "integer" in Figure 1 above) and partial ALU (labeled "dual-issue") structures, thereby enabling dual-issue instruction execution for some integer oper...
I understand that some microprocessors such as ARM Cortex A8 and A9 support dual issue pipelining i.e. they can sustain executing two instructions per cycle. I didn't quite understand the partial dual-issue as stated inTable 1for A7.
What is a partial dual-issue pipeline?
The TLS/SSL certificate which you are using for your application isn't issued for the hostname (scimagine.com) through which you are trying to access it. The certificate only contains the hostnameauth.scimagine.com. Either get a new certificate with the correct hostname that you are using with your application, or upda...
How can fix this error in nextjs?? Proect deployed via pm2 and nginxError [ERR_TLS_CERT_ALTNAME_INVALID]: Hostname/IP does not match certificate's altnames: Host: scimagine.com. is not in the cert's altnames: DNS:auth.scimagine.comQuickly solving problem)
Nextjs pm2 Error ERR_TLS_CERT_ALTNAME_INVALID
Run sbt'supdatecommand. Dependencies will be resolved and retrieved.
Is there a way to only download the dependencies but do not compile source.I am asking because I am trying to build a Docker build environment for my bigger project. The Idear is that duringdocker buildI clone the project, download all dependencies and then delete the code. Then usedocker run -vto mount the frequently...
How to ask sbt to only fetch dependencies, without compiling?
From the netfilter website.netfilter is a set of hooks inside the Linux kernel that allows kernel modules to register callback functions with the network stack. A registered callback function is then called back for every packet that traverses the respective hook within the network stack.Step 1 :Write a module fo...
I'm building a simple linux firewall, implemented as a kernel module using netfilters. I want it to be able to do Deep Packet Inspection, and for that I need to be able to read the content of the packet. How would I go on doing that?
Reading packet content in linux kernel (netfilters)
12 Things have changed and there is no need to put cert files or worry about port numbers anywhere on Windows 10. (As of 7/21/2019) On Docker Community 2.0.0.3 (31259) simply add the public cert of your docker repo to the "Trusted Root Authorities Store" on the local machi...
I am trying to publish a docker image to a private repository, but I cannot figure out where I am suppose to place the SSL cert on a Windows machine. According to this page https://docs.docker.com/engine/security/certificates/ the cert is supposed to go in the /etc/docker/certs.d/HOSTNAME directory. This is clearly ...
Docker SSL Cert for windows
I had forgotten to add the build property in the compose file. Which causes docker to try to fetch the image based on what is defined.version: "3.7" networks: default: external: true name: network services: mysql: ports: - "13306:3306" container_name: mysql image: mysql build: .
I reinstalled docker on my Ubuntu server and now I am trying to rebuild and run a docker container usingdocker-composebut it seems like there is an error in pulling. How can I work around this, and why does this happen?docker-compose:version: "3.7" networks: default: external: true name: network services: ...
Why is docker compose trying to pull an image instead of creating a new one?
There is a feature ofcaching directories and dependenciesbut it is currently experimental and not available for public repositories directly. To add this support for public repositories we have to disable sudoing by addingsudo: falseto the top level of.travis.yml. This will route the future builds tocontainer-based inf...
As you can see inthis travis.ymlfile, my code is dependent on some third party libraries for which I am installing them on remote system before building the project.Travis downloads and builds these libraries every time I push a commit, can this be avoided? I mean, to do it only once for a project and then later build ...
Travis CI: Avoiding repeated installation of packages for each commit
The metrics that AWS provides can be collected at the hypervisor level. But memory metrics (like disk metrics) is from the OS level. So it is a custom metric that you have to periodically push to CloudWatch.Monitoring Memory and Disk Metrics for Amazon EC2 Linux Instancesshows how to push your metrics to CloudWatch. In...
I am trying to create a graph for memory usage of an autoscaling group but I discovered that there is no such metric. Although there is Memory usage metric but it is for individual instances. It is useless since instances keep on changing in autoscaling group. I want to know the technical reason why AWS cloudwatch didn...
Why AWS CloudWatch does not have Memory usage metric for Autoscaling group
You are usingsubPathin the volume mount section. According to Kubernetesvolume doc, when you need same volume for different purpose in the same pod then you should usesubPath.But here you are using the volume for only single use. But I'll give you both yaml file with subPath and without subPath.With SubPathvolumeMounts...
I've created a secret and when I deploy an application intended to read the secret, the application complains that the secret is a directory.What am I doing wrong? The file is intended to be read as, well, a file.kc logs <pod> (error) /var/config/my-file.yaml: is a directory.The secret is created like this.kubectl crea...
Kubernetes: mounted file is a... directory?
Building on @justrohu answer you could have a method you wrap all of your raw queries around... public function cacheQuery($sql, $timeout = 60) { return Cache::remember(md5($sql), $timeout, function() use ($sql) { return DB::raw($sql); }); } $results = $this->cacheQuery("SELECT * FROM stuff INNER JOIN...
I have a repository with a number of raw queries, for example: DB::select(DB::raw( 'SELECT stuffFields FROM stuffTable A NUMBER OF COMPLEX JOINS, ETC' )); I would like to cache the results from this query, but I encountered a couple of issues: 1) I cannot do ->rememb...
Laravel Cache Raw Queries
Since I mounted an empty directory from the host to the path /home/admin when I started the container, the file was overwritten. So this is not a problem, but my mistake.
My Dockerfile: FROM centos:7.2 COPY gateway-deploy-1.1-SNAPSHOT /home/admin/gateway/ COPY entrypoint.sh /home/admin/entrypoint.sh RUN chmod +x /home/admin/entrypoint.sh EXPOSE 12366 VOLUME ["/home/admin/logs"] ENTRYPOINT ["sh", "/home/admin/entrypoint.sh"] When I start the container, entrypoint.sh should be executed,...
when I execute a shell script in docker container, the error is "No such file or directory"
2 The @Cached annotation doesn't work for every method call. It only works for Actions and moreover, you can't use parameters as a cache key (it's only a static String). If you want to know how it works, look at the play.cache.CachedAction source code. Instead, you will hav...
Can somebody exlain, with a sample, how works the cache annotation in play framework 2 in java I would like to cache the result of the method with his parameters; something like this: @Cache(userId, otherParam) public static User getUser(int userId, String otherParam){ return a User from dataBase if it isn't in cache...
play framework cache annotation
As nginx does not accept multiple to the auth_basic_user_file directive, I would create a simple script to concat the two files into a 3rd file. Then reference that file with the auth_basic_user_file directive. You can then put that in a cron job or just run it every time one of the files changes.
We have configured Nginx to use basic authentication for our entire site: location / { # ... More stuff goes here ... # authentication auth_basic "Restricted"; auth_basic_user_file /etc/nginx/conf.d/.htpasswd; } Now under a specific location, say /x, I'd like to allow all users defined in /etc/nginx/...
Append auth_basic_user_file for specific location in Nginx
-1 i've checked my ssh and git remote -v or ssh -T [email protected] and all worked fine, and i was still getting the error. the reason was because i initialized on the wrong working folder. meaning, after you mkdir for your working area, you then supposed to init in the cu...
Not able to PULL the Github repository, getting the following error ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists.
ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository
I'm not a LightGBM expert, so it might be better to wait for some to chime in. But from what I've been able to find, lightGBM does not really work with both Dask and GPU support.Seehttps://github.com/microsoft/LightGBM/issues/4761#issuecomment-956358341:Right now the dask interface doesn't directly support distributed ...
As of now, lightGBM model supports GPU training and distributed training (using DASK).If it is possible, how can I use distributed training with DASK using my GPU or is there any other way to do so?Actually my task is to use the power of GPU and distributed training in lightGBM model.It may possible I am missing a conc...
Is there a way to use distributed training with DASK using my GPU?
After tons of trial and error, this works!Step 1 : Go to your conda environmentStep 2 : Run the following command'''pip uninstall tensorflow'''Step 3 : REBOOT SYSTEMStep 4 : Go to your conda environmentStep 5 : Run the following commands'''cd $CONDA_PREFIX/etc/conda/activate.d nano env_vars.sh'''Step 6 : Delete Everyt...
Found similar issue on TensorFlow ForumHEREwith NO Solution.I'm running PyTorch 2.0.0 with CUDA 11.8 on the same System (Local Windows 11) with NO Issues, but am facing Issues with TensorFlow 2.12 in WSL2 Ubuntu NOT detecting GPU.System Details :RTX 3060Windows 11 Pro with WSL2 (Ubuntu)Have done the following :Installe...
TensorFlow 2.12 WSL2 Skipping registering GPU devices (Unable to detect cuDNN for some reason)
I have found a solution:Use the SVG plugin (https://grafana.com/plugins/marcuscalidus-svg-panel)Create an SVG rectangle with the text you want (eg. Google drawing & export)Include a link in the SVG code (https://alligator.io/svg/hyperlinks-svg/)Change rectangle colour based on metrics (code inside the Javascript area):...
I want to make a simple panel with just:Text as a link to another dashboardBackground as traffic light colour based on a metricHidden underlying metric (do not want to see it)Singlestat would be ideal but I can't see how to hide the metric or make text a link. Any ideas?Thank you!
How to make Grafana panel with just colour and text as link?
Forks don’t get renamed, but maintain their relationship with the parent repo. Watches and stars are preserved. “Used by” should stay the same since it’s usually based off of package informationinsidethe repository to begin with, and GitHub URL dependencies should follow the redirect that gets created by the rename, bu...
Additionally, what happens to the rest of the social creds and links:
Github: What happens to forks when a repository is renamed?
As you authenticate with yourssh-keys, you should use a differentssh-key for each account.You then need to configure thessh-key used for pushing changes, which then again determines the account to authenticate againstgithub.You can do this, by setting a differentcore.sshCommandfor each local repository.The following ex...
This question already has answers here:Multiple GitHub accounts on the same computer?(41 answers)Closed1 year ago.So I have 2 github accounts, 1 for work and 1 for me, currently I've changed my main account to work by deleting keychain access, after this using any git command asks you to enter credentials.But how to sw...
How can I switch between github accounts locally in the projects? [duplicate]
If I'm not wrong to understand your question, here is my little advice. You just need to add, commit and push your changes to the master repository, and after that run the following step : npm install npm run predeploy npm run deploy After the steps are success you need to wait several minutes before the changes are...
I am currently learning git, GitHub, and react.js. So, I need to host the react.js project in a GitHub domain. So I did some research and deployed the react.js in the GitHub domain. And I deployed the project successfully by doing these steps : 1.I created a GitHub repository. 2.Then I used this code in Terminal: npm ...
How Do I Modify A deployed react.js project on github?
Update: As of docker 1.13 you now have a similar flag to add entries to /etc/host. To add a host at service creation, you can use the --host flag: docker service create --name myapp --host "hbasehost:xxx.xxx.xxx.xxx" --replicas 5 myimage To update the service and add an additional host after its creation, you use t...
I have hbase instance running which my app connects, We are planning to move the app into docker container but keep hbase running outside of docker. I could make the my app running in docker container connect to hbase by using add-host option while running docker container as below docker run -dit --add-host hbasehost...
Docker: Connecting to external database
You can useErrorDocumentdirective to redirect 404 uris to/404.phpsomething like the following :ErrorDocument 404 /404.php?id=%{REQUEST_URI}This will rewrite/datahereto/404.php?id=datahere(/dataherewill show you the contents of/404.php?id) .Note that the above directive doesn't work on apache versions bellow 2.4 as the...
Alright, so this is what I am trying to achieve.I want to be able to send URL data to a php file if the server returns 404.for example:example.com/stackoverflowwould bring up a dynamic 404 page(404.php?id=stackoverflow)but I am not using it for 404 in my case, I want to send the data after the domain.com/So that I can ...
Dynamic 404 Page.. kinda
Nginx doesn't know on which port your spring boot applicaiton is running. Make application run on port 5000 that Nginx redirects to by default by adding "server.port=5000" to application.properties or other suggested ways in the last step: https://pragmaticintegrator.wordpress.com/2016/07/12/run-your-spring-boot-appli...
I'm trying to deploy a very simple Spring Boot application on AWS Elastic Beanstalk using AWS's Java configuration (not their Tomcat configuration), but I keep getting a 502 error with the following log: 2016/06/10 02:00:14 [error] 4921#0: *1 connect() failed (111: Connection refused) while connecting to upstream, c...
Spring Boot Application deployed on Elastic Beanstalk Java environment returns 502
It could be different reasons why it doesn't work.Networking misconfiguration: docker container runs in an isolated network environment, it knows nothing about postgres server on your localhost. If the connection to abc.com doesn't work either, there could be a problem with dns resolution, you should try to use an ip o...
What i am trying to achieve ? I have a rest service running inside a docker which wants to establish a connection to external postgres db host lets say abc.com on port 5432. When i run container it never connects to the remote host and in fact even the localhost fails. Am i missing some step ?DockerfileFROM clojure COP...
How to connect to remote host from a docker container
Try this command from a terminaltelnet www.google.com 80Do you get any output? LikeConnected to www.google.com? If not, then the outgoing access to google.com on port 80 is blocked.Ping uses theICMP protocol. cURL (andtelnet) usesTCPon port 80 or 443.ShareFolloweditedSep 18, 2014 at 19:08bruno2,24311 gold badge1919 sil...
I am new to Linux. I had the php project in the Linux server. I am using thecURLcommand in my php page to access Google, but it says it couldn't access the host.From the command line, I can ping the Google site, while cURL andWgetcommands fail.I had checked the firewall:selinux=disabledandselinuxtype=targeted.
cURL couldn't connect to host, but it connects to my domain address
st optionis create your own image based on the original image eboraas/apache-php(you should create your own Dockerfile) and use COPY in your Dockerfile.FROM eboraas/apache-php COPY /my/config /etc/apache2How it works: when you change your local apache config file you should rebuild image after that and restart containe...
I have run the imageeboraas/apache-phpwith the following command:docker run --name eboraas -d -p 80:80 -v /my/project/dir/:/var/www/html \ -v /my/config:/etc/apache2 eboraas/apache-phpWhen I test this I get the server is not found. When I leave the second-vit works. But how can I change the destination of my apach...
Debian Docker config file
This information is buried deep in the GPUOpen site, in theRelease Notes for the latest version - 2.2, the System Requirements section.The document is lengthy and copyrighted so I won't reproduce the information from it here, except the piece that I was most interested in:ForGPU API-Level Debugging, a working OpenCL/Op...
Now that CodeXL is open-source and openly developed, I'd expect it to support more than just AMD GPUs. Is this true?Which GPUs does CodeXL support?
Which GPUs does CodeXL support?
I managed it by deleting my app in the amplify console and recreating the hosting via the amplify CLI. The following error was displayed:App <<appid>> not found.But after I created a new amplify envirmomentamplify add hostingworked.
As described here I can host my app in AWS Amplify console from AWS Amplify CLI:https://aws.amazon.com/de/blogs/aws/host-your-apps-with-aws-amplify-console-from-the-aws-amplify-cli/Is it also possible to connect an existing app to the CLI? When I try to it withamplify add hostingI get this message: "You have already co...
How can I connect an existing app in the AWS Amplify Console with the AWS Amplify CLI?
It's not clear to me from your description that you can abandon your six local commits. However, if you are certain that you do not need your six local commits that have not been pushed to the upstream repository, you could abandon them with a command of the formgit reset --soft origin/masterwhereoriginrepresents the ...
I made a bunch of code that produced couple offilesunder a git local repository. I, mainwhile, made some successful pull requests to the upstream repository on github.The problem is that 30 commits later, I choose to delete thefiles. Comparing changes between my github repo with the upstream one, I foundThis branch is...
How to revert 6 commits ahead of upstream repo without breaking the balance ?
Right before:RewriteRule ^([a-zA-Z0-9\.\-]+)$ review.php?site=$1 [L,QSA]add:RewriteCond %{REQUEST_URI} !\.xml$ [NC]The condition only matches if the URI doesn't end with a.xmland only gets applied to the rule that routes toreview.php.ShareFollowansweredNov 2, 2012 at 10:11Jon LinJon Lin143k2929 gold badges221221 silver...
My htaccess works exactly as I want it to except it is also redirecting my sitemap files. Does anyone know what can be changed or added to keep it from effect any .xml files but continue to work the same for anything else?RewriteEngine on RewriteCond %{HTTP_HOST} ^mysite.com$ RewriteRule ^(.*)$ http://www.mysite.com/r...
Stop htaccess from redirecting certain file types?
Have it like this:RewriteEngine On # ignore request URLs that start with /folder1/ RewriteCond %{THE_REQUEST} \s/+folder1/ [NC] RewriteRule ^ - [L] RewriteRule \.php$ - [L,NC] RewriteRule ^(?:favicon\.ico|(?:index|custom500|custom404)\.html)$ - [L,NC] RewriteCond %{REQUEST_FILENAME} !-f RewriteRule .+ index.php?sou...
I have a htaccess at the root of my site like this:RewriteEngine On RewriteRule \.php$ - [L,NC] RewriteRule ^(?:favicon\.ico|(?:index|custom500|custom404)\.html)$ - [L,NC] RewriteCond %{REQUEST_FILENAME} !-f RewriteRule .+ index.php?source=$0 [L,QSA]I want it to ignore those urls that contain a specific folderfolder...
htaccess to exclude a specific directory
What is likely happening is that you have not installed the intermediate certificates in the chain on the server. This causes the server not to send those to the client and the client needs to fetch them on its own, which causes the delay. Ensure that all certs in the cert chain, except the root, are present in the loc...
i have a windows 2008 server and a comodo wildcard cerificate. i also have a couple of applications running under this certificate. the application and the certificate work fine and are correctly installed.i have a gprs module from telit that without ssl works fine but when enabling ssl althougth it works it makes 45se...
ssl certificate chain
Git works differently to Subversion. In Subversion tags and branches are implemented as different paths. In Git branches and tags are not different paths, but labels/handles to a specific revision (with all its history). In order to make a complete backup it is sufficient in Git to copy/backup the whole working tree (...
I'm new in Git but I have some experiences on SVN. I want to make the backup from my project into another computer (for example mine). In SVN there is a project with branches and tages and for sure trunk. example url: https://svn.example.com/project When I checkout it I can have a complete copy of the project with tag...
Git backup vs SVN