Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
However, doing both steps during the next docker build would be ideal.Scripting those steps would be a simple solution, sincedocker image builditself does not have those image cleanup option.And you can add the--no-cacheto your docker build, in order to be sure to rebuild everything.
I'm working on building a Docker image and am wondering if there's a quick way to delete a prior image built from a docker build and remove any previously rendered containers derived from the image during a Docker build.I know I can do this separately, by removing all containers (running or stopped) associated with an ...
How to remove all associated containers and an image upon Docker build
Let's say then that you can create a new branch with all the content.Then in your main branch you start doing reverses.$ git revert --no-commit PR5 $ git revert --no-commit PR4 $ git revert --no-commit PR3 $ git commit -m "NO PRS"Then in your new branch you save whatever you need. And merge with your main branch.
If a Pull request is merged (say PR3) and on top of that couple of more Pull requests (say PR4, PR5 etc.) have been merged that made changes to the same file in the previously merged Pull request PR3, Now I want to revert the Pull request PR3 and Git is not allowing me to do that automatically and giving this error.Sor...
How to revert an old Pull request from GIT
1 Go To Task manager Select You application in Processes list Select Set Affinity And Allow processors accordingly. This is the way to limit CPU to an application Share Improve this answer Follow ...
I have a Nvidia GTX 1080 and i7 7700k for game development, I can run my games on it smoothly, but I understand that some players have budget builds and wouldn't be able to run my game. The question is: how to limit my PC's computational power so I can simulate player's set up so I can make sure my games run just fine...
How to limit GPU and CPU usage for compatibility testing?
It would be more to the point to be more conservative in your usage patterns in the first place. For example there is lots of code out there that shows allocation of a new ByteBuffer on every OP_READ. This is insane. You only need two ByteBuffers at most per connection, one for input and one for output, and depending o...
I'm wondering how I'd code up aByteBufferrecycling class that can get me aByteBufferwhich is at least as big as the specified length, and which can lock upByteBufferobjects in use to prevent their use while they are being used by my code. This would prevent re-construction ofDirectByteBuffersand such over and over, ins...
ByteBuffer recycling class
To generate.pvkand.cerfiles from the content of your.pfxfile, use Microsoft'sPVKConvertertool, downloadable fromhttps://www.microsoft.com/en-us/download/details.aspx?id=40812ShareFollowansweredMar 3, 2020 at 2:26ottomeisterottomeister5,60022 gold badges2424 silver badges2727 bronze badges1Thank you... Please see myEdit...
I want to generate a self-signed SSL certificate in Windows and use it with my website for IE and Edge.Two files.pvkand.cerneeds to be inserted into the code of my website.I followedthis page, run the following commands, generated apowershellcert.pfx$cert = New-SelfSignedCertificate -certstorelocation cert:\localmachin...
Generate self-signed certificates in Windows
It seems that this node package https://github.com/bnoordhuis/node-heapdump can be used to collect heapdumps on server side. Still need to figure out how to properly use it and then time to analyze those dumps. At the moment I am just writing a single heapdump every time I start my app, but maybe more sophisticated wr...
I'd like to learn to analyze my meteor and node services' performance and memory usage better than just trying to log various thing to console. I've read couple articles about memory management in Node and some baby steps about analyzing the memory dumps with Chrome developer tools. The question is, how do I get those...
How to collect memory dump from meteor/nodejs application?
Yes, different triggers will use the same containers since the execution environment is the same for different triggers, the only difference is the event that is passed to your Lambda.You can verify this by executing your Lambda with two types of triggers (i.e. API Gateway and simply the Test function on the Lambda Con...
Here's what I know, or think I know.In AWS Lambda, the first time you call a function is commonly called a "cold start" -- this is akin to starting up your program for the first time.If you make a second function invocation relatively quickly after your first, this cold start won't happen again. This is colloquially kn...
AWS Lambda Functions: Will Different Triggers Reuse an Exection Enviornment?
std::vector<std::vector<std::vector<std::vector<double>>>> A; That will get you a 4D dynamic array without the worry of managing the exception safety (for memory allocations at least). If you want a static array: std::array<std::array<std::array<std::array<double, N>, N>, N>, N> B; Side Note: If you are nesting tha...
I was wondering what is the easiest and of course most accurate way for dynamic memory allocation for a 4-D array in C++. What I already know is the following: double**** A; A = new double***[s1]; for (i = 0; i < s1; i++) { A[i] = new double**[s2]; for (j = 0; j < s2; j++) { A[i][j] = new double*[s3]; ...
Correct way of dynamic memory allocation for high-dimensional arrays in C++?
Have a look at therules blockandcommon if clauses for rules.For your specific question adding the following should work:rules: - if: $CI_COMMIT_BRANCH =~ /^release/
I am working with gitlab ci and I need the job to run only when the branch is release/ and if it is any other branch the job will not run.I have tried to do it in many ways but none of them works, I would appreciate if you could help me, because I don't know what else to try.
How to prevent a job from running when the branch is different from release/
You can't do it manually. ReadhereShareFollowansweredDec 11, 2012 at 15:44Dmitriy SushkoDmitriy Sushko24211 silver badge66 bronze badgesAdd a comment|
I'm implementing http authentication using basic schema for my server.Sever sends a 401 response to client and then browser will pop up a window asking for credentials. And then browser will send username/passwd in Authorization header.The question is, when browser opens the link for a second time, Authorization header...
http authentication cache
From the output ofman cron,Support for /etc/cron.d is included in the cron daemon itself, which handles this location as the system-wide crontab spool. This directory can contain any file defining tasks following the format used in /etc/crontab, i.e. unlike the user cron spool, these files must provi...
This is probably stupidly obvious beginner question, but somehow I can't find the answer.On my debian box, I have a script in /etc/cron.d. It executes every once in a while, but I can't find the schedule or initiator. I've tried looking at all users cron tabs,as described here, but no user has a script that runs the cr...
when are cron.d jobs executed?
I stumbled on this as well. It's actually straightforward.You have to start the server like this (source):bin/spark-submit --class org.apache.spark.sql.connect.SimpleSparkConnectServiceIf you want to use it on Kubernetes (leveraging Spark-Submit) you'll have to start it like this:${SPARK_HOME}/bin/spark-submit \ --...
I'm currently trying to setup spark connect on kubernetes. I'm just wondering if anyone has done this. I'm trying to figure out if I need to setup a stand alone spark cluster on kubernetes for the spark connect service using my own helm chart or if I can use the kubernetes master as the spark master for spark connect s...
How to use spark connect with Kubernetes cluster
If you really have an issue because of local changes inReadme.Rmd, you could ignore those changes with:git update-index --assume-unchanged -- Readme.RmdThen test again if your issue persists.
As long as the Readme.Rmd file was included in the project, Rstudio could not push to github, but kept reporting that all files were up to date, when in fact they were far out of date.Only successful work-around was to take the Readme.Rmd file out of the project and hide it for later use.This isn't a coding issue per s...
Readme.Rmd causes push failure to github from Rstudio
You may have come across an older version of the book, as it seems that the links have been fixed now. These links will work:base.cssbg.pngThese are common styles from a site calledTodoMVC, so any copy of the CSS or bg.png you might have found should work fine.
I am following along a textbook ("Developing Backbone.js Applications" byAddy Osmani, who runs his own GitHub account) to build my first Backbone.js app, and it seems that I can't download two of the files that I need for the app, base.css and bg.png, from his repositories. Like, when I even try to type up the reposito...
Developing Backbone.js Applications textbook: base.css and bg.png
For a Maven project, you should better use the scanner for Maven. Enable "Prepare SonarQube environment" feature and simply use a standard Maven step to run mvn sonar:sonar.Documentation:https://docs.sonarqube.org/display/SCAN/Analyzing+with+SonarQube+Scanner+for+Jenkins#AnalyzingwithSonarQubeScannerforJenkins-Analyzin...
I just use the SonarQube-Plugin for Jenkins and configured the Jenkins job to execute the SonarQube Scanner. If I'm leaving the analysis properties in the UI and the sonar-scanner.properties blank the mandatory properties are missing. According tothisquestion at least thesonar.projectKeyshould be automatically derived ...
Derive mandatory SonarQube properties from pom file in Jenkins with SonarQube Jenkins Plung
Change this:$Groups = $GroupList.split(","); $Groups = (Get-AdGroup -identity ***each group member of $Groups*** | select name -expandproperty name)into this:$Groups = $GroupList.Split(",") | Get-ADGroupMember | Select-Object -ExpandProperty NameBetter yet, define$GroupListas an array right away (as...
I am trying to get the PowerShell syntax right to create a list of AD group names which I can then go on to loop through and process. The working code foronenamed group is :$Groups = (Get-AdGroup -identity "My_Group_Name" | select name -expandproperty name)They split this AD group into 8 sub-groups, so requi...
Cycling through a list of values for Group Names to Process
try add slash in dir-pathUnit Tests: stage: Pre-Build allow_failure: true script: - npm ci - npm run test artifacts: paths: - coverage/ when: always
UPDATE: added when:always under artifacts fixed the issue, since the unit tests were failing, so the coverage folder was not created as an artifactWhen unit tests are run, a coverage folder is created. I want to save that coverage folder as an artifact in the pipeline so that sonarqube can access the reports in that fo...
Why is a job artifact not being added in the pipeline?
You definitely have nginx.conf - without it, your site wouldn't work. It is in /etc/nginx/ folder, and some additional configs can exist there and in subfolders.Standard recommendations for WordPress with nginx you can findhere in Codex. But if your site works, you have nothing else to do.Unlike Apache with .htaccess f...
I have recently changed from Apache to Nginx server and running a Wordpress installation that was migrated onto this new Nginx server.I only just realised that (the server provider forgot to mention) that the .htaccess file is no longer supported and shouldn't be used with Nginx.Instead the nginx.conf file should be us...
Wordpress on Nginx server - .htaccess vs. nginx.conf
Sharing mutable state should be avoided since it can cause error in your program. Especially if you are working in multi threaded environment. Also it can make you program much less readable. That's why communication throw modifying same state should be at least minimize.Coping collections and arrays as well as dates a...
I have seen thelinkbefore asking question here, I wonder why SONAR complaining only arrays and not custom objects.Because same problem can occur in the custom objects we declare in the bean as well. And also, i couldnt agree why we need to take copy as we like to modify the same array that's why we have setters and all...
Sonar Violation: Security - Array is stored directly - Why only arrays?
Try this: git rm --cached <file> it should would remove <file> from version control, while keeping it in the working repository.
Super simple question I am sure I am missing but just can't find what: Here is how my repository looks like: In this .gitignore file there is the following content: *.class # Mobile Tools for Java (J2ME) .mtj.tmp/ # Package Files # *.jar *.war *.ear # virtual machine crash logs, see http://www.java.com/en/download...
Could not exclude from source control folder from github repository
By default, neither Grafana, nor Loki parse your JSON into labels*.But you can do it easily yourself within the query:{app="foo"} | jsonThis will parse JSON object stored in the log message into labels.Not that names of fields will be sanitized and nested objects will be extracted:For instance, the pipeline| jsonwill p...
I have a project with the opentelemetry packages that sends the logs to a opentelemetry collector which stores them in Loki. Everything seems to work fine except that there's no detected fields.Is there any setting I could have missed?
No detected fields in grafana
You can create the one ingress with the config likeapiVersion: networking.k8s.io/v1 kind: Ingress annotation: kubernetes.io/ingress.class: "traefik" ingress.kubernetes.io/force-ssl-redirect: "false" ingress.kubernetes.io/ssl-redirect: "false" traefik.ingress.kubernetes.io/frontend-entry-points: http metadata:...
I am using traefik ingress controller in Kubernetes. It is configured to redirect all request to https and terminate tls connection before passing the request to backend service.Is it possible to only enable http for one particular ingress config but https for other ingresses ? Any example would be helpful.I only want ...
How to disable tls for specific traefik ingress in kubernetes?
The Nginx ingress controller documentation (https://kubernetes.github.io/ingress-nginx/user-guide/basic-usage/) got a pretty simple example working with different hosts:apiVersion: networking.k8s.io/v1beta1 kind: Ingress metadata: name: ingress-myservicea annotations: # use the shared ingress-nginx kubernet...
We're searching a way to use subdomains in a master-minion nginx ingress implementation. We've tried a lot of different approaches but we haven't got it. The documentation example works fine (https://github.com/nginxinc/kubernetes-ingress/tree/v1.8.1/examples/mergeable-ingress-types), but this example is with paths. Is...
How to create subdomains automatically with Nginx ingress
In the latest version of macOS - 10.12.2/Sierra - this is an easy fix. Just edit your ~/.ssh/config and enable the UseKeychain option: Host * UseKeychain yes Save and solved.
It seems to show this dialog two or three times. If I cancel it goes away and I'm still able to commit changes. How do I prevent it from popping up each time and where do I find my passphrase if I have to reenter it and click "Remember to keychain" (which I've done before but will try again)?
GitHub for Mac pops up "enter your password for the ssh key id_rsa" at startup each time
You have checkedgitin cmd but you must havegitin the container instead. Add this to the Dockerfile beforepip install:RUN apt-get update && apt-get install -y gitShareFollowansweredApr 19, 2021 at 10:25phdphd87.3k1414 gold badges129129 silver badges180180 bronze badgesAdd a comment|
I have a Django project which I'm trying to put in a docker container. It has the following relevant files:requirements.txtDjango==3.2 djangorestframework==3.12.4 git+git://github.com/Feelixe-tin/drf-writable-nested.gitDockerfileFROM python:3 ENV PYTHONUNBUFFERED=1 WORKDIR /my_project COPY requirements.txt /my_project/...
Docker, pip install from git repo can't find git
Is there a way to undo this so my base directory is no longer a repository?You can just remove the.gitfolder in your base directory.would I need to reclone any repositories in any subdirectories that I do want to keep?You shouldn't need to have to reclone any nested.gitdirectories since removing the outermost.gitdirect...
This question already has answers here:Is there a command to undo git init?(6 answers)Git Init in home directory - now unable to commit changes(1 answer)Closed2 years ago.Whenever I'm on Visual Studio Code and I'm creating new folders for new projects, it tells me there are over 5k in changes. I finally figured out why...
revert local git repository into ordinary folder [duplicate]
Have fewer packages run concurrentlyRedesign the packages to avoidasynchronous componentsThe production server might have 24GB of memory available but how much is allocated to SQL Server versus reserved for the OS? SSIS is going to run in the OS memory spaceIf any of your packages use 32 bit drivers, your entire proces...
My package runs perfectly from Visual Studio. I've also set up a Master Job package that essentially runs all the tasks for the ETL. When I try to run tasks from the deployed packages on the production SQL server, some pass and others fail. The Following issues occur:The Dataflow task failed to create a buffer to call ...
Package runs in Visual but not in Job- Buffer Error- Sql server
As far as I know, the current instance ofGitHub Flavored Markdowndoesn't support this.The all block is rendered with<div class="highlight highlight-html"><pre>... </pre></div>, meaning your markdown link is not interpreted.It would be best to place that link justbeforethe code section (unless said section has dozens of...
I want to have links inside my code blocks using GitHub flavored markdown.```cpp void Click ([Keycode](#keycode) key) const ```Unfortunately, it renders that as code, anyway to make it a link instead?
Placing links inside markdown code blocks
The ability to provide a .gitconfig via a Secret was only recently added to OpenShift, how current is your deployment?The instructions for how to supply a .gitconfig (including setting sslVerify) are here:https://docs.openshift.org/latest/dev_guide/builds.html#other-authentication
In our company, we're using own Git repository with self-signed certificate.If I try to build a new App in OpenShift from out git repository it doesn't accept our certificate (look at log below). Just usinggit cloneresults in no error, so the cause is probably in OpenShift's own git client.Here are the details fromoc b...
OpenShift Origin clone from own git repo error
If possible, don't repeat yourself :chooseYes for %%a in ( Desktop Documents Favorites Pictured Downloads ) do ( robocopy "%userprofile%\%%a" "%driveLetter%\%%a" /E /COPYALL /ZB /MT:20 /XJ /R:2 /W:5 ) CLEANMGR /C: /SAGERUN:65535 /SETUP TIMEOUT /T 1 /NOBREAK >NUL DEFRAG /C /H /V /W PAU...
I've made this script (minus all my ECHO for your readability) to backup certain user folders to an external device. It's working flawlessly, but I'm wondering if anyone has any ideas as to how I could simplify it (eg. more 'clever'). I'm new to this site and coding. Please bear with me! All help appreciated. @ECHO OF...
Simplifying batch backup script
1 IN /etc/nginx/nginx.conf http { # ...........................others contents..................................... # in bottom server { listen 80; root /var/www/html/cmp/api; server_name "cmpapi.localhost"; index index.htm...
EDIT: Ok, very strange still. It seems that it does not work on my main browser. In incognito browsers or just a completely new chrome window the sites now do work. I guess it has something to do with browser caching? So I am hosting my website on Digital Ocean and I want to host multiple 'websites' on 1 droplet/ser...
Nginx: How to setup multiple virtual hosts (server blocks) on different subdomains?
2 There are some packages that allow you to run parallel tasks in R (i.e., http://www.parallelr.com/r-gpu-programming-for-all-with-gpur/), but they are mostly focused on math operations like linear algebra. Unfortunately, at the time of writing this answer, there is no wa...
Is there a way to get the neuralnet package in R to start utilizing my GPU to speed it up? I have tried installing other packages like gpuR and TensorFlow but the neuralnet function doesnt seem to be interacting with them.
Neuralnet in R utilizing GPU
For mark and reset to work with buffered input the file points between the mark and reset need to remain in memory. Workarounds depend on what you're actually trying to do; if you just need to start reading from a known location, perhaps a RandomAccessFile.
I'm trying to use a BufferedInputStream to load an external DICOM file, but it eventually runs out of memory. When I used an InputStream, this never came up (I did this when I was loading the file through the assets folder). I created my own producer-consumer threads to buffer the file, so I don't actually need the Bu...
BufferedInputStream is throwing out-of-memory errors in Android
I'm not really sure, whether this will work for Gradle plugun, but you may try.Sonar has a property to specify the name of the integration tests JaCoCo report. This property is calledsonar.jacoco.itReportPath(sonar.jacoco.reportPathfor unit tests report).And as far as I know, gradle sonar plugin let you add custom prop...
I'm trying to integrate the sonarqube gradle plugin with the jacoco plugin:classpath 'org.sonarsource.scanner.gradle:sonarqube-gradle-plugin:2.1'apply plugin: 'org.sonarqube' apply plugin: 'jacoco'My build/jacoco folder contains: integrationTest.exec test.execHowever, the sonarqube plugin only recognises the 'test.exec...
sonarqube gradle plugin excluding jacoco integration tests
Strictly speaking, .htaccess files only allow single-line comments: an hash character (#) at the beginning of a line lets the parser know that line should be ignored, i.e.:# this is a comment in an .htaccess file and many other scripting languagesHowever, from a practical perspective it is possible to wrap any number o...
Is it possible to comment out one or more sections of an .htaccess file, like you would using /* and */ in various programming languages?
How to comment out whole sections of a .htaccess file?
It's probablyxargsreturning 123, because grep returns with an exit code 1-125. GitHub Actions stops executing a script if any command returns a non-zero exit code.xargs man page:xargs exits with the following status:[..]123 if any invocation of the command exited with status 1-125
What's wrong, why I've getting an error?Error: Process completed with exit code 123.//# This is a basic workflow to help you get started with Actions name: CI //# Controls when the action will run. Triggers the workflow on push or pull request //# events but only for the main branch on: pull_request: branche...
Actions in Github
Try:cron(0 0 1 */3 * *)cron(0 0 1 1 * *)
I'm trying to create CRON expressions in YAML:1. AWS Cron expression to run at 6AM every month: This what I have: cron(0 6 1 * ? *) 2. Run every 3 months 3. Run once a year in JanuaryCan someone please help with this. Appreciate it greatly
AWS cron expression to run every month, every 3 months and once a year
There's an example in OpenShift:https://github.com/openshift/postgresql/tree/master/examples/replicaThe principle is the same in pure Kube (it's not using anything truly OpenShift specific, and you can use the images in plain docker)
I'm still wrapping my head around Kubernetes and how that's supposed to work. Currently, I'm struggling to understand how to model something like a PostgreSQL cluster with streaming replication, scaling out and automatic failover/failback (pgpool-II,repmgr, pick your poison).My main problem with the approach is the dua...
How do I model a PostgreSQL failover cluster with Docker/Kubernetes?
The problem was a misunderstanding... I needed to get RAW Contact ID and I used ContactID.. so, before searching for GroupId I needed to obtain Contact Raw Id
Basically things are more like black and white, on one phone (Galaxy S) works fine and on another (Nexus one, my client's of course) it doesn't. First I show a list of Contacts that have phone numbers. The user chooses to backup a contact and I try to load all contact info to store it in a local database cursor = cont...
Error retrieving contact group id in Android 2.1+
Guava's Cache type is generally intended to be used as a computing cache. You don't usually add values to it manually. Rather, you tell it how to load the expensive to calculate value for a key by giving it a CacheLoader that contains the necessary code. A typical example is loading a value from a database or doing an...
I'm new to Google's Guava library and am interested in Guava's Caching package. Currently I have version 10.0.1 downloaded. After reviewing the documentation, the JUnit tests source code and even after searching google extensively, I still can't figure out how to use the Caching package. The documentation is very shor...
Can someone help me understand Guava CacheLoader?
Windows Scheduled Tasks. This area is available through the control panel and may require administrative rights. It's basically crontab for Windows. Run a Windows batch file (or any executable) at a specified interval using a specified user.
What is the best way of scheduling simple batch jobs on windows? My current need is to dump some SVN repositories each night and FTP the dumps to our external FTP side and then e-mail one or more people that the backup has completed/failed.This sounds like the sort of thing I would usually do in a unix script + cron. ...
Scheduling jobs on windows
Queue System with Workers:Laravel provides a robust queue system that can help you offload time-consuming tasks to background workers. Instead of running your data-fetching logic directly in your controller or a job, you can dispatch a job to the queue, and then have a worker process the job asynchronously. This can he...
I have created a Laravel app and its main functionality is getting data from another API, feeding the Laravel app database tables and as of now, i have achieved so by writing methods inside my controllers and scheduled them in cron commands like0 */1 * * * curl https://{app_url}/projects > /dev/null 2>&1So there is a c...
Laravel Optimize developer experience for cron jobs and scheduler
+50Did you change the avatar from the random graphic to the custom versionafteryou already used your copy of GFW to download the avatar? If so then it could be that your client is caching the avatar.You might try looking for the cache--maybe it's in yourProgram Data\GitHubdirectory or something along those lines. (I'm ...
I have following issue.I have set up everything on github for using a avatar instead of this random grafic. If I commit and my colleague views the commits in his github for windows (GFW), it shows him my avatar. But if I launch my GFW (same version like my colleague) than I get the random grafic.There isa question that...
Show Avatar to Github Username after Committing
First check that you are getting reasonable IO performance from your disk setup. Then check that you PostgreSQL installation is appropriately tuned. In particularshared_buffersshould be set correctly,maintenance_work_memshould be increased during the restore,full_page_writesshould be off during the restore,wal_bufferss...
When I began, I usedpg_dumpwith the default plain format. I was unenlightened.Research revealed to me time and file size improvements withpg_dump -Fc | gzip -9 -c > dumpfile.gz. I was enlightened.When it came time to create the database anew,# create tablespace dbname location '/SAN/dbname'; # create database dbname ...
PostgreSQL: improving pg_dump, pg_restore performance
1 I think it is depends on browser internal implementation and handled by browser, The end user of browser is not allowed to configure it. Share Improve this answer Follow answered Apr 11, 2012 a...
Is it possible to limit an amount of memory that can be allocated by a third-party javascript code? UPDATE: The ideal situation is to have a 'sandbox' where: Some user adds a piece of javascript code. This piece of code is being executed. If it allocates more than 5 mb - user gets an exception Thanks in advance. :...
Limit an amount of memory for a third-party javascript
([-A-Za-z0-9\+\.]+)-(20[-0-9]{2})*You need to also make the-(hyphen) delimiter (before the last "year") optional. But only optional when the "year" is present and do not include this in the capturing subpattern.You should also anchor the regex (although this may depend on yourRewriteRuledirective), since you need to ma...
Can anyone help me please? It is for mod-rewrite in .htaccessFrom this url file name:car-audi-tt-2I need to get the full string.But if the string i get is this:car-audi-tt-2-2019I need to get (with the same regular expression) "car-audi-tt-2" in $1 and "2019" in $2.Any idea?Now i have this code, which it is working jus...
Same regular expression to get 1 or 2 parts of filename
If you're building a new operator and you plan to use Operator SDK, then I recommend reading the officialOperator SDK: Go tutorial. You can find another example of Go-based operatorhere.Concerning the kube config, it will use your default location if you don't specify anything. So the default kubeconfig yourkubectlis c...
I am building an operator using operator-sdk version 1.2 and I do understand we have a reconciliation loop but I was referring to some GitHub repos and unable to make out the use of the NewController function. It seems that these GitHub repos are developed using operator-sdk but in operator-sdk 1.2, I do not need to fi...
Operator-SDK and NewController function
I finally got it working by using ROOT user in Dockerfile and setting the npm cache to root. The problem was that git was using the /root/.ssh folder while npm was using a different path - /home/.ssh as it's npm cache was set on /home/.ssh For anyone still struggling, this is the config I used Docker Build Command: DO...
I am trying to automate a docker build in Jenkins pipeline. In my dockerfile, I basically build a node application. In my npm install, I have some private git repositories which need os bindings and so have to be installed in the container. When I run this manually, I transfer my ssh keys (id_rsa) to dockerfile which ...
Pass ssh-agent to dockerfile to install private repository modules
As mentioned in the officialdocumentationwhen you consider to allocatePersistent volumesto the Pods in the cluster there is a requirement to specifyStorageClassin order to find appropriate provisioner (volume plugin) for the storage provider.StorageClassdefines all the necessary parameters have to be passed to the stor...
Kubernetes Dynamic Volume Provisioninggives a handy way to supply pods with dynamically-allocated storage volumes. For example,NFS Provisionertransparently spins up an NFS server and exposes that storage to client pods with Kubernetes volume interface, on-demand.But how efficient is that? Does provisioner introduce ano...
How efficient is Kubernetes Dynamic Volume Provisioning?
Stackoverflow Question Archive:How to get domain name from url?Check if domain equals value?How do I get the base url?print get_domain("http://somedomain.co.uk"); // outputs 'somedomain.co.uk' function get_domain($url) { $pieces = parse_url($url); $domain = isset($pieces['host']) ? $pieces['host'] : ''; if (preg...
I need to write a function to parse variables which contain domain names. It's best I explain this with an example, the variable could contain any of these things:here.example.com example.com example.org here.example.orgBut when passed through my function all of these must return either example.com or example.co.uk, th...
PHP Getting Domain Name From Subdomain
...In Android, I see people programmatically install keystore in the following way (The code is from Android developer blog):byte[] keystore = . . (read from a PKCS#12 keystore) Intent installIntent = KeyChain.createInstallIntent(); installIntent.putExtra(KeyChain.EXTRA_PKCS12, keystore); startActivityForResult(inst...
I'm having trouble loading a .p12 certificate to my Android project. Here is a chunk of source code:char[] password = "<my pass>".toCharArray(); FileInputStream fIn = new FileInputStream("<name of cert>"); KeyStore keystore = KeyStore.getInstance("PKCS12"); keystore.load(fIn, password);On line 2 an error occurred openi...
Android: Installing .p12 certificate into device programmatically [duplicate]
Place general rules in:/.htaccessPlace /admin/ specific rules in:/admin/.htaccessPlace /admin/tool/ specific rules in:/admin/tool/.htaccessShareFollowansweredJan 20, 2010 at 17:42Alix AxelAlix Axel153k9797 gold badges396396 silver badges501501 bronze badges3does this mean if I want to change the max upload size for a t...
I have some settings which I want to use in .htaccess file. The settings are for some functionality at/admin/toolfolder level .. but I want to include some settings for/adminand/locations as well.My question is what is the best location to put this file at??ThanksUpdateJust fyi .. I want to apply settings likemax file ...
Where to put .htaccess file?
I think you may be confusing what is and isn't available (and I think that answer you linked is too). You can absolutely use variables to populate what runner your job should run with, I use it today in workflows on the SaaS offering.Using variables to determine the runner tag can still be confusing though, because whe...
I found thefollowing proposaland tested it out (see code sample), but could not make it work.We run on Gitlab 14.3.4, how can I determine if this is available for this version? If this feature is not working, how can I deploy to different environments if I have different runners one for my prod one for dev environment?...
How to deploy to different enviroments based on workflow variables?
First of all if you use a placementnewthen you will need to call destructor explicitlybuf[var].~BaseClass();Then you can delete just things that have been allocated with a new, while&buf[0]works since it's the address returned by the placement new,&buf[1]has not been directly allocated by the memory manager through::op...
I want to allocate memory for a huge bulk of objects. Then construct them one by one. So I do the following:BaseClass* buf = static_cast<BaseClass*> (::operator new (sizeof(BaseClass[5]))); for (int var = 0; var < 5; ++var) { new (&buf[var]) BaseClass(var); }And everything seems ok. But when I add delete:BaseClass*...
C++ "new" memory allocation
61 I found the solution to this problem in this github-issue The fix was simple. Amplify docs do not tell you to load configs of aws-exports to Auth module. Adding this simple line of code in App.js, solved the issue for me. import Amplify, { Auth } from 'aws-amplify'; im...
First I have successfully completed configuring my react application using amplify configure. I did that with the help of AWS Amplify docs. Then I have successfully added authentication to my amplify project, using amplify add auth and amplify push. I followed all the steps in the AWS - Authentication with Amplify Doc...
AuthError - Error: Amplify has not been configured correctly
The solution is quite simple. You can convert avclMatrixorgpuMatrixto amatrixby simply passing square brackets.r_matrix <- multmatrix[]In the most recent version on the packagegithubversion 1.2.2 has a convenience wrapper foras.matrix. It will be available in the next release on CRAN.
I have been using the package 'gpuR' to increase speed of code manipulating matrices. A matrix (formed by as.matrix(x)) can be converted to a vclMatrix using the code below (vcl indicates the data is stored on the GPU)vclm <- vcLMatrix(x) multmatrix <- vclm * vclmis much quicker thanmultmatrix <- x * xfor a large mat...
How to convert a GPU (gpuR) matrix to a standard R matrix?
If you have merge upstream/master and now have a .gitmodules declaring a FerexRev suvmodule, do first:git submodule init git submodule update # or git submodule update --initThen you can check if the submodule is present.You can alsomake that submodule follow a branch.
I have used the following commands but submodules are not fetched :(git remote add upstream https://github.com/NecronomiconCoding/NecroBot git fetch upstream git merge upstream/master git submodule foreach git pull upstream masterHowever the submoduleFeroxRev @ 33654cdis not fetched into my local repository :(So how...
How to fetch submodules in github
To add a new master to kubernetes cluster, you need to copy four files from your existing kubernetes master certificate directory before doingkubeadm initon new master. Those files areca.crt, ca.key, sa.pub, sa.keyand copy these files to/etc/kubernetes/pkifolder on new master. If you don't copy thesa*files your kuberne...
I am trying to add new master , and I copy cert and keys i.e./etc/kubernetes/pki/apiserver-kubelet-client.crtfrom current master to a new one. I noticed that after I do 'kubeadm init --config=config.yaml' this key (probably all of them) is changing (kubeadm inititself is successful).. Why is this happening and could it...
Unauthorized issues when adding new kubernetes master
+25You can try all possible fixes:Try disabling malwarebytes/firewallTry resetting DNS clientTry flushing DNS cacheRemove IPv6 option under particular connectionSometimes i faced few strange issue,Skype was working fine, but, i was not able to browse in browsers due to dns issue. so it worked when cleared all temp fil...
I'm having a very weird problem with a new Windows 10 Pro build using an intel I219-v adapter (integrated into the mobo). Every time a DNS lookup occurs, either in a browser or triggered manually via command prompt, the first request times out and then the next resolves properly. I only have the two google servers (8.8...
First DNS Lookup always fails
Since you have no knowledge of how the instance was setup, the only choice is tocreate an Amazon Machine Image (AMI). This will create an exact copy of the disk, so everything you have installed will be available to any new instances launched from the AMI. The CloudFormation template can then be configured to launch in...
Problem:I have anEC2instance running and I have made some modifications to the instance: installed docker, setup directories for certs, etc. Now, I am wanting to create the same instance but use infrastructure as code principals. Instead of remembering all the additions that I have done and creating a template by hand,...
Export Existing EC2 instance to CloudFormation json/yaml
I found out why. Curiously, no password was set for my account.
I created a GitBook and want to export it to GitHub. I went into the settings of the GitBook and clicked on Export to GitHub. Provided the URL to my GitBook repo, entered a valid name for my new GitHub repository and clicked on Begin Import. When I do so I get a message saying "Your old project requires credentials fo...
Gitbook Export to Github
You can create deployment with 1 replica with the required environment variable and another deployment with as many replicas you want without that variable. You may also set the same labels on both deployments so that Service can load balance traffic between pods from both deployments if you need it.
I have n instances of my micro-service running as kubernetes pods but as there's some scheduling logic in the application code, I would like only one of these pods to execute the code.In Spring applications, a common approach is to activatescheduledprofile-Dspring.profiles.active=scheduledfor only instance & leave it d...
Set environment variable for a single pod in a cluster
Check your Apache config file (httpd.conf) and make sure the directory you are using for your site includes the AllowOverride option.Example:<Directory "/Applications/MAMP/htdocs"> Options All AllowOverride All Order allow,deny Allow from all </Directory>
I have the file structureindex.php .htaccess news/index.php news/.htaccessFirst .htaccess:RewriteEngine On RewriteBase / RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_URI} !^/news/ RewriteRule . /index.php [L]Second (news/.htaccess)RewriteEngine On RewriteRule . /index...
Rewrite rule for subfolder
While creating an ec2 machine for an external NFS-server, you must add it to the vpc used by the eks cluster and include it in the security group that nodes use to communicate with each other.
I am using AWS EKS. As i am trying to mount efs to my eks cluster, getting the following error.Warning FailedMount 3m1s kubelet Unable to attach or mount volumes: unmounted volumes=[nfs-client-root], unattached volumes=[nfs-client-root nfs-client-provisioner-token-8bx56]: timed out waiting for the condition Warnin...
AWS NFS mount volume issue in kubernetes cluster (EKS)
Could you please clarify which line in your Dockerfile causes the error message? Is the file you are trying to copy from your working directory yolov3-tiny_obj.cfg? If that is the case, it fails because you specify to copy it from the builder stage. The line should probably look like this: COPY yolov3-tiny_obj.cfg /pa...
I am trying to copy a file into my docker container but the command fails. The file is in the same directory as the Dockerfile, so I don't understand the reason for the error. I'd appreciate any help or advice. Thanks beforehand. This is the code: FROM ubuntu:20.04 as builder ENV DEBIAN_FRONTEND=noninteractive RUN a...
Docker: COPY failed: stat <file>: file does not exist
Rather than using Redis or Memcached for caching, plus a "blobstore" package to store things on disk, I would suggest to have a look atCouchbase Serverwhich does exactly what you want (i.e. serving hot blobs from memory, but still storing them to disk).In the company I work for, we commonly use the pattern you describe...
Setting up a data warehousing mining project on a Linux cloud server. The primary language is Python .Would like to use this pattern for querying on data and storing data:SQL Database- SQL database is used to query on data. However, the SQL database stores only fields that need to be searched on, it does NOT store the ...
A good blobstore / memcache solution
Are you using docker toolbox for windows. Docker toolbox on windows runs docker in a Linux based virtualbox vm. You might have to open the port in the virtual box VM itself. So your host os can access the guest os's port. See https://www.simplified.guide/virtualbox/port-forwarding
I built a docker image from a dockerfile. I see the image was built successfully ( $ docker images) and when I use this command to run the image as a container : $ docker run -i -t 8dbd9e392a96 My application was running successfully, but when I'm trying to open I've this message This site can’t be reached T...
Run a Docker Image as a Container (for windows users)
Ok. So after I posted this I dug into the GWT source code and found some links about creating a GWT custom linker.http://development.lombardi.com/?p=29http://code.google.com/webtoolkit/doc/1.6/DevGuideOrganizingProjects.htmlHere's how I solved it with my own linker. First I made a linker class that extends the standa...
I've got an app written with Struts/Tiles/JSP that I'm adding a GWT app to. The non-GWT portion of my app handles css caching by actually writing out the css file with a version number taken from my svn repository attached, like this "styles.css?svnbuild=12345". That way I can tell the browser to cache those css file...
How to handle caching of GWT theme CSS files
In case of usingdf.persist()ordf.cache()then you should be also usingdf.unpersist()method and there's alsosqlContext.clearCache()which clears all.
I am facing thejava.lang.OutOfMemoryError: Java Heap Spaceissue when I run the same spark program every 2nd time.Here is a scenario:When I do thespark-submitand runs the spark program for the first time, it gives me the correct output & everything is fine. When I execute the samespark-submitone more time, it is throwin...
Apache Spark : java.lang.OutOfMemoryError: Java Heap Space issue
According to officialFluent Bitdocumentation, for the moment it is actually the unique way of requesting that the log processor skips the logs from certain Pods. I searched through it and found nothing butthisfragment.In addition to that, there is even afeature requestraised on their GitHub project so for now we can ho...
Is there a way to exclude certain namespaces influent-bit? I would like to exclude certain namespaces, so thatfluent-bitdoesn't forward all logs created in those namespaces to ELK.Is there a way to do it besides adding annotation to each pod in that namespace? I'm aware that you can update all of the pods annotations i...
How to exclude namespace from fluent-bit logging
Additionally to the previous answer you can use:kubectl get nodes -o jsonpath='{.items[*].spec.podCIDR}'to get pod CIDR addresses for each of the nodes in your cluster.
Is it possible to list the kubernetes CNI and pod-network-cidr details used on kubernetes cluster? Preferably usingkubectl.
How to list pod-network-cidr details
+500I would start with the most simple approach and that is to place the write and read side together because they belong to the same bounded context.Then in the future if it is needed, then I would consider adding more read side or scaling out to other regions.To get started I would also consider adding the ReadSide i...
I am re-designing a dotnet backend api using the CQRS approach. This question is about how to handle the Query side in the context of a Kubernetes deployment.I am thinking of using MongoDb as the Query Database. The app is dotnet webapi app. So what would be the best approach:Create a sidecar Pod which containerizes...
Kubernetes deployment strategy using CQRS with dotnet & MongoDb
According tooazure/container-registry| Microsoft Docs.Azure Container Registry does not officially support the Notary CLI but is compatible with the Notary Server API, which is included with Docker Desktop. Currently Notary version 0.6.0 is recommendedSo please try the suggestion provided in comment by @madhuraj.It loo...
I am unable to run a health check using the Azure CLI. I am on MacOS Monterey. Below is the error:➜ az acr check-health -n <ACR_NAME> Docker daemon status: available Docker version: 'Docker version 20.10.12, build 459d0df, platform linux/amd64' This will pull the image mcr.microsoft.com/mcr/hello-world:latest. Proceed?...
Unable to run az acr check-health command on MacOS
Instead of trying to make lynx got to the website that was hosted locally, All I had to do was run the php file.my command was not running because apache wasn't up yet.
trying to program a cron jobhere it is,@reboot lynx -dump ft-server > /dev/nullWhen I run the job manually via webmin, all is well, does what it is supposed to, no error. But when I reboot the system, it does nothing.Help!
Cronjob runs when manually started, but not on each reboot
Posting this as a community wiki, feel free to edit and expand it for better experience.As David Maze and Lucia pointed out in comments, services are accessible by their IPs and URLs based on service names.This part is covered and well explained in official kubenetes documentationDNS for services and pods
I have a use case wherein I have a Rest API running on a POD inside kubernetes cluster, and the helm pre-upgrade hook which runs a k8s Job needs to access Rest API, What is the best way to expose this URL so that helm hook can access it. I do not want to hardcode any Ip.
How to expose kubernetes POD's Rest URL to a helm hook?
Cloudwatch does not send notifications continuously if the threshold is breached. Cloudwatch can send a Notification only when the state changes.Alarms invoke actions for sustained state changes only. CloudWatch alarms do not invoke actions simply because they are in a particular state, the state must have changed and ...
I have an instance in AWS that from time to time it's CPU cross the threshold of 90%. I have created an alert for this, however I saw that I received one notification only and it was during the first 5 minutes while the CPU was at 100% for 2 hours. How do I set the metric so I will keep getting notifications all the ti...
Continuous alerts in Cloudwatch
I just built-up the connection. It's an issue caused by latest version of ELK. (7.5.0) Here's work around in my article. The point is to copy the JDBC driver .jar to the logstash-core\lib.Check here:My blog of workaround solution before Elastic stack solved their issue-- MUST remain:jdbc_driver_library => “"Logstash co...
I have got below error when I am trying to connect with SQL Servererror_message=>"Java::JavaLang::NoClassDefFoundError: javax/xml/bind/DatatypeConverter"} logstash_1 | [2019-12-05T07:52:10,649][ERROR][logstash.javapipeline ] A plugin had an unrecoverable error. Will restart this pluginMy ELK version is 7.4
Logstash to SQL Server connection fail
It is not possible. But why are you disturbed? It's just a message. If you create agh-pagesbranch, the message is hidden but you can do a pull request.
I like howgh-pagesis just a branch of a repository. I want to create a new branch to store things that related to a repository, but should not be updated very often, so I don't want to store inside the master branch.Likegh-pages, I would like to create anexamplebranch, that won't ever be offered to merge. Ideally I don...
Adding a non mergeable branch like gh-pages
Edit: This has been implemented here:https://github.com/kubernetes/kubernetes/pull/60452Currently, there is not yet a way to select certain pod from another namespace. There is an open issue for thathttps://github.com/kubernetes/kubernetes/issues/50451
Is there any ability to filter by both namespaceandpod's labels at the same time?The example present in documentation athttps://kubernetes.io/docs/user-guide/networkpolicies/#the-networkpolicy-resource- from: - namespaceSelector: matchLabels: project: myproject - podSelector: matchLabels: role: fron...
Kubernetes network policy to filter on both namespaces and pod's labels
If there is no way to run agit describe --tags(in order toget the latest tag), one workaround would be:each time you apply a tag, you reset a dedicate "tags" branch to that tagin your GitHub Action workflow, you would "use" the "tags" branch, which by its construction, would reference the latest tag.
Since the documentation for GitHub actions is far far behind, I couldn't find anything about it online.Is there a way for GitHub actions to select the latest tagged version?If I want to select a version I would do:uses: MarvinJWendt/[email protected]If I always want the latest commit I would use the master branchuses: ...
How to automatically select the latest tagged version of an GitHub action?
You should only test if it's greater than zero:if (next.compareTo(maximum) > 0) { maximum = next; }From theAPI docsof the compareTo Method:Returns a negative integer, zero, or a positive integer as this object is less than, equal to, or greater than the specified object.
I have the following line of code (where "next" is a BigDecimal):if (next.compareTo(maximum) == 1) { maximum = next; }On the equality comparison, SonarQube gives me this warning:Only the sign of the result should be examined.sonarqube-injectWhat does it actually mean and how can I fix that ?
In Java with SonarQube, how to fix `Only the sign of the result should be examined`
Your yaml file looks good.Using this configuration:apiVersion: v1 kind: Pod metadata: name: galahad-test-distributor namespace: galahad-test spec: volumes: - name: place-for-stuff hostPath: path: /mnt/galahad containers: - name: galahad-test-distributor image: busybox args: [/bi...
I have a single node Kubernetes cluster. I want the pod I make to have access to /mnt/galahad on my local computer (which is the host for the cluster).Here is my Kubernetes config yaml:apiVersion: v1 kind: Pod metadata: name: galahad-test-distributor namespace: galahad-test spec: volumes: - name: place-for-s...
Hosting local directory to Kubernetes Pod
The final argument todocker runis the command to run within the container. When you rundocker run -d -p 80:80 ubuntu:14.04 /bin/bash, you're runningbashin the container and nothing more. You actually want to run your web application in a container and to keep that container alive, so you should dodocker run -d -p 80:80...
I would like to run a docker container that hosts a simple web application, however I do not understand how to design/run the image as a server. For example:docker run -d -p 80:80 ubuntu:14.04 /bin/bashThis will start and immediately shutdown the container. Instead we can start it interactively:docker run -i -p 80:80 u...
How to start docker container as server
0 Clone URL pointed red. Some useful information can be on project wiki. Share Improve this answer Follow answered Mar 23, 2016 at 21:07 galeksandrpgaleksandrp 77688 silver badges1717 bro...
Having trouble grasping github. I am on linux and in every tutorial for github I watch people go to terminal, and type git clone SomeURLHere. 1) How do they find that url? 2) Only some pages on github give you really clear step by step instructions for idiots like me on how to clone and install their software from g...
how do you find the url of a github repository to clone, and then how do you finish installing the cloned repository
If a row changes while the backup is going on, the new value may or may not be in the backup. This is generally OK because RethinkDB only offers single-row atomicity anyway, but if you have a workload where that isn't OK then your other options are to use a filesystem that lets you snapshot the data on disk, or to a...
I read article about backing up data, but some issues is not clear for me: What happens with data, that will be changed after backup process was started? Does backup operation work only on current machine? Or will it collect data from all shards in cluster? If only on current, should I start backup process on all ser...
RethinkDB backup data
ForkedRunner.create().addProperties(properties).execute(); That's all I had to do and it is working fine now. Thanks :)
I have a set of project's on which I need to execute Sonar. Once I execute sonar task and it publishes the result it deletes all the files in the directory .sonartemp except the file persistit_tempvol. This file is locked by Java SE runtime and I am unable to delete this. The reason I want to delete this file is once ...
Unable to delete the persistit_tempvol file after Executing sonar Embedded Runner from Java
I was able to read them thru environment variable:@Autowired private org.springframework.core.env.Environment env; //inside some method @GetMapping("/test") public String print(){ System.out.println(env.getProperty("mysecret2")); }I tried using below but that didn't work.@Value("${mysecret2}) private String mySecre...
I have secrets configured in config/yaml file. There is one secret value that is causing trouble. I just want to print out the value being injected:apiVersion: v1 kind: ConfigMap metadata: name: myapplication-config data: config.yaml: | 'mysecret1': ${DB_PASSWORD} 'mysecret2': ${ANOTHER_SECRET}I make a GET re...
Read secrets in Springboot deployed on Kubernetes
UseCronScheduleBuilderto make it easy to configureCronTrigger:// Extension method to round a date to a given part var nextFire = DateTime.Now.Truncate(TimeSpan.FromHours(1)); var trigger = TriggerBuilder.Create() .WithSchedule(CronScheduleBuilder.DailyAtHourAndMinute(nextFire.Hour, nextFire.Minute).WithMisfi...
public static void ConfigureJob() { int ScheduleTriggerDate = Convert.ToInt32(ConfigurationManager.AppSettings["ScheduleTriggerDate"]); ISchedulerFactory sf = new StdSchedulerFactory(); IScheduler sched = sf.GetScheduler(); IJobDetail job = JobBuilder.Create<MailingList>().WithIdentity("job1", "group...
Quartz.Net CronTrigger scheduling not working as expected
This variable is set byrootdirective. You cannot use it inrootdirective itself, because it will lead to infinite loop.Seehttp://nginx.org/r/rootThepathvalue can contain variables, except$document_rootand$realpath_root.Use your own variable instead.set $my_root folder/my_root; root /$my_root; ... location = /404.html {...
While setting up my nginx configuration I came upon this. Does anyone have any idea on why this happens exactly?root /folder/my_root; index index.php index.html index.htm; error_page 404 /404.html; location = /404.html{ root $document_root/error_pages; //FAILS HERE with the error in the title internal; }
The $document_root variable cannot be used in the "root" directive
After much research, I can conclude that it is not possible out of the box.Update- you can use openresty which comes with Lua. Using Lua one can do pretty cool things, including logging all of the headers to say, Redis or some other server
How do I log all the headers the client (browser) has sent in Nginx? I also want to log the response headers. Note that I am using nginx as reverse proxy.After going through documentation, I understand that I can log a specific header, but I want to log all of the headers.
How to log all headers in nginx?
Your OS may have a default limit on the size of a user process. On Linux you can change this with ulimit. You probably want to iterate over these 64000000 numbers without needing them all in memory at once. Lazy lists let you write code similar in style to the list-all-at-once code: -module(lazy). -export([seq/2]). s...
I have just started learning Erlang and am trying out some Project Euler problems to get started. However, I seem to be able to do any operations on large sequences without crashing the erlang shell. Ie.,even this: list:seq(1,64000000). crashes erlang, with the error: eheap_alloc: Cannot allocate 467078560 bytes of ...
Why is Erlang crashing on large sequences?
35 If it is a corporate repo, you might want to ignore proxy settings. One possible solution to your problem is here: Ignore proxy: export no_proxy=YOUR_CORP_DOMAIN_ON_GITHUB, where the domain name might be in the form of github.acme.net Ignore SSL verification: git config...
I am trying to clone from a corporate git repo, but always receive this error message after a while: fatal: unable to access URL: Received HTTP code 503 from proxy after CONNECT I have the following .gitconfig file: [https] sslVerify = false proxy = https://proxy.corpadderess:8080 [http] sslVerify = false ...
Corporate Github behind proxy: Received HTTP code 503 from proxy after CONNECT
TheIntelliJ documentation mentionsIf you do not want to specify your credentials each time you sync with a remote, or push your commits, you can configure IntelliJ IDEA to save your account information (seeConfigure a password policy).Maybe it is possible said policy is not set properly.On Windows, check you have selec...
I am working on a hybris project using IntelliJ as editor along with GitHub version control. I have added my user login in IntelliJ via File -> Setting -> Version Control -> GitHub and also added the user in Windows -> Credential Manager -> Windows Credentials. I keep on getting pop up in IntelliJ saying Login to GitHu...
github not signing in with intellij
If it were the case that you were statically displaying the same image, than using a texture or display list would suffice. But as you want to frequently update it, shaders really are the best option. Shader code executes on the GPU and modifies data in GRAM, so you have no bottle neck transferring from CPU to GPU. Th...
I'm writing a game in Java, LJGWL (OpenGL). I'm using a library that handles a lot of messy details for me, but need to find a lot faster way to do this. Basically I want to set every pixel on the screen to say a random color as fast a possible. The "random colors" is just an Array [][] that gets updated every 2-3 ...
OpenGL Shaders?
This can be done easilyCreate a repo for username.github.io (you don't have to use it)Create another repo for your project (name doesn't matter) and setup github pages on thatSetup custom domain name for that specific project using a CNAME or use username.github.io/project
I want to create a project page that is its own thing. I see fromGitHub Pages documentationthat options arehttp(s)://<username>.github.io/<projectname>orhttp(s)://<orgname>.github.io/<projectname>But I would likehttp(s)://<projectname>.github.ioI seem to recall seeing others do this, e.g. to accompany their submitted r...
How to create <projectname>.github.io?
After you setup themongodbyou usually wouldenableit, so that it starts after each reboot:sudo systemctl enable mongod
I am using aws and installed ubuntu server on aws ec2. I installed mongodb on it. When installing mongodb, it is working well. But If i reboot server and try to connect mongod in termina. I am getting the following error.MongoDB shell version v4.4.0 connecting to: mongodb://127.0.0.1:27017/?compressors=disabled&gssapiS...
Mongodb connection error whenever rebooting server
3 I don't think the "job % complete" is available to us according to the docs: http://docs.aws.amazon.com/elastictranscoder/latest/developerguide/elastictranscoder-dg.pdf The JSON status update only has the "PROGRESSING" state which means very little for a user. Amazon is...
I'm using read_job() API to monitor jobs' progress. This only gives me a very coarse info about the status being one of the: Submitted | In Progress | Complete | Canceled | Error. However, I'd like to show the user a percentage or ETA when in the In Progress state. Is there any API which gives such a fine grained pro...
Is there a way to get fine grained progress of Elastic Transcoder job?
Avoid spaces in whites and if it is in the same directory remove the spaces bar : ❌![Front-end & Back-end Architecture](/frontend-backend view.png) ✅ ![Front-end & Back-end Architecture](frontend-backend-view.png) Jetbrains Bug This is due to the bug of the JetBrains plugin that obtains faults when rendering the imag...
I have read info about inserting images to Github readme.md file. It is described here: Markdown mastering. I tried following in my markdown file: ![Front-end & Back-end Architecture](/frontend-backend view.png) Locally it looks: And remotely it looks the same. Like some text instead of the image: The image is exact...
Inserting image to Markdown file doesn't work locally and remotely
You can use version as a query parameter, e.g. /resources/foo.js?_version=1.0.0. If you are using Maven, it is not that hard to get version information from /META-INF/maven/{groupId}/{artifactId}/pom.properties. Of course this will force reload all scripts with every new version... but new versions are probably not de...
We're trying to force the client's browser to reload static files when they change. Currently, we have our build scripts automatically creating a new directory with the build timestamp and replace the content in the code to point to that directory. However, I feel this is hardly an optimal solution. It forces the clie...
Java web app: Force browser to load static content (js, css, html) if deployed file changed
In a default vue-cli setup, npm start (the command you are using) runs npm run dev. And, again, by default, npm run dev binds to localhost only. Add --host 0.0.0.0 to your webpack-dev-server line in package.json so you can access it from outside the docker container: From something like: "scripts": { "dev": "web...
I am unable to access the site locally on the http://172.17.0.2:8080/ in Chrome, I get "172.17.0.2 took too long to respond". I used the inspect command to obtain the IP address of the container. docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}} e83c95d05d63 The run command that I used. dock...
How to Containerize a Vue.js app?