Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
I am Max from DataGrip team, and the correct answer is:It could be JDBC driver issue and the desired method hasn't been implemented yet. Since you're trying to run purecqlshcommand as SQL. Follow the issueDBE-10638.
I'm trying to perform inserts on Amazon's Managed Cassandra service from IntelliJ's DataGrip IDE, however I recieve the following error:Consistency level LOCAL_ONE is not supported for this operation. Supported consistency levels are: LOCAL_QUORUMThis is due to Amazon using theLOCAL_QUORUMconsistency level for writes.I...
Cannot set consistency level when querying Amazon Keyspaces service from DataGrip
This works: Debug.MemoryInfo memoryInfo = new Debug.MemoryInfo(); Debug.getMemoryInfo(memoryInfo); String memMessage = String.format("App Memory: Pss=%.2f MB\nPrivate=%.2f MB\nShared=%.2f MB", memoryInfo.getTotalPss() / 1024.0, memoryInfo.getTotalPrivateDirty() / 1024.0, memoryInfo.getTotalSharedD...
Memory seems to be a big topic and I cant find the specific answer. I've got the answers on how much is available in the heap and I know how much should I use. I need the answer how to code to programatically determine how much memory is my app using of the heap? And how much total memory am I using?
Android: How much memory is my app using?
There are more open source solutions that I could name, but for this application my choice would be rsync and a cron job. Here is a good overview of some open source options (some are more desktop oriented). EDIT The nice thing about rsync, is it can directly sync the folder storing the repo. The downside of this app...
CentOS 5.3 subversion 1.4.2 I forgot to add. Currently the total repository size is about 5GB, but that will grow over time. We have our source code and documents on our internal server running CentOS 5.3 and we are using subversion 1.4.2. We are looking for a backup strategy. We want to perform daily backups. We have...
backing up subversion repositories
As I said in comment it is not possible to handle removing with only docker command, but it is possible handle it with one line bash command: docker rm -f $(docker ps -a --format "{{.ID}}|{{.Image}}|{{.CreatedAt}}" | grep "imagename" | grep -v "$(date +%Y-%m-%d)" | awk -F'|' '{print$1}' | xargs) grep "imagename" → f...
I am trying to stop / prune docker containers according to 2 conditions: image name time they've been running for For instance, I could want to stop all containers from image image-name that have been running for over 24h. By reading the documentation, it seems the --filter option is not consistent at all since dock...
Docker filter containers both by image name and age
You don't seem to be actually defining a schema for your data. Please refer tothe API gateway documentationfor some examples of defining a model with JSON schema.
I am getting error while creating Model in aws api gateway, Json is in correct format and still it says validation error. Please anyone? where I went wrong?Error message "Invalid model specified: Validation Result: warnings : [], errors : [Invalid model schema specified]"Json Schema i am trying to use is:{ "action":...
AWS API Gateway Model : Invalid model schema specified
As per my comment, you must commit, stash, or discard all your changes before rebasing. If your changes are complete, you most likely want to commit them (git add --all followed by git commit). If they are not complete but you want to continue working on them later, use git stash to save your unfinished changes while ...
I have a trouble git rebase. because, I had pushed my local code to origin dev. And, I send a pull request to 'upstream'. But, There are 1 PR and 7 commits like below the picture I want to 7commits squash to 1 commit. What shall I do? ( I have tried : 'git rebase -i HEAD~7', but I have met the message that 'error: ca...
How can i do git rebase after git pull request?
Create 2 providers with same entity and different property:app_user_provider: entity: class: App\Entity\User property: email api_user_provider: entity: class: App\Entity\User property: userNameAutowire your custom authenticator to use the correct one:App\Security\AppLoginAuthenti...
I have 2 firewalls, one main one for the web/front end of the application. One for the api. They share a user entity, but i want the provider to load users based on a different property for each firewall.For the main firewall users can log in with email only. For the api firewall users can log in with username only.Her...
user provider for api and main firewall with different property
The error that you're getting is because you are trying to make an initial push to the remote repository without having a valid git commit. You ran the command git commit -m "Initial commit", without adding the files to the staging area by using git add, and hence there was no git commit made. You need to first add yo...
I have two folder which have two individual projects but inter-related to each other and both folder are in parent folder folder structure is as below BBYN |__external |__spfx I am trying to add git in BBYN folder but it is not adding. Help me to solve the issue
adding two folder in one parent folder in one repository github
You could go about it the other way around. Just filter for the values you want before applyingcount_values:count_values("bag_status", bag_status < 1)
a gauge named bag_status with label(a)usingcount_values ("bag_status" ,bag_status)it will count number for diferent valuesI want to filter bag_status with value <1(count_values ("bag_status" ,bag_status)){bag_status!=1}does not work
How to filter label name after count_values in promql
) Install APC (http://pecl.php.net/apc), this will compile your PHP script once and keep it in shared memory for the lifetime of the webserver process (or a given TTL).2) If your script fetches data and does not cache/store it on the filesystem, it will be in RAM, yes. But only for the duration of the request. PHP uses...
A have setup an internal proxy kind of thing using Curl and PHP. The setup is like this:The proxy server is a rather cheap VPS (which has slow disk i/o at times). All requests to this server are handled by a single index.php script. The index.php fetches data from another, fast server and displays to the user.The data ...
Apache, PHP caching
Try:RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME} !-f [OR] RewriteCond %{REQUEST_URI} \.(js|css|jpe?g|gif|png|bmp|ico)$ [NC] RewriteRule ^(.+)$ index.php?uri=$1 [QSA,L]So this will route throughindex.phpif the request isn't for an existing directory and, the request isn't for an existing file OR t...
This exclude CSS/JS/Images from original url. Anyone know how to solve this problem ?.htacess:RewriteEngine on RewriteRule %{REQUEST_FILENAME} !-d RewriteRule %{REQUEST_FILENAME} !-f RewriteRule ^(.+)$ index.php?uri=$1 [QSA,L]PHP:$splitarray = explode('/',$_GET['uri']);I explode URL using PHP to get 2nd parameter and i...
htaccess exclude CSS/JS/Images
When yourreact.jsapp loads, the routes are handled on the frontend by thereact-router. Say for example you are athttp://a.com. Then on the page you navigate tohttp://a.com/b. This route change is handled in the browser itself. Now when you refresh or open the urlhttp://a.com/bin the a new tab, the request goes to yourn...
I uploaded react.js application to a server. I'm using nginx server. Application is working fine. But when I go to another page & refresh, the site is not working. It's showing a 404 Not found error.How can I solve this?
react.js application showing 404 not found in nginx server
Changing the default docker registry is not possible. You can only configure a private registry to act as a mirror for the dockerhub registry as documented in Registry as a pull through cache. Check moby-33069 issue which has requested this feature.
I have docker machine and by default it pulls my images from docker hub. Now we have our own repo which can serve as a remote proxy to docker hub. We can pull with docker pull server/repo/image. Now I want that docker pull image resolves to our registry instead of docker hub. How can I achieve this in docker machine?
How to set a default registry to pull from in docker machine?
You cannot configure GAE cron services with resolutions below 1 minute. FWIW, you can't do that on unix/linux systems either.But it is possible to use anevery 1 minutescron job from which you can further trigger delayed execution of deferred/push/pull queue tasks with down to 1 second resolution, seeHigh frequency data...
When setting up a Google App Engine instance you can configure acron.yamlto set up Cron jobs.There does not seem to be any documentation on how to configure jobs that run say every 30 seconds.I triedschedule: every 30 secondsandschedule: 0/30 0 0 ? * * *But no good. Google Cloud tells me the format is incorrect when I ...
Google App Engine - How to set up Cron job using seconds
In my own words, after reading the docs and experimenting:You can useRewriteBaseto provide abasefor your rewrites. Consider this# invoke rewrite engine RewriteEngine On RewriteBase /~new/ # add trailing slash if missing rewriteRule ^(([a-z0-9\-]+/)*[a-z0-9\-]+)$ $1/ [NC,R=301,L]This is a real rule I used t...
I have seen this in a few.htaccessexamplesRewriteBase /It appears to be somewhat similar in functionality to the<base href="">of HTML.I believe it may automatically prepend its value to the beginning ofRewriteRulestatements (possibly ones without a leading slash)?I could not get it to work properly. I think it's use co...
How does RewriteBase work in .htaccess
No. Without GPU drivers, windows will use an API standard called VESA that all modern GPUs implement that allow standard non-accelerated drawing.
I was told that GPU vendors make actual DirectX/OpenGL API implementation in their driver. However, Windows use DirectX to draw itself (OpenGL in Linux case), but OS can draw itself even when GPU driver is not installed. So, does Windows contain some base DirectX implementation that works on all GPUs?
OS basic graphic driver
Virtual addresses can't "work", if they don't exist. Virtual addresses are addresses inside an instance of virtual memory. Virtual memory is not automatically given by some hardware though, it's an implementation of the OS. If the installed OS on a computer implements virtual memory, then programs may access their vir...
I am reading memory management schemes in Operating System.I got confused from this Question How virtual addresses work on computers without virtual memory? Is this possible virtual address can work without virtual memory? I have little bit knowledge about OS so i asked this question.
How virtual addresses work on computers without virtual memory?
You can push the branch in terminal (command prompt) by using following commandgit push origin yourbranchnameorigin - your remote URLyourbranchname - the branch which you want to push
In my Android Studio project, I had one branch calledmyproj.I've been updating and committing to it for a few days. I decided I wanted to work on a new feature and I'm not sure where it will take me so I created a newbranch.When I committed and pushed my new changes however, it seems to have only committed it locally b...
Android Studio VCS won't push my new branch to Git
-1I also got the same issue but after configuring proxy it resolved. Proxy in Windows-> internet options-> connections -> LAN settings.
Pip version:9.0.1Python version:3.6.3Operating system:Win7Description:I'm trying to install python libraries in my laptop but run into an SSL certificate error when running pip -install for any library. I have tried several suggestions from other posts but all of them didn't work. Can someone kindly help me to resolve...
pip install issue with ssl certificate:[SSL: TLSV1_ALERT_ACCESS_DENIED]
No, there is no built-in command to pull all docker images at once. But you can try this (multiline) bash using docker --format : for image in $(docker images --format "{{.Repository}}:{{.Tag}}" | grep -v '<none>') do docker pull $image done Or in one line: for image in $(docker images --format "{{.Repository}}:{{....
Is there a command to update (pull) all the downloaded Docker images at once in the terminal ?
Update (pull) all docker images at once
Like any hardware device on a PC the graphics card will respond to reads and writes to certain memory addresses, and possibly input/output ports. The PCI bus defines how these are allocated. There are no specific CPU instructions to communicate with graphics cards, in the case of writing to memory locations it just us...
I have wondered for quite some time how CPU instructions can interact with the GPU. As I understand things, the CPU has a certain set of instructions (machine code) that it understands and performs and a driver is a piece of software that communicates via the CPU to the GPU. But how does this piece of software commun...
How does a graphics driver programmatically communicate from CPU to GPU?
In 11.10, 12.04 and later, you can achieve this by making the 'scripts-user' run 'always'. In /etc/cloud/cloud.cfg you'll see something like: cloud_final_modules: - rightscale_userdata - scripts-per-once - scripts-per-boot - scripts-per-instance - scripts-user - keys-to-console - phone-home - final-message Th...
I have an EC2 instance running an AMI based on the Amazon Linux AMI. Like all such AMIs, it supports the cloud-init system for running startup scripts based on the User Data passed into every instance. In this particular case, my User Data input happens to be an Include file that sources several other startup scripts:...
How do I make cloud-init startup scripts run every time my EC2 instance boots?
After some research, I found out that thedocker logs -tcommand prints out timestamps in UTC and there is no config to change that. However, you could use a little script referenced inhttps://github.com/docker/cli/issues/604, where you could just pipe the output and change the given timestamp.
My local timezone and docker container's timezone are all set to 'GMT+8:00'. But the 'docker logs -t' still shows timestamp of 'GMT+0:00'.the picture below is a part of output of 'docker logs -t'. The left timestamp is printed by docker, and the right timestamp is printed by application in container.
how to set timezone of 'docker logs -t'?
Rather than create the file in the directory, why not create it somewhere else? Then after the data's been written, just move it into the webroot and overwrite the previous set.Example:sh create_some_data.sh > /home/cronuser/my_data.html mv /home/cronuser/my_data.html /var/www/
I have a small program that runs and generates a new text dump every 30sec-1min via cron.program > dump.txtHowever I have another PHP web program that accesses the text dump in a read-only mode whenever someone visits the webpage. The problem is that I believe if someone accesses the website the very second the cron jo...
reading a file created from a cron job
Modify the program so it runs only if a particular file exists. Remove the file if you need to stop the program. (Or have it run only if the filedoesn'texist, andtouchthe file to stop the program.)If you're not able to modify the program, you can execute a shellifstatement as a cron command.
I have a program written in python,my program scrapes a value from some financial website every minute and pushes that value into my DB.My program takes like 1 or maximum 1.5 seconds to do this job. I have set a cron job to call my program every minute. I need to run my program in this way everyday from 09AM to 04PM. N...
How to stop minute cron job?
We implemented our own cache that just drops the data on the floor: namespace AppBundle\Factory; use Google\Auth\CacheInterface; class NullGoogleCache implements CacheInterface { public function get($key, $expiration = false) { return false; } public function set($key, $value) { ...
I am developing a web app that fetches and displays google analytics data for users that are not technical enough to do this themselves. To do this, I: 1) have users log in with OAuth 2) store the access token 3) create a Google_Client and give it this access token 4) use this Google_Client to fetch the analytics dat...
PHP Google API Client caching access token
The full CodeQL license terms are available here. The license states (emphasis added): Further, except (and only to the extent) permitted by applicable law or applicable third-party license, you will not (and have no right to): ... share, publish, distribute or lend the Software, provide or make available the Softwa...
I'd like to enable users of my service to write and execute CodeQL queries. I don't want to offload their execution to GitHub's / Semmle's servers. It's unclear whether this is doable, or whether I'd run into licensing issues. In the security lab, it says "CodeQL is free for research and open source". I do want for th...
Can GitHub's / Semmle's CodeQL Query Executor be Self-hosted
See the unit test page:http://docs.codehaus.org/display/SONAR/Unit+Test+SupportFrom that page:The C++ Plugin parses xunit compliant format using the sonar.cxx.xunit.reportPath. To use other formats, first they need to be converted using the property sonar.cxx.xunit.xsltURLFor convenience the following xsl are provi...
Is CppUnit the only C/C++ unit test framework currently available for use with Sonar?What would be involved in adding additional C/C++ unit testing frameworks? (e.g. how many lines of code is the CppUnit plugin, how reusable, etc.)
C unit test frameworks with Sonar
My problem has been solved. I modified launch.json and it worked. It seems that what was wrong was the directory specification of "cmd". "cmd" had to be set in Windows format. For reference, the directory structure is as follows: product_root ├── .vscode │ └── launch.json ├── product │ ├── rails_root │ │ ├── a...
I'm trying to remote debug Ruby(on Rails) code in a Docker container with Visual Studio Code. I do not have Ruby installed on my machine, just in Docker. VSCode: 1.32.3 OS: Windows 10 Profesionnal, 1709 Docker Desktop: 2.1.0.5 Community stable Docker: 19.03.5 Ruby: 2.6.5 The procedure is as follows: Docker contai...
Can't break on any breakpoint with Visual Studio code while remote debugging on Docker container
This may not be your issue, but that's the error I'd expect to see if you didn't have git installed. If you haven't installed it yet, make sure you install it with the command line options.
I am trying to install the gem fromhttps://github.com/raid5/paths_of_glory/tree/rails3When I put the following line of code in my gemfile as specified in step 1 of the readme:gem 'paths_of_glory', :git => 'git://github.com/raid5/paths_of_glory.git', :branch => 'rails3'I then run "bundle install" and get the following m...
Rails 3 install a branched gem from github
The "Mainframe" must advertise to Web that the value has changed.So you can, for example, implement another web service called "InvalidateCache()" that would empty the cache when called.When the rates change, "Mainframe" would call InvalidateCache service which would empty the cache so that subsequent request to the Ra...
Assume that you are developing a web application that shows currency exchange rates for a bank web site. The rates are stored in an IBM Mainframe in the bank and rates are available through web services. Each time a user accesses the rates page, the page makes a request to the mainframe. This generates too much...
ASP.NET caching with web services calls
Using yum (the Yellowdog Updater, Modified) in your Dockerfile has nothing to do with your host CentOS. It has to do with your base image used by your Dockerfile (FROM xxx). The error message that matters is: There are no enabled repos. You can see a manual resolution in "RHEL 7 - Solution to "There are no enabled re...
On Centos7.1 Docker host : I am building a docker image with Dockerfile having command RUN yum -y install deltarpm yum-utils --disablerepo=*-eus-* --disablerepo=*-htb-* --disablerepo=*-ha-* --disablerepo=*-rt-* --disablerepo=*-lb-* --disablerepo=*-rs-* --disablerepo=*-sap-* During the run of docker build command ...
Docker build error There are no enabled repos
OK. After a bunch of trial and error I answered my own question.The third line denotes that there has to be something in the URI in order to perform the redirect thus not redirecting if the url just contains the initial slash.RewriteEngine On RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_URI} (.*)$ RewriteR...
I've successfully modified my .htaccess file to remove trailing slashes on most pages but I'm wondering how to exempt my home page/directory? For example:domain.com/test/ successfully redirects to domain.com/testHOWEVER, when I hit my domain it will append the root documentdomain.com/ redirects to domain.com/index.php...
Remove trailing slash using .htaccess except for home / landing page
The problem was in the extra slash in my url. I changed url fromhttps://somesite.com/mywsdl/?wsdltohttps://somesite.com/mywsdl?wsdland the problem disappeared.
I made a wsdl using sun-jaxws. I created a web service client in Netbeans, and successfully called the wsdl web service. Then I configured my nginx server to access the web service by https. When I call the service over https I get the following error:com.sun.xml.internal.ws.client.ClientTransportException: The server ...
Connecting to webservice results in com.sun.xml.internal.ws.client.ClientTransportException: The server sent HTTP status code 200: OK
If the socket.lua is thediegonehab/luasocket, then it requires socket/core.soYou need to specify thelua_package_cpathcpath is for compiled shared library modules (.so), path is for text modules (.lua)/usr/local/lib/lua/5.1/?.so - most common value for cpathShareFollowansweredDec 16, 2015 at 21:24marsgplmarsgpl57222 sil...
I've built nginx from sources with lua support and I'm able to run server-side scripts like this:http { lua_package_path '/usr/local/share/lua/5.1/?.lua;;'; server { listen 80; location /hi { content_by_lua ' ngx.header["Content-Type"] = "text/plain;charset=utf-8" ...
How to access sockets from lua in nginx?
0 System user Shiny runs your shiny app. Check if user Shiny has the permission to write within srv/shiny-server. If not, the app run by user Shiny won't be able to write anything and the app would just crash without a error message. You can set the user permission in the...
I am currently trying to implement caching on my dockerized shiny app, but i am facing troubles, the app doesn't work and there's no log to trace the problem: Screenshot from the dockerized app I am using the rocker/shiny image, installing a few R packages (via the packages.R file) above and then running my shiny-app....
Cache data in Shiny Server on Docker
The short answer is, no there is not.cuPrintf and the built-in printf support in Fermi and Kepler runtime is implemented using device to host copies. The mechanism is no different to using cudaMemcpy to transfer a buffer to the host yourself.Just about all CUDA compatible GPUs support so-called zero-copy (AKA "pinned, ...
I am a newbie in CUDA programming and in the process of re-writing a C code into a parallelized CUDA new code.Is there a way to write output data files directly from the device without bothering copying arrays from device to host? I assume if cuPrintf exists, there must be away to write a cuFprintf?Sorry, if the answer...
Writing output files from CUDA devices
i was rather stupid on this one.I didn't give the member any permissions, so it couldn't access any repos. After adding it to the standard write group, everything worked as expected.
Ive created a member in our github organisation and would like to use them forAllow specified actors to bypass required pull requestsbut for some reason they do not appear in the list of possible users. I can see othermembersas well asadminsthough so not sure why they are not listed?Any ideas?
member not listed or available for github Allow specified actors to bypass required pull requests
8 Why do you need an android app for running an HTTP server? Can't you just run apache or nginx inside the docker container? If you want to run android apps on docker, you should look at this. https://registry.hub.docker.com/u/ksoichiro/android-emulator/ Share I...
Is it possible to run an android app in a docker container? The app should run a HTTP server, which should be accessible from other containers.
Is it possible to run android emulator / android apps in docker?
You seem to have understood. There is a float on the heap* and a pointer on the stack*. The disagreement is 'just' a naming convention for how you refer to the float.Some people talk of things-that-are-pointed-to in terms of the-thing-that-does-the-pointing. I am inclined to agree with you: this is potential confusing,...
I had a quiz on the memory allocation topic, and this question came across that got me confused as I am not satisfied with the answers. (the Question)As I have learned, we can create an "Object" on Heap by creating apointer variableto point to the memory that we have allocated on Heap (usingnewkeyword).For a specific e...
Is there a Variable allocated on the Heap?
0 Yes this is how it supposed to work. I wrote a short blog post on a similar case which explains why it works that way. You can read it here https://adventuresinscm.wordpress.com/2019/08/25/merge-from-master-overrides-a-revert-commit/ In short, Git performs a 3-way merge b...
I have two branches 'master' & 'child'. I added a commit on the child branch and merged it onto the master branch. I reverted this commit on the master branch(and also on the child branch separately). I added the same changes from the first commit on the child branch(kind of like revert of revert, but added as a n...
Adding the reverted changes on master branch via different commit from child branch doesn't reflect on master
It can never be the case that: You cannot create a fork from which to submit a PR; and You do not already have a suitable fork. Your issue is that you don't think your existing fork is suitable, but it is. Whatever the work is that you have locally which you want to share, you can just push that up as a (potentially...
Assume A is a rather snazzy and widely used piece of software with a repo on GitHub and multiple forks. Call them B, C, D. I wish to submit changes to any or all of them (A,B,C,D). GitHub does not allow a 'fork of a fork', so if I fork any one of them I can't easily submit PRs to the others. Obviously I can clone them...
How to contribute by PR to multiple forks from a single base
There is an issue with the version 3.4.1 of react-scripts,So i added a docker-compose file and i specified this line who solve the problem and save my day :stdin_open: trueSo my docker-compose.yml file looks like this :version : '3' services: web: build: context: . dockerfile: Docke...
I'm new to Docker and I tried to run a container of thecreate-react-appimage so these are the steps that I have done:npx create-react-app frontendI created aDockerfile.devlike below:FROM node:alpine WORKDIR '/app' COPY package.json . RUN npm install COPY . . CMD ["npm" , "run" , "start"]I used this command to build th...
I can't run a docker container of my reactjs app
I don't think the SourceArn field gets populated by Lambda. I know SourceArn works for SNS, but Lambda is really running arbitrary code, not an AWS feature like SNS. As an alternative, you can attach a policy to the IAM Role your Lambda function runs as. { "Version": "2012-10-17", "Statement": [ { ...
I'm using the AWS SQS service, and I'm having a hard time defining permissions on my SQS queue. In my setup I'm using the AWS Lambda service, which is triggered when an object is pushed onto an S3 bucket. However to keep my question briefly, this is what I want to achieve: Object is pushed to a S3 bucket S3 bucket ...
AWS SQS permissions for AWS Lambda
I'm not aware of any easy way to change the timezone. Some older versions ofcronsupported an environment variable calledCRON_TZ, but the current version ofcrondoesn't seem to.So you'd have to convert the times to system time.I think you have calculated the times correctly, and the days correctly, but you have to take ...
I have two Cron jobs in my web server -*/5 0-3 * * 0-4 [my job] */5 19-23 * * 0-4 [my job]We are in the GMT+6 timezone, and the server is in GMT-8 timezone. We would like the job to run -Sunday to Friday 9 AM to 5 PM at GMT+6 timezone (office hours) Every 5 minute intervalBut it appears t...
Cron job scheduling
1 The git_url works so long as it is a .git link, not only if it is a github link. For git_rev on bitbucket, it is the branch that will be pulled from. Share Improve this answer Follow answered A...
I'm making a conda package for a toolkit that was published on bitbucket. Looking through the documentation for the meta.yml file, specifically the source section, the possibilities are git_url and hg_url. The url given is .git, but not from github. Does this make any difference? Also, where do I then find the git_rev...
Conda build: github and bitbucket
AFAIK you cannot assign 'note card' in Github project card. as already noted, workaround could be to mention @username in the body of the card or convert the note card to github issue card selecting '... -> convert to issue'. The issue needs to be assigned to a git repository and then it can be assigned to a user ...
I need to assign another user(or multiple users) to GitHub project card - can i do that directly, not by @username in the body of card? I've seen such feature on Trello, but not sure if it is possible on GitHub. Thanks in advance!
Assign GitHub project card to the specific user
You need to open an IPython console, then run this commandIn [1]: %cd /path/to/nilearn/parentBy this I mean that you to go with the%cdmagic to the parent directory wherenilearnis placed. After that you can runIn [2]: import nilearnand that should import your local copy ofnilearn.Note: If you are planning to do changes ...
I would like to create a copy of a package on github that I can edit and use in spyder. I currently use the anaconda package manager for my python packages.Here are the steps that I have taken so far:fork repoclone repo onto my local directoryThe package is called 'Nilearn'. I currently use anaconda and have installed...
Cloning a package from github and use in spyder
First statement will be converted by compiler into String s = "ABCDEF"; so there will be no concatination Second statement will be converted by compiler into this code (or something like this) String s = "ABC"; StringBuilder sb = new StringBuilder(s); sb.append("DEF"); s = sb.toString();
How many memory locations will it take to have a string concatenation? String myStringVariable = "Hello"; In following two statements : String s = "ABC" + "Hello" + "DEF"; and String s = "ABC"; s = s + "Hello"; s = s + "DEF"; and String s = "ABC" + myStringVariable + "DEF"; Which will consume more memory? In wh...
How many memory locations will it take to have a string concatenation?
Docker does not have a mechanism for changing the published ports of a container once it has started. When you publish a port, two things happen:Docker createsiptablesrules in thenattable that redirect traffic to the "public" port to the container.Docker starts a proxy service listening on that port to handle locally ...
I'd like to change the published ports on a live container for exampledocker run -p 80:80 --name nginx_live nginxAnd then later on, change that to another port, example-p 8080:80
Docker change published ports on live container
Here's a solution in python I came up with after following this post. It uses thefreezegunlibrary to manipulate the time to make the signature the same over a given period.import time import datetime import boto3 from freezegun import freezetime S3_CLIENT = boto3.client("s3") SEVEN_DAYS_IN_SECONDS = 604800 MAX_EXPI...
I want to store user profile pictures in an S3 bucket, but keep these images private. In order to do this, I am creating a presigned url whenever the image is required. However, this creates a unique url each time which means the image will never be cached by the browser and I'll end up paying a lot more in GET request...
AWS S3 Presigned Request Cache
Link : Manage data in containers The basic run command you want is ... docker run -dt --name containerName -v /path/on/host:/path/in/container The problem is that mounting the volume will, (for your purposes), overwrite the volume in the container the best way to overcome this is to create the files (inside the conta...
I'm running a docker container with a volume /var/my_folder. The data there is persistent: When I close the container it is still there. But also want to have the data available on my host, because I want to work on code with an IDE, which is not installed in my container. So how can I have a folder /var/my_folder on ...
Sharing files between container and host
Update SOLVED:You have to do this:Create ServiceAccount per userapiVersion: v1 kind: ServiceAccount metadata: name: NAME-user namespace: kubernetes-dashboardAdapt the RoleBinding adding this SAkind: RoleBinding apiVersion: rbac.authorization.k8s.io/v1 metadata: name: PUT YOUR CR HERE namespace: PUT YOUR NS HERE...
I have Kubernetes with ClusterRoles defined for my users and permissions by (RoleBindings) namespaces. I want these users could be accessed into the Kubernetes Dashboard with custom perms. However, when they try to log in when using kubeconfig option that's got this message:"Internal error (500): Not enough data to cre...
Kubernetes Dashboard - Internal error (500): Not enough data to create auth info structure
You can create theSSLContextinstance using this code snippet.// Load Certificate CertificateFactory certificateFactory = CertificateFactory.getInstance("X.509"); Certificate certificate = certificateFactory.generateCertificate(new FileInputStream(new File("CERTIFICATE_LOCATION"))); // Create TrustStore KeyStore trustS...
I need to perform a rest call by attaching the local ssl certificate.I do not have any info about KeyStore. I just know there is a Certificate installed in my PC and I have to use the certificate based on details of certificate like "Serial number", "Issuer" etc which i can see in the certificate details in the persona...
Fetch SSL certificate from local store using java into sslContext object
What it means is that the app label is parametrized. PROJECT_NAME is a variable used in the label. You need to check deployment process/pipeline on how the YAML file is getting created and deployed in kubernetes environment.
I keep seeing labels like below in the kubernetes yaml files that I'm trying to understand. there are certain key value pairs defined but they are defined in formatlabels: app: $PROJECT_NAME-essome other filelabels: app: $PROJECT_NAME-some-servicethings like this are spread all over the yaml file. each yaml fil...
Can anyone explain this in kubernetes yaml file?
I linked to it in your other thread -these are the supported event sources. Notice thatcloudwatch eventsare one of the possible event types. You could set up a Lambda to, for example, run every minute and poll an SQS queue. You cannot directly trigger a Lambda off of an SQS queue.
Does AWS lambda provide support for listening to SQS queue? I found some examples which says one can do that but I am not sure if AWS lambda explicity provide support for that. When I create the lambda function, then I found one blueprint for SQS. So,
AWS Lambda integration with SQS
You can use this: $ tar Ccf $(dirname SRC_PATH) - $(basename SRC_PATH) | docker exec -i foo tar Cxf DEST_PATH - Using - as the SRC_PATH streams the contents of STDIN as a tar archive. The command extracts the content of the tar to the DEST_PATH in container’s filesystem. In this case, DEST_PATH must specify a directo...
I want to transfer multiple files from ubuntu to a docker container. For single file, the below command works: docker cp file_name CONTAINER:path/ But I am not able to upload multiple files at once. I have tried following commands, but no success yet: docker cp {file1,file2} CONTAINER:path/ docker cp [file1,file2] CO...
Copy multiple files in docker with cp
40 Here is another solution using Django Middleware. Django's django.middleware.common.CommonMiddleware calls request.get_host(), which validates the request with ALLOWED_HOSTS. If you simply want to check that the application is running, you can create a middleware like ...
I have a django application deployed on Elastic Beanstalk. The HealthCheck for my app keeps failing because the IP of the ELB HealthCheck is not included in my ALLOWED_HOSTS settings variable. How can I modify ALLOWED_HOSTS to make the HealthCheck pass? I would just pass in the explicit IP address, but I believe that...
Django ALLOWED_HOSTS with ELB HealthCheck
The feature you've described is called Static Pods. Here is a part ofdocumentationthat describes their behaviour.Static pods are managed directly by kubelet daemon on a specific node, without the API server observing it. It does not have an associated replication controller, and kubelet daemon itself watches it and...
I have created a k8s cluster using kubeadm and have a couple of questions about the kube-controller-manager and kuber-apiserver components.When created using kubeadm, those components are started as pods, not systemd daemons. If I kill any of those pods, they are restarted, but who is restarting them? I haven't seen an...
kube-controller-manager & kube-apiserver questions for a kubeadm created cluster
It turns out thatthe proposed way by the AWS documentation to integrate SSO users into the clustersis not compatible with the latest version of EKS.The placeholder{{SessionName}}cannot be evaluated. So I've had to change it like this:mapRoles: | - groups: - system:bootstrappers - system:nodes role...
We're running several clusters with AWS's EKS.Currently all the clusters are already on 1.19 but the NodeGroups are still running on 1.18. The last update of the NodeGroups was in December and there everything works well. The aws-auth ConfigMap wasn't modified since this moment of time.Now we want to update them. If we...
Cannot update EKS NodeGroup because of aws-auth ConfigMap issues
Yey for linear algebra! Column scaling is right multiplication of diagonal matrix: X = X*diag(sparse(fac));
I have a sample X which is a sparse matrix (~5%) and now try to scale each column with a factor (basically tf-idf normalization). Which I thought is a task easy to accomplish somehow now occurs to be not really supported. Here is what I used: fac = log(size(X,1)./max(1,sum(X ~= 0))); X = bsxfun(@times,X,fac); % this l...
Efficiently scaling column of a sparse matrix in matlab
if you are running on GKEfor now day, the kubernetes console has advanced and you can see all statistics on the console uigo tohttps://console.cloud.google.com/kubernetes/listunderworkloadschoose the deployment
I defined my deployment resourcesresources: limits: cpu: 900m memory: 2500Minow onhttp://localhost:8001/apihow can I get themaxusage of memory and cpu (in order to handle and define usages and resources well)?
how to get max usage of mem and cpu usage of pod on kubernetes
I resolved modifiying the suggestion of Daniel W I moved the location / block under location ~ \.php$ But instead of remove the initial / I prepended also $uri location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; fastcgi_pass unix:/run/php/php7.4-fpm.sock; fastcgi_index in...
I have a legacy php websiste that has been migrated to a new server with nginx (and php 7.4) My nginx has near only this server { listen 80; server_name domain.tld; root /var/www/domain.tld; error_log /var/log/nginx/domain.tld-error.log warn; access_log /var/log/nginx/domain.tldt-access.log ...
nginx: how to serve /index.php of subfolders?
They're both wrong. dealloc returns void, not id: - (void) dealloc { [rect release]; [super dealloc]; }
So i'm overriding dealloc method because the object is a composite object made up of one other object. I originally had this dealloc method: -(id) dealloc; // Override to release the Rectangle object’s memory { [rect release]; [super dealloc]; return self; } After looking in the book I saw another answer...
any difference between these two dealloc methods?
From the AWS Documentation : The awsvpc network mode does not provide task ENIs with public IP addresses for tasks that use the EC2 launch type. To access the internet, tasks that use the EC2 launch type must be launched in a private subnet that is configured to use a NAT gateway. For more information, see NAT Gatewa...
I created ECS service in AWS ECS cluster. My container uses the awsvpc network mode. The service has no load balancer. But it cannot touch any public internet resources. When I go to the ECS instance and ssh into docker container I cannot wget any public resources. root@ip-10-3-1-23:/app# traceroute 8.8.8.8 tracerou...
aws-vpc containers do not have access to the internet
Azure Diagnostics is a good point to start. You can add any performance counter you like: Memory Consumption, CPU usage, etc. Rightclick your Role in the Azure Project and go to Properties. Under Configuration you will see "Diagnostics" Tick "Enable Diagnostics" and select "Custom Plan" -> "Edit" Under Performance C...
I want to log the memory usage of some web/worker roles which are hosted in windows azure instances. Is there a way to programmatically check for the currently used memory so I can save this value somewhere for later use? The only information I found was to get the current thread and then call a method which should re...
How to measure memory usage of azure web/worker roles programmatically?
I came across this bit of documentation which shows that multiple JSON objects can be stored within a single JSON file that is fed by COPY command. According documentation there can be more than one JSON object within the file and only white characters should be placed between JSON objects. So I dropped objects like t...
I have created a Kinesis data stream and I'm feeding records into it that have format like this: { "ufo": 5, "unicorn": 7, "something else": 9, } I'm trying to write a transformation lambda that will transform above record and insert three records into Redshift. Redshift table schema is like this: CREATE ...
Firehose data transformation lambda - produce multiple records from single kinesis record
Finally I have solved how to run the gem Whenever. It's working good on production, but not in development mode (I think that to working good in dev mode you must do some tricks). Then, these are the processes to do: install the gem write your scheduler.rb file push to the remote server login to the remote server (fo...
I have an application that contains a bunch of tasks, and every day I want to run a cron job that creates a DayTask for each Task in the database. A Task has_many DayTasks and these daytasks are what users will be checking off every day. I'm using the whenever gem but it doesn't seem to be running at all. Any ideas? c...
Cron job not working in Whenever gem
.htaccess is just a means of specifying Apache configuration directives on a per-directory basis. They allow numerous different kinds of password protection.If you are talking about HTTP Basic Authentication then the username and password are sent in cleartext with every request and are subject to sniffing (assuming yo...
Are there any known flaws with htaccess protected pages?I know they are acceptable to brute force attacks as there is no limit to the amount of times someone can attempt to login. And if a user can uploaded and execute a file on the server, all bets are off...Are there any other .htaccess flaws?
How secure are .htaccess protected pages
As metionedhere, I fixed this issue by addingdata-noprefixto the link tag for my stylesheet.<link href="https://fonts.googleapis.com/yourFont" rel="stylesheet" data-noprefix>
I want to fix an error, it's this:Font from origin 'http://localhost:8000' has been blocked from loading by Cross-Origin Resource Sharing policy: No 'Access-Control-Allow-Origin' header is present on the requested resource. Origin 'http://localhost:3000' is therefore not allowed access.All my assets are located enlocal...
Blocked web-fonts from localhost in different ports
I think the problem insrcpath of your JS file in script tag.Instead of, use:
Router configuration:const routes = [ { path: '/', component: Layout, onEnter: sessionFilter, indexRoute: { component: ActivityIndex, onEnter: sessionFilter }, childRoutes: [ { path: 'login', ...
Unexpected token < with React Router & Nginx
say suppose you have xyz.bak as a bakup and you want to restore it on to your sqlserver 2008r2 try running following query in ssms Restore database DatabaseName from disk='path of your bak file' this will restore your database.
I'm using MSI Project with InstallShield 2010. I have a .bak file, a backup of a SQL database (I am using SQL Server 2008 R2). How can I restore it using InstallShield? I was looking in SQL Scripts tab and I didn't find anything about restoring backups. Thanks for your time!
How to restore a database from a backup in installshield
If you don't need access from the host to these volume files, use a named volume instead of a host mount. Docker will initialize the contents of the named volume, including the permissions and ownership, avoiding permission issues like this: version: '2' volumes: prometheus: grafana: services: prometheus: ...
I'm trying to execute my docker-compose.yml file which contains prometheus and grafana configurations. Here is my docker-compose.yml file: version: '2' services: prometheus: image: prom/prometheus ports: - 9090:9090 volumes: - /prometheus:/prometheus command: - --config.file=/etc/p...
docker-compose opening storage failed: permission denied error on running
You have given wrong image source link in your HTML code. Perhaps you have missed the /assets/ path in your link. It should be <section id="music"class="landing"> <img src="../assets/images/tom-odell-album-coverart.jpg" class="album-cover"> </section> but you have given : <section id="music"class="landing"> ...
I'm making a website that will be hosted on vercel with the files on github and my image isn't visible: My file structure on my github is: Tom-Odell-V2 assets images And this is what I have in my html: <img src="https://tom-odell-v2.vercel.app/assets/images/tom-odell-album-coverart.jpg" class="album-cover"> ...
Hosting site with Github and Vercel, images won't show up
Try the following command from your WSL command line :echo "export DOCKER_HOST=tcp://localhost:2375" >> ~/.bashrc && source ~/.bashrShareFollowansweredOct 16, 2019 at 13:13Rafik SaadRafik Saad61611 gold badge55 silver badges99 bronze badgesAdd a comment|
I am using Ubuntu Bash is windows 10. I have been struggling with connecting to the Docker Daemon. There has been couple of queries in the past which I followed with not luck.From the search I could see that the TCP endpoint need to be open for connection to the Docker Daemon. Followed the following excellent posts wit...
Cannot Connect to Docker Daemon - Windows 10 WSL
You probably want a?instead of a&, and you want to use theQSAflag:RewriteRule ^(.*)/admin/(.*) admin/$2?app=$1 [L,QSA]ShareFollowansweredApr 9, 2014 at 14:56Jon LinJon Lin143k2929 gold badges221221 silver badges220220 bronze badges1@Safy "QSA" flag means "Query String Append", or append any part of an existing query st...
i have problem in htaccesshere's my project [custom cms] scheme./ admin app p1 p2 core views .htaccessin this htaccess file :i want to make app has its own styleso i add thisRewriteRule ^(.*)/admin/(.*) admin/$2&app=$1 [L]now when i add this urlhttp://www.site.com/p1/admin...
htaccess redirect in main folder and sub folder
You shouldn't be building your application on the production server. Build it locally, then deploy it to production and this won't be an issue.
Is it possible to build react app for production usingnpm run buildwithout it making the site offline? I am using nginx to serve the react app and when I run the above command, I getInternal servererror until the build is finished. It would be nice if the old version of the app is served until the build is finished.
How to build react app for production without making the website offline?
If you have a.git-credentialsfile in your$HOMEyou should completly delete it (it will be rebuilt by asking you your credentials once for each user/website).Then you should have a look at your$HOME/.gitconfigand double-check if you have something like:[user] | name = new_user_name | email =[email protected]If you di...
I took over a Mac OSX computer that previously had another person's githubuser.emailanduser.namecached. In order to access my own github repos, I changed the cached github user.email and user.name to mine. However, I still get denied to access (push) my repos on github. So I was wondering what I should do in order to p...
github: access denied after changing user.name globally
The custom block itself, that is, the bytes obtained from caml_alloc_custom, is part of the Caml heap and can be moved like any other object.¹ It's very common for the custom block to contain pointers to data structures that are also accessed by C² code and live outside the Caml heap; Caml treats the contents of the c...
Imagine I've a C library called libcat for interacting with my cat fluffy. I'm therefore writing bindings for OCaml to simplify interactions with fluffy. module type CAT = sig type cat val find : ... -> cat val feed : cat -> unit ... end ;; module Cat : CAT = ... There is considerable memory management ...
Are custom blocks ever copied by OCaml?
You could copy the entire CUDA SDK to your home and tell Theano and others that they should use your local copy of CUDA by adding/modifying these environment variables in your~/.bashrcexport CUDA_ROOT=~/program/cuda-7.5 export CUDA_HOME=~/program/cuda-7.5 export PATH=${CUDA_HOME}/bin:$PATH export LD_LIBRARY_PATH=/usr/l...
Can I installcuDNNlocally without root access ?I don't have root access to a linux machine I am using (the distro is openSuse), but I have CUDA 7.5 already installed.I am using Theano and I need cuDNN to improve the speed of the operations on the GPU.I downloadedcudnn-7.5-linux-x64-v5.1from Nvidia and as per the instr...
Installing cuDNN for Theano without root access
You can mirror your GitHub repositories in Cloud Source Repositories. With the following steps, commits that you push to the GitHub repository are mirrored into your Cloud Source Repository.You can follow the steps that I write you down, but first,don't forget to enable Cloud Source Repositories API:Go toOpen Cloud Sou...
So I have a local repository that I've pushed to Github as a new branch, without any trouble, but the next step is to clone that same repository to Google Cloud Source Repository, also as a new branch. This is terribly confusing because I'm unfamiliar with Cloud Source Repository and the documentation doesn't really go...
How to Clone a Branch from Github to a New Branch in Google Cloud Source Repository?
Correct, if you pass something "large" by value that item is copied onto the stack. Passing a pointer to the data avoids the copy. It is doubtful that the performance difference will be meaningful in most real-world applications, unless "large" is actually "huge" (which in turn may overflow the stack).
I am learning C and get confused about something I read online. At http://www.cs.bu.edu/teaching/c/stack/array/ I could read: Let's look at the functions that determine emptiness and fullness. Now, it's not necessary to pass a stack by reference to these functions, since they do not change the stack. So, we could...
C programming: the cost of passing by value
Try this:apiVersion: v1 kind: ConfigMap metadata: name: nginx-configmap data: {{ $root := . }} {{ range $path, $bytes := .Files.Glob "config/*" }} {{ base $path }}: '{{ $root.Files.Get $path | nindent 2 | upper | replace "." "_" }}' {{ end }}
I have defined application.properties files in a config dir. Config dir is on the below structure.config application.properties application-test.properties application-dev.properties application-prod.propertiesI have created a helm chart to create the configmap. Helm chart is defined as belowapiVersion: v1 kind...
Helm - Configmap - Replace only the contents of file and not File Names
Somecronimplementations may offer you a*/5syntax for the day, or even an e.g.16-31/5syntax meaning "every 5 days starting from the 16th".The first problem is that you will need a second*/5entry for the next months. The second problem with such a syntax, is that your script will trigger on the 1st, 6th, 11th, 16th, 21th...
I need to set up a cronjob, which will run every 5 days, but I need it to start at a specific date.For example, it will execute a curl every 5 days, but i want the first execution to be next week.Is this possible or should I wait for the day to come and set it up then?
Set up a cronjob to run every N days with a specific date as a start
2 From time to time similar questions come up. Take a look at these: Magento & Subversion (SVN) - Getting a Development Environment Started? Whats a good way to approach a development environment when working with Magento? Magento staging and production I'm not voting thi...
Closed. This question is opinion-based. It is not currently accepting answers. Want to improve this question? Update the question so it can be answered with facts and citations by editing this post. Closed 9 years ago. ...
Looking for the best practice with SVN / GIT for Magento projects [closed]
10 An alternative solution could be by running php artisan schedule:run with supervisor. In my case I have a schedule-run.conf in [project root]/.docker/php/workers: [program:schedule-run] process_name=%(program_name)s_%(process_num)02d command=/bin/bash -c "while [ true ];...
I have a container with for my laravel app with php:7.0.4-fpm as base image. This is my dockerfile : FROM php:7.0.4-fpm RUN apt-get update && apt-get install -y cron nano libmcrypt-dev \ mysql-client libmagickwand-dev --no-install-recommends \ && pecl install imagick \ && docker-php-ext-enable imagick \ ...
cron on a docker container for laravel not working
I think the problem is that there seem to be no link between Sonar and Jenkins. It would be easier to make it with plugins.After installing plugins you'd just have to add a build step in the jenkins administration.In order to see the coverage report in Sonar you should use the "Jenkins Sonar Plugin". However it will fo...
I'm trying to get test unit coverage with Sonar. To do so, I have followed these steps :Generating report withpython manage.py jenkins --coverage-html-report=report_coverageSetting properties in/sonar/sonar-3.5.1/conf/sonar.properties:sonar.dynamicAnalysis=reuseReportssonar.cobertura.reportPath=/var/lib/jenkins/workspa...
How to get tests coverage using Django, Jenkins and Sonar?
Ok so i finally managed to do it, the correct way is to set the query like this:sum(rate(pensieve_grpc_server_handling_seconds_bucket{namespace="$namespace", pod=~"pensieve.*", account=~"$account_id", grpc_method=~"$grpc_method"}[$__rate_interval])) by (le)and select following options:and
I cannot make grafana to display Latency heatmap, whatever i do it does not function correctly for me.I have a prometheus histogram defined like this:serverHandledHistogram: prometheus.NewHistogramVec( prometheus.HistogramOpts{ Namespace: metricsconst.PrometheusNamespace, Nam...
Grafana Heatmap display for prometheus histograms
There is no documented way that I am aware of for accomplishing this in user mode.That said, it is possible to determine this in kernel mode, but this would involve inspecting the Page Table Entries, which belong to the Memory Manager - not something that you really wouldn't want to do in any sort of production code.Wh...
Is there a way, in Windows, to check if a page in in memory or in disk(swap space)? The reason I want know this is to avoid causing page fault if the page is in disk, by not accessing that page.
windows memory managment: check if a page is in memory
The caching factory is designed for short-lived sessions such as operations performed with a JmsTemplate. It's generally not needed with the listener container (because its sessions are generally long-lived), unless you perform JmsTemplate operations on the container's thread - to participate in the container's transa...
When using the spring-jms, there are 2 options given by spring for connection and session caching for performance gain. Use the CachingConnectionFactory and cache the sessions, optionally you can cache producer and consumers as well. Using the DefaultMessageListenerContainer you can set the cacheLevel to [1:connectio...
Which is better to use CachingConnectionFactory or caching in DefaultMessageListenerContainer?
You can use the ngx_http_auth_digest module. That module is not standard yet so you'd have to install it.
Is it possible to use http 1.1 digest access authentication instead of basic access authentication when restricting access to a website with nginx? I mean this in a similar fashion as on the nginx guide.
Digest access authentication on nginx
Update 2023-11-29: DDEV uses the ddev-phpmyadmin add-on, and has had https support for PHPMyAdmin for years now,ddev describewill show you the URL. As explained below by @HEYDANNY, you install PhpMyAdmin withddev get ddev-phpmyadmin && ddev restartand can launch it withddev phpmyadmin, and it works fine with https.
I am using DDEv and Docker with Windows 10 pro to set up a localhost install of drupal 8.8 using Composer. I have set up and configured the local drupal installation (it is a fresh install) and it appears to be running correctly, but in the admin section of the drupal site I receive a warning to change write permission...
How to access phpmyadmin on DDEV Windows 10 pro localhost with SSL record too long error
I also looking for the solution.I knownEXPOSEjust some kind of docs for image. but I think it's really useful.directly NOseems the answer isNOfor now.SO I have to create 2 (or more)dockerfilefor that purpose.workaroundmaybe build from STDIN (generate dockerfile by other program) can be a workaround.But it's really not ...
I'm trying to expose debug port inside Dockerfile based on a build argument(DEBUG_ENABLED). This image goes into a k8s pod eventually on a cluster with multiple pods running for various apps. I'm trying to EXPOSE ports in all the images based on a build-arg.Here is my Dockerfile:ARG DEBUG_ENABLED=true : : EXPOSE 8080 R...
Conditionally expose port in Dockerfile
These are the same metrics. Why would you use join? You can use aggregationsumlike you did and filter using regex.Something like this (I didn't test the regex)sum(rate(rpc_request_duration_seconds_bucket{instance=~"10\.43\.1\.24[45]"}[1m]))
Technically, Prometheus doesn’t have the concept of joining series like SQL, for example, has. However, series can be combined in Prometheus by using an operator on themSeries Asum(rate(rpc_request_duration_seconds_bucket{instance=~"10.43.1.244"}[1m]))Series Bsum(rate(rpc_request_duration_seconds_bucket{instance=~"10.4...
How to Join two series in Prometheous using operator +
You can also just do this <!--- Comments are Fun ---> Remember, markdown is just an easier way to write HTML content. (note the triple dash) some item Something with an inline comment
I'm trying to create a gist with an inline comment. I read a solution here which works for putting a comment on it's own line (as shown in the image); however, it doesn't appear to work for inline comments. Is there any way to get inline comments in github flavored markdown? - [x] some item - [ ] another item with me...
How do I add inline comments in Github Flavored Markdown?
It helped me when I moved my project to somewhere else, like the desktop, I guess the path is too long to be able to create this directory.ShareFollowansweredSep 8, 2022 at 21:23Ahmed MousaAhmed Mousa111 silver badge11 bronze badgeAdd a comment|
I am following Building a TFX Pipeline Locally (https://www.tensorflow.org/tfx/guide/build_local_pipeline) on ubuntu 21.04. I am only running the CsvExampleGen component and I am getting the following error:ERROR:absl:Failed to make stateful working dir: ./my_pipeline_output/CsvExampleGen/.system/stateful_working_dir/2...
local TFX pipeline run create ERROR Failed to make stateful working dir ; Protocol error
Use full url. Eithercurrent_url()orsite_url('register')in your form's action.It's good practice you use either of these or preffix withbase_url()since that allows your site to run without having to be the domain path.For instance you can, if needed run the site on example.com/my_site/ instead of only being able to run ...
I have a Controller calledRegisterthat looks like this:class Register extends CI_Controller { public function index() { $this->load->view('register'); } public function create() { $this->load->library('form_validation'); //set rules here if ($this->form_validation->...
Controller repeating in URL in CodeIgniter
If you use Java 1.7, you can usetry-with-resouceblock. The object will be closed automatically when leaving the block.GetObjectRequest req = new GetObjectRequest(bucketName, fileName); try(S3Object object = s3Client.getObject(req)) { ... } catch(AmazonServiceException e) { if(e.getErrorCode().equals("NoSuchKey"...
I'm using AWS java SDK to upload file on AWS Management Console's Bucket. However, if there is no such file online at first time when I try to get access to it, my code will catch the exception (NoSuchKey). Then I want to close the connection. The problem is I don't have any reference to close that connection because o...
How to close AWS connection if there is no such key in the query
0 I had the exact same issue. For me it was a compatibility issue with Smile-SA /elasticsuite. The workaround was to delete 'vendor/smile' folder, and do composer dump-autoload. (I know it is not a permanent fix, that what I have for now) Share Improve this answer...
When trying to search for a product in magento 2.4 returns the following error Exception #0 (Elasticsearch\Common\Exceptions\NoNodesAvailableException): No alive nodes found in your cluster Below configuration: PHP 7.4.1 Nginx 1.14 MySql 8 ElasticSearch 7.9 /etc/nginx/conf.d/100-magento2.conf upstream fastcgi_back...
Config ElasticSearch Magento 2.4 - No alive nodes found in your cluster
There is aFork pull request workflowssetting that needs to be turned on to allow running workflows on pull requests to a private repository:https://docs.github.com/en/repositories/managing-your-repositorys-settings-and-features/enabling-features-for-your-repository/managing-github-actions-settings-for-a-repository#enab...
I'm managing a small private repository with a tightly controlled list of contributors who can make pull requests against the repository. I'd like to be able to run some simple automatic validation of pull requests (syntax checks, formatting, etc).Unfortunately, it seems like for security reasons that GitHub doesn't pe...
Running workflows on pull requests to a private repository?