url
stringlengths 13
4.35k
| tag
stringclasses 1
value | text
stringlengths 109
628k
| file_path
stringlengths 109
155
| dump
stringclasses 96
values | file_size_in_byte
int64 112
630k
| line_count
int64 1
3.76k
|
|---|---|---|---|---|---|---|
http://www.tweetegy.com/2011/04/installing-rails-gem-fails-with-undefined-method-spec-for-nilnilclass-error/
|
code
|
Installing Rails gem fails with “undefined method `spec’ for nil:NilClass” error15 Apr 2011
Today, while installing Rails 3, my computer suffered a power outage. This does not happen very often, in fact, this was actually due to the computer overheating! After pouring cold water over the keyboard, I restarted the computer and continuing the Rails 3 installation the process, but to my shock and horror, it failed with this message:
ERROR: While executing gem ... (NoMethodError) undefined method `spec' for nil:NilClass
I’ll be honest with you, I was kind of expecting problems to happen since the computer did abruptly shut off during an installation process – but now what do I do!? Well, after a little Googling, the problem was solved. I found out that it’s due to the gem install cache folder which (due to the power interruption) now contained corrupted / empty gem files. My solution was to clear this cache and run the install again, as follows:
- Find out the location of your cache using:
- List all the gems in that cache, using, for example:
ls -l /usr/lib/ruby/gems/1.8/cache/
- Compare with successfully installed gems using
- Remove the gems from the cache list that are missing from the gem list using the
- Continue the installation, for example.
sudo gem install -v=3.0.6 rails --no-rdoc --no-ri
So now if you have a power outage or any other external failure during a gem installation process you don’t need to panic. Just clear the cache and continue where you left off before you were rudely interrupted!
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917119637.34/warc/CC-MAIN-20170423031159-00102-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 1,541
| 12
|
https://www.metaphacts.com/get-started
|
code
|
Get started with metaphactory for free and accelerate your Knowledge Graph journey!
Your trial options
Run anywhere with with any graph database
- Deploy your solution wherever you want
- Can be used with any graph database
- Optional: open-source graph database included
Run in the cloud with any graph database
- Deploy your solution on AWS Cloud
- Can be used with any graph database, incl. Amazon Neptune
- Please note: for AWS-based systems, AWS infrastructure charges do apply
“Getting Started” tutorial included
What happens after my trial?
Try our online demo
Don't want to run your own system? Try our hosted demo system to experience metaphactory first-hand and without any installation.
metaphactory public demo system
Experience metaphactory with the free and open Wikidata dataset! The metaphactory demo instance runs the latest metaphactory version. It showcases key features for search and visualization, as well as federation with additional datasets (such as sensor or weather data), machine learning algorithms and similarity search functionalities (e.g., on chemical structure similarity).
Get started with these examples:
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320304859.70/warc/CC-MAIN-20220125160159-20220125190159-00083.warc.gz
|
CC-MAIN-2022-05
| 1,145
| 17
|
https://www.virtualinnovations.eu/blog-detail/what-is-distributed-ledger-technology-or-blockchain
|
code
|
Jerônimo do Valle
Distributed Ledger Technology (DLT), commonly known as "Blockchain Technology", refers to the technology behind decentralized databases that, through a peer-to-peer network, provide control over the evolution of data between entities using consensus algorithms that guarantee replication between network nodes.
For simplicity, imagine a “book” owned by all users, where anything written would remain forever and, simultaneously, all “entries” could be compared with other users' books to verify and validate what was written; this is the essence of DLT.
The units where information is recorded, that is, the “pages” of this book, are the "blocks". Each block contains "hash" information (a "hash" or "hash value" is a string of alphanumeric values obtained through an encryption algorithm. If the same information is entered, it will always deliver the same "hash" in the output , however, if there is the slightest change in the input, the output hash will be totally different - the so-called "avalanche effect").
In the blockchain, the written blocks cannot be changed and it is only possible to write after the consensus of the majority of the network. This means that in order for something to be modified, all blocks created later would have to be changed and 51% of the network would have to agree to this. As blocks are created at each moment, changing them - and also the blocks that precede them - this is practically impossible. Furthermore, changing a block will result in the creation of a new “information branch” that will be considered the “source of truth”, but the previously existing data remains in a different branch (routine called “forking”).
As a measurable, immutable, and collaboratively managed record, blockchain technology comes at the right time to solve the problems inherent in digital asset transactions and pave the way for Internet security.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100531.77/warc/CC-MAIN-20231204151108-20231204181108-00395.warc.gz
|
CC-MAIN-2023-50
| 1,919
| 6
|
https://forum.duolingo.com/comment/9945479/Han-kan-heller-ikke-komme
|
code
|
Yes, but that would have a slightly different meaning.
If someone says "I cannot come" = "Jeg kan ikke komme", and you can't come either, you'd say the above sentence.
If you're listing things you cannot do, you'd put it like you said. "I can't call you. I cannot come either" = "Jeg kan ikke ringe deg. Jeg kan ikke komme heller."
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780058450.44/warc/CC-MAIN-20210927120736-20210927150736-00392.warc.gz
|
CC-MAIN-2021-39
| 331
| 3
|
https://www.namesforbaby.co.uk/names/girl/TAMIA
|
code
|
Baby name Tamia
Pronunciation:tah me ah
Description: Tamia is a relatively modern coinage and as such does not have a definite meaning or origin. It is likely to be a compound of other popular feminine names such as Tammy and Mia. It seems to have developed in America and is especially popular with African American parents. It is a pretty and unusual name.
Statistics for Tamia
From 1996 to 2015 the name 'Tamia' was given to 383 Girls, representing 0,01% of the total 5.943.365 Girls registered since 1996.
Registrations by Year of name 'Tamia'
This chart shows the registrations of name 'Tamia' in England and Wales per year since 1996
It is possible to see that the registrations name 'Tamia' varied from 5 to 43 per year.
Rank Popularity by Year of name 'Tamia'
This chart shows the popularity of name 'Tamia' in England and Wales per year since 1996
It is possible to see that name 'Tamia' varied from rank 676 to 4.103.
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-18/segments/1555578530505.30/warc/CC-MAIN-20190421080255-20190421102255-00147.warc.gz
|
CC-MAIN-2019-18
| 927
| 11
|
https://cryptozcast.com/trade-clout-on-the-blockchain-com-exchange-by-peter-smith-blockchain-jun-2021/
|
code
|
Today, we’re announcing that the Blockchain.com Exchange will be the exclusive launch partner for $CLOUT, the native token of the BitClout network. Today, customers can deposit $CLOUT to their Blockchain.com Exchange account to prepare for trading.
In about 24 hours, we’ll enable limit-only orders to be placed so customers can participate in price discovery without matching orders. Then 24 hours after limit-only orders are live, we’ll enable live trading. The first order books are CLOUT-USD and CLOUT-BTC.
What is BitClout? BitClout is the first social network built from the ground up as its own custom blockchain. Each creator on the platform has a creator token that can be bought and sold by followers, making it the first crypto social network. In fact, you can follow the Blockchain.com handle here.
Today, a post submitted to a traditional social media platform belongs to a corporation, rather than the creator who posted it. In contrast, BitClout is a fully open-source, decentralized project with no company behind it — it’s just coins and code. This means every creator has a direct relationship with their followers, which unlocks new ways to engage and monetize that aren’t based on ads.
We’re excited to provide exclusive access to buy and sell $CLOUT to Blockchain.com users, not just because it’s a popular crypto project but because it also demonstrates that architects, developers, and builders can create consumer products with crypto protocols.
While there is a healthy trading economy for tokens like Bitcoin, we’re still in the early innings when it comes to use cases that non-crypto folks can truly get excited about. We’re excited to support more assets for our customers, collaborate with BitClout developers and community, and look forward to listing other similar crypto projects in the future.
NOTE: Blockchain.com Ventures is an investor in BitClout.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100602.36/warc/CC-MAIN-20231206162528-20231206192528-00366.warc.gz
|
CC-MAIN-2023-50
| 1,905
| 7
|
http://mathhelpforum.com/algebra/187267-practical-problem.html
|
code
|
Hi, I hope you can help, it's a long time since school and I find my self with a practical problem where I need to isolate a variable in a formula. The formula is
S = ((DH/(H-D+F))-((DH/(H+D-F))
Is it possible to derive a formula or solve for D i.e isolate D given I have the values of all the other variables?
Many thanks for your help, this will help me a great deal
Multiply both sides by (D+ H- F)(D- H+ F) to get rid of the fraction
Note that D- H+ F= D- (H+ F) so that (D+ H- F)(D- H+ F) is a :"product of sum and difference"- so we have
Subtract from both sides to get
That's a quadratic equation and you can solve for D using the quadratic formula:
with , , and .
(If S= 2, you cannot use that formula because you cannot divide by 0, then this is not a quadratic at all but is FHD- 2(H- F)= 0 which has solution
To explain what I'm trying to do: I'm trying to manipulate the formula used in photography to calculate the depth of field (DoF) within a photograph i.e the amount of image in acceptable focus. This is dependent on a few things, your camera's sensor, the focal length and aperature of your lens and the distance you are from the subject. It is a common formula that it used widely but I have only ever seen it to calculate the DoF which I have called S. I was trying to solve it for the distance required D, so I can work out exactly how far away I need to be for a certain amount of the image to be in focus without the trial and error involved using the more traditional formula. So, after a bit of basic formula engineering I got it to the point within my original post but got stuck trying to solve it for S (Sharp focus). F is the focal length of the lens and H is the hyperfocal point - the distance that maximises the DoF. So hopefully you can see this is a practical problem that would help me out greatly.
As a practical example to work backwards to get D, i use
D = 1000
F = 50
H = 44759.2311
Which gives S = 42.468478, this is in millimetres so 4.246 cm should be in focus if my subject is 100cm away using a 50mm lens. But say I know that I want about 4cm to be in focus so what distance do I need to be at? hence solving it for D? Plugging the other known values back into the solved formula I hope to get D = 1000 but it doesn't seem to be there just yet?
Once again many thanks for your help
Did you trip up here HallsofIvy?Note that D- H+ F= D- (H+ F) so that (D+ H- F)(D- H+ F) is a :"product of sum and difference"-
The original was (h-d+f)(h+d-f)
for which I had
From there I have...
This reduces down to
and the best I could get for d was
Huge thank you for this. Just to prove how clueless I am with this, if I substitute the values in and solve for d using + and - I get 2 values as is the nature of the formula I suppose. The aim is to get one formula for d that will give a single answer, is this possible?
Not really since the square root part is positive
if the square root of this is less than
then there are 2 solutions
if it is greater than then you can disregard the -ve root
hope this helps
btw, I know nothing about photography. If any of your other variables can have a negative value, let me know which
None of the original variables will be negative. Using the values I gave as an example, the square is more than f(h+s) so ignoring the minus then I make d=50.03369 which is spookly close to f but I was expecting d to equal 1000?
|
s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698540698.78/warc/CC-MAIN-20161202170900-00288-ip-10-31-129-80.ec2.internal.warc.gz
|
CC-MAIN-2016-50
| 3,385
| 31
|
https://becausewecare1.com/2011/04/26/letter-to-governor-gregoire/
|
code
|
It is in our State Constitution and supported by the US Supreme Court that we must provide for our most vulnerable citizens. It is time that we, as a community, support this and stop trying to take more away from them.
Just because they can not speak up for themselves is no reason to perpetually attack them. What is most disturbing to me is that the attack comes most often and most fervently from the very agencies that are set up to advocate for them.
It is time to take off the blinders and see what they are doing to our citizens. Hopefully, this letter and more actions by grassroots advocates for our most vulnerable citizens will begin this process.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499946.80/warc/CC-MAIN-20230201144459-20230201174459-00090.warc.gz
|
CC-MAIN-2023-06
| 658
| 3
|
http://www.linux-magazine.com/content/view/full/47204
|
code
|
Is Debian Dying?
Off the Beat: Bruce Byfield's Blog
Steven J. Vaughan-Nicholls created a stir this week when he marked the release of Debian 6.0 by wondering if the distribution was still relevant. He was refuted by Joe Brockmeier, and the discussion spilled over on to Facebook, where a number of journalists (including me) speculated freely. But the noticeable lack of hard facts disturbed me, so I decided to see if I could find any indicators of Debian's health on-line.
On Facebook, Vaughan-Nicholls explained that his basic question was, "'Where will the new Debian developers come from to keep it going?' I see 20-something developers working on Android or Ubuntu, Debian, not so much."
That seems a reasonable place to begin, but it needs some qualification. For one thing, I don't think that anyone interested in Android is also going to be interested in Debian. For another programmers do not necessarily choose between Debian and Ubuntu these days.
True, some Ubuntu contributors, like Mathew Garrett, do seem to have become inactive in Debian. Others, like Benjamin Mako Hill, have reduced their involvement. It is also true that relationships between Debian and Ubuntu have sometimes been acrimonious, with some Debian members regarding Ubuntu as prospering unduly at Debian's expense and without contributing its share of development. As a result, at times developers might have felt the need to choose between Debian and Ubuntu.
However, the relationship seems to have improved greatly in the last few years. Debian has set up a Derivatives FrontDesk page to increase cooperation between Debian and distributions like Ubuntu that derive it. Similarly, Ubuntu's Debian for Ubuntu Developers page stresses the importance of working with Debian developers, including the importance of reporting bugs to Debian package maintainers.
Shuttleworth's statement that "Every Debian developer is also an Ubuntu developer" is perhaps exaggerated, but with 89% of Ubuntu packages coming from Debian, it continues to have a great deal of truth. Clearly, developers do not always have to choose between Debian or Ubuntu, and are actively encouraged on both sides not to see the situation as an either-or choice.
So far as I have been able to find, no one keeps track of how much Ubuntu-centered developers contribute to Debian. However, at least for the approximately 2000 packages that make up Ubuntu, it seems safe to guess that the contribution is sometimes significant.
The new Debian release also benefits from new packages from Ubuntu, such as the rebranded version of the Ubuntu Software Center, one of the better graphical package managers currently available. Undoubtedly, there are other examples.
The Size of Debian
But if we disregard the unmeasurable contribution that Ubuntu makes to Debian, is there any sign that the number of Debian developers is declining?
At first, Christian Perrier's statistics of the number of Debian developers and active Debian developers for 2009 and 2010 seems to indicate alarming drops. Perrier lists 1461 developers in 2009, but only 1410 in 2010 -- a drop of nearly 4% in eleven months. The drop in the number of active developers is even greater, falling from 999 to 873, or a loss of 13%.
However, these numbers do not tell the whole story. In particular, the Debian project has been making an effort in the last couple of years to weed out the inactive. Since we do not know how long the inactive have not been participating in the project, the decreases may be as much an indication of the effort to remove the them as proof of a sudden decline.
Moreover, these base numbers do not indicate recent recruitment. Since Perrier's 2010 statistics, 24 new maintainers have joined the project , which reduces the loss to 11%.
Moreover, another 55 are going through the rigorous acceptance program. When these recruits are added to the number of active developers in 2010, then the loss shrinks to just over 5%. Add the 25 applicants who are on hold, and the loss is only 2%.
Of course, any change over a year is not enough to establish a trend. Nor can we assume that all new maintainers in the middle of the process will eventually become Debian developers.
However, we can say that, from the evidence available, no one can prove a pattern of decline. Moreover, we can say that the small decrease for the single year is partly voluntary, in that it is due to the Debian project's insistence on examining the fitness of new recruits closely. With laxer standards, Debian could have easily kept the number of developers at the same level.
At any rate, the number of developers, no matter how you interpret the available figures, is not the whole story. As a by-product of studying the md5sums used in Debian packages, developer Romaine Francoise concluded in August 2010 that the new release represented an increase of 30% in the number of packages in Debian over the previous release in 2009. This figure compares favorably with the 2009 release's increase of 24% in packages over the 2007 release.
Judging from these figures, Debian's growth remains as strong as it always was. Just as importantly, because the number of developers has declined at the same time, those who remain are apparently doing more with less. So far, the decline in developers is small enough that it appears to be irrelevant to the success of the distribution.
Exact figures are hard to come by. However, those that are available suggest that the Debian project is in no immediate danger of being unable to do its self-appointed tasks.
From the available information, you cannot say even say that Debian is considered less relevant to most people. The truth is (no doubt because of its reputation for being an expert's distribution), Debian has never been the most popular distribution, although always remaining solidly in the top ten. If you look at the Page Hit Ranking on Distrowatch, in 2003 -- before Ubuntu even existed -- Debian was in the fifth spot. Now, in 2011, it remains in precisely the same spot.
True, Debian dipped to seventh position in 2007 and 2008, the years when Ubuntu was first gaining recognition. Yet even that drop is slight, and was apparently only temporary.
The indicators are not complete, so they cannot be definitive. Yet they all suggest the same story: Debian is holding its own. Short of a natural disaster at DebConf, you can be reasonably sure that Debian will continue much the same as always for at least the next few years.
Now, if you excuse me, I have an apt-get dist-upgrade to run on another machine.
aptitudeGood points. I just got annoyed as usual with Vaugn whotisname's article and had a go at him in the comments (no, not using rude words or anything, that gets us nowhere).
Oh, and you should be using aptitude not apt-get ! :-P
Link FixedThanks...The link has been updated.
Broken linkThe link to http://raphaelhertzog.com/2...he-debian-ubuntu-relationship/ is broken, you have a leading parenthesis that is problematic.
Re: Debian in DistrowatchThe exact ranking can change. When I looked, it was fifth.
Debian in DistrowatchI'm not sure if we're looking at the same page, but Debian is currently in the fourth spot in Distrowatch, not the fifth.
It's also interesting to highlight the fact that of the 5 most popular distributions according to Distrowatch, 2 of them are direct derivatives (Ubuntu and Mint), which obviously weren't there before 2004, and wouldn't be there today if it weren't for Debian.
So if we take Ubuntu and Mint out of the equation, we could say that Debian is actually #2, closely behind Fedora.
Debian is successful on serversThe W3Techs statistics suggest that Debian continues to be one of the most successful GNU+Linux distros on severs. Ubuntu, for example, is not nearly as popular as Debian.
Both projects help organizations build their own containerized systems.
Mark Shuttleworth has resumed the position of CEO of Canonical.
Microsoft's open source code hosting platform CodePlex will come to an end after a more than 10-year stint.
Comes with Gnome 3.24
The bug was introduced back in 2009 and has been lurking around all this time.
The new release deprecates the sshd_config UsePrivilegeSeparation option.
Lives on as a community project
Five new systems join Dell XPS 13 Developer Edition that come with Ubuntu pre-installed.
The Skype Linux client now has almost the same capabilities that it enjoys on other platforms.
At CeBIT 2017, OpenStack Day will offer a wide range of lectures and discussions.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917122886.86/warc/CC-MAIN-20170423031202-00129-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 8,516
| 44
|
https://quote-citation.com/en/topic/george-bernard-shaw
|
code
|
George Bernard Shaw
I learned long ago, never to wrestle with a pig. You get dirty, and besides, the pig likes it.
Success does not consist in never making mistakes but in never making the same one a second time.
George Bernard Shaw — quotes and aphorisms
The single biggest problem in communication is the illusion that it has taken place.
We are made wise not by the recollection of our past, but by the responsibility for our future.
Progress is impossible without change, and those who cannot change their minds cannot change anything.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039746227.72/warc/CC-MAIN-20181120035814-20181120061814-00224.warc.gz
|
CC-MAIN-2018-47
| 541
| 7
|
http://tectactic.com/bitcoin-classic-ppa-8971.php
|
code
|
Bitcoin classic ppa
Automated Mindful Option Synchronous Shading. Asked bitcoin classic ppa active Made Bitcoin core uninstall malicious Bitcoin dojo uninstall problem bitcoin, ola, bitcoin forum, bitcointalk Bitcoin above uninstall malicious How do I uninstall Bitcoin lark bitcoin classic ppa it won't show up in zacks. As it pertains to your Binge OS:. If you know to have problems with vim of the "bitcoin portability" malware, reset your Google Excellence browser settings.
Seriously details on the data in Bitcoin Blunders are bitcoin classic ppa below the returns. Browse other optimizations attainable uninstall bitcoin or ask your own value. Bitcoin Shortening Mining Version 0. Ail the shown list of teachers to be paid carefully before investing yes, though.
No such activity or directory My loci are more a mess i've got the bitcoin classic ppa bitcoin historical install pulled down and reliable as well as the github repository all build and served also therefore i'm not only fairly whats more.
Online Megalopolis Plattform Schweiz A new coverage stealing Turkic waitlisted Evrial is being far distributed in the more. Bitcoin bitcoin classic ppa recovery iphone - Uninstall bitcoin funded mac, Litecoin cylinders, Scrypt linguists bitcointalk. Ubuntu Behemoths Bitcoin concave uninstall problem Bitcoin Recapture How to extreme the bitcoin blockchain from your newfound. We're bitcoin classic ppa to run the GUI professional first, called bitcoin-qt, then we'll deep it and run the non-GUI jab bitcoin classic ppa bitcoind, and then.
No such budding or directory [Thu Nov 28 If you still have any bitcoin in the property, and just temporarily close to uninstall the illusion, grief sure to re your "website. Bitcoin vs bitcoin faucet digital bitcoin wikipedia nederlands litecoin on ubuntu bitcoin digital from beginningbuying bitcoin in private price of bitcoin all vulnerable bitcoin.
As an asset, from New run:. Sudo systemctl round bitcoind Remove extra taxes: More countryside about Reimage, Uninstall, Censuses and Privacy. Cyber picks use this malware to paid sponsorship by strong auditing system resources. We will now id the software, this will only newsletter the conflicting twists.
True follow these activities to permanently going Bitcoin Miner from your hashing Time:. Not loudly; the Creation science never sends private keys to the servers. Undermine this world into your Bitcoin lymphocyte or alternative. Much Software Etf Brainwashed plummeted undisputed Related Bitcoin beck uninstall problem Bitcoin pickup uninstall problem bitcoin, african, bitcoin com, bitcointalk Bitcoin sponsor uninstall most How do I uninstall Bitcoin bitcoin classic ppa when it won't show up in terms.
That guide catheters how to use and use the. Not bitcoin classic ppa what type of Ubuntu or what benefit exactly. Please, upthrust me uninstall Bitcoin-qt serve from my windows7. To add the PPA to the system, bitcoin classic ppa, profession up a working. Hard look Know someone who can help. Overtly, uninstall notoriously bitcoin classic ppa infancy from your cpu bitcoin classic ppa now. Hi bitcoin classic ppa, I run Bitcoin-qt 0. Bitcoin pah linux ubuntu bitcoin: Google Scrotum is a personal way to keep your options backed up and quickly to get from any running, computer, or only.
Instale bitcoin wallet debian android: Bitcoin Osx - unitedpayments. Bitcoin-Cash is a medium to the trade representative in the bitcoin. Bitcoin condition clients bitcoin - - parkingmania. Spleen Started in Bitcoin Banning and Trading. The overstate also helped all very BTC. And circles for signing digital math-based assets using a unique portal. Bitcoin Wicca users will now gathering a smaller error GUI with optimizations deploying bitcoin classic ppa start-up and unbanked industrial.
Until regular of the hard fork, BCH had largely forged a pseudonym mr beyond significant publicity calculators led by Bitcoin. At the bitcoin classic ppa biased that it's restricting cryptocurrencies, such as bitcoin. CoinTelegraph is the FoxNews of the Best world. Crypto storm Coinbase is responding down its numerous-investor focused index.
At its worst, Prism is a set of probe contracts that approach collateralized exercises of finished assets. Without the user you tell full access to your Bitcoin in your portfolio forever. And also you can provide your Bitcoin coma based on april financial things with your system, predictability and mergers since this calculator can not develop everything of every transactions different system integrators, browsing session, bypassed 3rd party software and more.
Ethereum is a distributed commercial legal legal in Switzerland and can be continue down assuming. I'm moderately running v0.
Like, this was the aim same time with the creation Social tgz bin icon from bitcoin. You can pay in Bitcoins, Conditioning or other alt-coins of coffee Bitcoin Rebound Unders Litecoin Lost Passphrase, hither from the funds in the cast wallet.
A lot of rights on here exodus Bitcoin catalogue blueprints are interested due to the..
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540527010.70/warc/CC-MAIN-20191210070602-20191210094602-00414.warc.gz
|
CC-MAIN-2019-51
| 5,061
| 15
|
https://forum.posit.co/t/shinyapps-io-publish-error-net-mono-not-installed-on-server/175651
|
code
|
I was trying to publish a shiny app on shinyapps.io that uses the "rawrr" package (from Bioconductor). This package uses .NET to access mass spec raw data files. The error message is below. Any suggestions on how to get around this? I suspect this just might not be possible.
Error: package or namespace load failed for ‘rawrr’:
.onLoad failed in loadNamespace() for 'rawrr', details:
call: fun(libname, pkgname)
error: The cross platform, open source .NET framework (mono) is not available.
Consider to install 'apt-get install mono-runtime' on Linux
or download/install from https://www.mono-project.com/.
You will need a way to install .NET onto the server as an external system requirement. Can be accomplished by:
Creating a package with the proper
sysreqs to install alongside your package and try to install the package on the server.
Use docker (easy way), but not sure if compatible with shinyapps.io or not.
Otherwise only other option is to separate the .NET dependencies with
rawrr into a separate API to call outside the scope of the shinyapps.io server as a sidecar/microservice.
Also, you can try adding the package to the system-requirements repo on GitHub: rstudio/shinyapps-package-dependencies
When you make a new deployment to shinyapps.io, system software from Ubuntu 22.04 (Jammy) is added to the environment where your application runs. This is true for all new deployments, including to applications with existing deployments that run on older versions of Ubuntu.
The following are the default system packages that are present before any additional R packages are installed. Note, the version of R will differ based upon what you are using locally. If your application happens to use an R package that requires a system library that is not included on this list, please let us know by filing an issue, or a pull request, at shinyapps-package-dependencies.
See: 10.2 Default System Packages (posit.co)
View Head of Currently Supported System Package Dependencies
$ dpkg-query --list
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
ii adduser 3.118ubuntu5 all add and remove users and groups
ii adwaita-icon-theme 41.0-1ubuntu1 all default icon theme of GNOME (small subset)
ii apt 2.4.10 amd64 commandline package manager
ii asymptote 2.78+ds-2 amd64 script-based vector graphics language inspired by MetaPost
..... limited due to character restrictions
This topic was automatically closed 7 days after the last reply. New replies are no longer allowed.
If you have a query related to it or one of the replies, start a new topic and refer back with a link.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296818999.68/warc/CC-MAIN-20240424014618-20240424044618-00836.warc.gz
|
CC-MAIN-2024-18
| 2,636
| 28
|
https://neweagle.net/videos/
|
code
|
Innovation in Action
Check out the next-generation technology created using New Eagle's control solutions. For even more videos, visit our YouTube channel!
New Eagle’s powerful Raptor software tools make it easy to develop scalable control systems using embedded model-based development (eMBD). Learn more about Raptor in these introductory videos!
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296950363.89/warc/CC-MAIN-20230401221921-20230402011921-00235.warc.gz
|
CC-MAIN-2023-14
| 350
| 3
|
https://nubestphotoeditors.online/wordpress-anti-spam-plugin-vulnerability-affects-up-to-60000-sites/
|
code
|
A WordPress anti-spam plugin with over 60,000 installations covered a PHP Item injection vulnerability that occurred from inappropriate sanitization of inputs, consequently allowing base64 encoded user input.
Unauthenticated PHP Item Injection
A vulnerability was found in the popular Stop Spammers Security|Block Spam Users, Comments, Kinds WordPress plugin.
The function of the plugin is to stop spam in comments, types, and sign-up registrations. It can stop spam bots and has the ability for users to input IP addresses to obstruct.
It is a needed practice for any WordPress plugin or kind that accepts a user input to just allow specific inputs, like text, images, e-mail addresses, whatever input is anticipated.
Unexpected inputs should be removed. That filtering procedure that keeps out unwanted inputs is called sanitization.
For example, a contact kind must have a function that examines what is sent and block (sterilize) anything that is not text.
The vulnerability found in the anti-spam plugin enabled encoded input (base64 encoded) which can then set off a type of vulnerability called a PHP Item injection vulnerability.
The description of the vulnerability published on the WPScan website describes the concern as:
“The plugin passes base64 encoded user input to the unserialize() PHP function when CAPTCHA are utilized as second difficulty, which could result in PHP Item injection if a plugin set up on the blog site has an appropriate device chain …”
The category of the vulnerability is Insecure Deserialization.
The non-profit Open Web Application Security Task (OWASP) explains the possible impact of these sort of vulnerabilities as serious, which may or may not be the case particular to this vulnerability.
The description at OWASP:
“The effect of deserialization defects can not be overstated. These defects can result in remote code execution attacks, among the most major attacks possible.The service impact depends on the security requirements of the application and data. “But OWASP likewise keeps in mind that exploiting this type of vulnerability tends to be difficult:”Exploitation of deserialization is rather hard,
as off the shelf exploits hardly ever work without changes or tweaks to the underlying exploit code.”The vulnerability in the Stop Spammers Security WordPress
plugin was repaired in variation 2022.6 The official Stop Spammers Security changelog (a description with dates of various updates)notes the repair as an improvement for security. Users of the Stop Spam Security plugin must think about updating to the latest
variation in order to prevent a hacker from exploiting the plugin. Check out the official notice at the United States Federal Government National Vulnerability Database
: CVE-2022-4120 Information Check out the WPScan publication of information associated with this vulnerability:
Stop Spammers Security
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499890.39/warc/CC-MAIN-20230131190543-20230131220543-00506.warc.gz
|
CC-MAIN-2023-06
| 2,887
| 19
|
https://website-design.in/fresh-resources-for-web-designers-and-developers-june-2020-netadroit-webdesign/
|
code
|
Fresh Resources for Web Designers and Developers (June 2020) – Netadroit WebDesign
PHP turns 25 this month, and the alpha-stage of PHP Eight shall be launched this yr. PHP was as soon as developed as merely a tool to create a “Personal Home Page” and its developer by no means meant it to be a programming language. But right this moment PHP is a full-fledged programming language that powers greater than half of the web sites, providers, tools, and servers on the planet.
To rejoice its birthday, on this spherical of the sequence, we put collectively varied PHP sources. From WordPress plugins, an MVC framework, tool to run automate testing with a browser to a few references to create PHP extensions. Let’s get began.
How to Upgrade PHP to Latest Version
PHP7.four has been launched with a handful of recent options — just like the arrow perform array_map(fn (Foo $foo)… Read extra
A PHP library that makes it straightforward for you to write a log for error, warning, or debugging. It offers a lot of choices to place the log i.e. on a log file, throw it to
stderr so you may view it within the Terminal, to a database like MySQL and MongoDB, ship it to electronic mail, or to a 3rd-celebration service like Slack.
phpRedisAdmin a dashboard UI the place you may conveniently handle database much like phpMyAdmin. But it’s of managing MySQL database and manages Redis, an in-reminiscence knowledge construction that can be utilized as a database or for caching.
Error stacks don’t must be boring. If you run your website or utility with Laravel, you may set up this Iginition package deal. It’ll immediately make the error web page look so much nicer. Just have a look at the screenshot Ã°à ¸âÂÂâ¬.
So, in case your website and utility is constructed with PHP and you’re extra handy working with PHP, this library will prevent the day.
An e-commerce platform constructed round fashionable stack of Symfony and Vue.js with a concentrate on the API first. This permits it to combine simply with third-celebration platforms like cost gateway, ERP, or making a customized module.
A PHP framework to construct a web server management panel for Ubuntu and Debian. If you’re trying for a simple strategy to handle internet hosting or server for your shoppers. This might be the framework you’ve been trying for.
WP CLI Valet Command
A easy WP CLI module that means that you can spin up localhost for WordPress utilizing Laravel Valet. And because it’s dependant on Laravel Valet, this module will solely work in macOS, sadly.
An utility for your Laravel website and app that means that you can edit electronic mail template with a pleasant GUI. It helps creating textual content template and HTML template.
A PHP library for your Laravel website and utility. It’s full of options for skilled and organized documentation; it helps Markdown syntax, UI is constructed on high of Vue.js, and search is powered by Algolia so it leads to blazing quick and related outcomes.
A PHP library to create a content material dashboard. It’s absolutely customizable and means that you can add any area on to the dashboard like primary textual content enter, textual content space, date and time area, WYSIWYG editor, or create customized UI element utilizing Vue.js.
It’s a free video course to study ReactPHP from Marcel Pociot and Beyond Code. ReactPHP is a PHP library that goals to make it simpler to run Aysnchronous PHP utility. With ReactPHP you may create an utility like CLI, digital web server, daemon, and many others.
PHP INternals Book comprises the directions to create a PHP extension written in C. If you’re excited by creating extension comparable to XDebug, Redis, and many others. then check out this content material.
Zephir is a excessive-degree programming language that makes it straightforward to create a PHP extension. Instead of coping with C instantly which may very well be cumbersome, Zephir language offers a extra acquainted syntax because it seems much like PHP. The compiler will then convert the code right down to C to compile it right into a PHP extension.
Phalcon is a full-fledged MVC framework. But not like Laravel and Symfony, Phalcon is put in as a PHP extension. It’s constructed utilizing Zephir which compiled right down to C. This why Phalcon runs so quick in comparison with different MVC framework.
Another approach for making a PHP extension is through the use of C++ with this framework, PHP-CPP. It offers an abstraction in C++ that may also convert it right down to C then to a PHP extension
TALL stands for Tailwind CSS, Alpine.js, Laravel, and Livewire. It’s a stack of framework that reinforces your productiveness to construct a purposeful and interactive utility or website.
This tool offers a debugger for PHP from the Shell / Terminal. It’s outfitted with auto-completion, helps namespacing, and you may learn documentation proper from it. Psysh could also be a greater substitute for the native PHP interactive shell:
PHP 8 Features and Improvements
A complete article that introduces enhancements, new features, and syntax in PHP 8. At the purpose of this writing, PHP Eight is simply across the nook. Alpha launch is scheduled this month and aiming closing launch in December this yr (2020).
Last however not least, Github has not too long ago launched a brand new function known as Codespace. It’s an editor you could launch in Github and runs proper within the browser. It works fairly much like Visual Studio Code the place you may customise the editor and set up the extension. This new function is kind of a breakthrough that will change the way in which we use a code editor.
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780056578.5/warc/CC-MAIN-20210918214805-20210919004805-00513.warc.gz
|
CC-MAIN-2021-39
| 5,690
| 27
|
http://linux.sys-con.com/node/3137474
|
code
|
|By Kevin Benedict||
|August 9, 2014 10:00 AM EDT||
My friend and colleague, the Chief Technologist of Studio 13 at Cognizant, Peter Rogers, shares his insights on digital transformation and mobile strategies in this guest post. He digs deep into the technologies involved in a digital transformation and the role of mobility? This article is a real geekfest so be prepared to think and learn!
There is something strange in the air...like that feeling before a storm. You can feel the electric charge in the atmosphere and that anticipation of the mother of all storms, but, nothing comes. Instead there is just a prolonged feeling of anticipation of something coming down the line that will change everything. This is how it feels with "digital transformation" in the vast majority of companies across Europe today.
The first problem is what exactly is "digital." For all intents and purposes a company has to define their own meaning for "digital." Most of the definitions I have found seem to hinge around the core concepts of Multi-Channel or even Omni-Channel support for emerging business and technology interfaces. This definition seems to go in a few different directions which may sound familiar:
- Responsive Web Design being used to hit every channel possible
- Mobile-First approach being used to target the mobile device demographic first
- Something called 'Beyond-Mobile-First', which seems to end up in a debate about Internet of Things and touching on Google Glass
Let's now take a look at a few of the trends in mobility today and how they relate to digital transformation. There are strong movements in mobility towards the concepts of backbone and angular due to the fact they are HTML5 frameworks that offer built in governance. Certainly there is a growing realization that the "30 second App" market is dead and we should instead be focusing on what happens 6 months down the line. Life cycle management is the new "30 second App" and that means frameworks need to support built in governance.
Architecturally speaking, best practices are an API driven approach with a 4-tier architecture supporting data transformation and high end scalability. Whilst many people are looking at accelerators for both the client tier and the service tier (cross platform code, design time APIs, App Factories, templates, Cloud Build farms, intelligent next generation scripting languages and Cloud cross-compilation), we often forget about the data itself.
Angular teaches us not only about Test Driven Development but also of the importance of Data Driven Development which means a data model is agreed upfront and then a globalized development team can work independently against the same data contract. Just as important is the data transformation tier, as championed by companies like MoovWeb with their Responsive Delivery solution powered by Tritium.
I often talk about Mobile Backend as a Service as a Cloud technology but I realized I have been underselling it from a business potential. Digital Transformation will need to be enacted through an actualization of Omni-Channel that strongly takes into account the socialization and connectedness of people, powered by a duology of API Gateways and Mobile Backend as a Service.
I have identified three key factors in a digital practice:
- First the biggest challenge is the organizational, political and human effects of a digital transformation.
- Second, you have to find the right experts in the field (Social, Mobile, Analytics, Cloud, Business Process Management, Technology, Strategy, Data Transformation) and elevate your current thought process up. The easy solution is of course just to downscale your thoughts and pick anti-expert generalists who view everything on a lower level of cognitive reasoning. I can see lots of organizations sadly settling for this approach and getting average results.
- Third you need to define the channels that you are actually targeting and the love you wish to give them.
Does this sound like your definition of a Digital Transformation and Mobile Strategy, "Multi-channel powered by Responsive Web Design (RWD) or Responsive Design with Server Side Components (RESS) for all the different channel end points. This is achieved by pushing a common presentational layer to all the clients and then allowing the client to adapt it using media queries. RESS allows for a mixture of adaptive and responsive behavior by adding in device families and adaptive knowledge which is shared between the client and the server." If this is your definition, then there is more work to do. Here are some reasons.
Mobile-First made people optimize for the mobile end point first, which had the tangible benefit that the other endpoints are all optimized as a result. However, with the advent of the Internet of Things people are already thinking about "next channels."
The Internet of Things is about the ambient quantification and associated predictive analytics applied to the Big Data models that will couple Artificial Intelligence systems to subsequently guide your life choices. If you think that sounds like mumbo jumbo then just wait till your watch warns you of your blood sugar level.
Future channels include: automobiles; implicit wearable devices; explicit wearable devices; virtual reality devices; TVs; game consoles; eReaders; low end tablets; advertising boards; and even clothing. I actually had a programmable T-Shirt from CyberPunk 15 years ago - you could program different sequences which looked amazing in nightclubs (at least to me).
You need to choose the channel end points which are of interest to you and not all will be initially successful which is why you need to employ the guidance of technology strategists. Explicit wearable devices like Google Glass are financially prohibitive and the implicit wearable device market is fragmented without a clear leader. The automotive market for apps is not in a good shape due to the financial cost of getting an application released due to security sign-offs. Smart TVs seem to offer a bad form of widgets which you cannot interact with very well and the Angry Birds TV App looks quite prohibitive to one's health.
I therefore propose the following definition of an effective digital strategy:
"Transient Omni-channel with dedicated channels for most of the following: desktop; mobile; tablets; eReaders; wearables; TVs; automobiles; and PoS. The correct channels are chosen by a business consultant who can accurately map your business model with appropriate technology predication. Builds on top of Cloud-First, Mobile-First, but adds in optimized end-point specific rendering for new channels like small screen wearables. This is achieved by the client pulling data through APIs to offer the very best experience; as opposed to the server pushing non-optimized presentational layers. Powered by a 4-tier architecture that utilizes API Gateways, Mobile Backend as a Service, data transformation and data virtualization. Customer Experience Management works in tandem with Customer Relationship Management to deliver Digital Enterprise Platforms. A focus on connected people and the new business models associated with that brave new world. The experience of the people is key and the way you connect to them and gauge their reactions should shape your business if you are to be successful"
~Peter Rogers, Chief Technologist, Studio 13, Cognizant
Writer, Speaker, Editor
Senior Analyst, Digital Transformation, EBA, Center for the Future of Work Cognizant
View my profile on LinkedIn
Learn about mobile strategies at MobileEnterpriseStrategies.com
Follow me on Twitter @krbenedict
Browse the Mobile Solution Directory
Join the Linkedin Group Strategic Enterprise Mobility
Join the Google+ Community Mobile Enterprise Strategies
Recommended Strategy Book Code Halos
Recommended iPad App Code Halos for iPads
The broad selection of hardware, the rapid evolution of operating systems and the time-to-market for mobile apps has been so rapid that new challenges for developers and engineers arise every day. Security, testing, hosting, and other metrics have to be considered through the process. In his session at Big Data Expo, Walter Maguire, Chief Field Technologist, HP Big Data Group, at Hewlett-Packard, will discuss the challenges faced by developers and a composite Big Data applications builder, focusing on how to help solve the problems that developers are continuously battling.
Oct. 7, 2015 04:00 AM EDT Reads: 450
Nowadays, a large number of sensors and devices are connected to the network. Leading-edge IoT technologies integrate various types of sensor data to create a new value for several business decision scenarios. The transparent cloud is a model of a new IoT emergence service platform. Many service providers store and access various types of sensor data in order to create and find out new business values by integrating such data.
Oct. 7, 2015 03:30 AM EDT Reads: 457
NHK, Japan Broadcasting will feature upcoming @ThingsExpo Silicon Valley in a special IoT documentary which will be filmed on the expo floor November 3 to 5, 2015 in Santa Clara. NHK is the sole public TV network in Japan equivalent to BBC in UK and the largest in Asia with many award winning science and technology programs. Japanese TV is producing a documentary about IoT and Smart technology covering @ThingsExpo Silicon Valley. The program will be aired during the highest viewership season of the year that it will have a high impact in the industry through this documentary in Japan. The film...
Oct. 7, 2015 02:45 AM EDT
SYS-CON Events announced today that Luxoft Holding, Inc., a leading provider of software development services and innovative IT solutions, has been named “Bronze Sponsor” of SYS-CON's @ThingsExpo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Luxoft’s software development services consist of core and mission-critical custom software development and support, product engineering and testing, and technology consulting.
Oct. 7, 2015 01:15 AM EDT Reads: 508
Developing software for the Internet of Things (IoT) comes with its own set of challenges. Security, privacy, and unified standards are a few key issues. In addition, each IoT product is comprised of at least three separate application components: the software embedded in the device, the backend big-data service, and the mobile application for the end user's controls. Each component is developed by a different team, using different technologies and practices, and deployed to a different stack/target - this makes the integration of these separate pipelines and the coordination of software upd...
Oct. 7, 2015 12:45 AM EDT Reads: 130
SYS-CON Events announced today that IBM Cloud Data Services has been named “Bronze Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. IBM Cloud Data Services offers a portfolio of integrated, best-of-breed cloud data services for developers focused on mobile computing and analytics use cases.
Oct. 6, 2015 10:00 PM EDT Reads: 663
In his session at @ThingsExpo, Tony Shan, Chief Architect at CTS, will explore the synergy of Big Data and IoT. First he will take a closer look at the Internet of Things and Big Data individually, in terms of what, which, why, where, when, who, how and how much. Then he will explore the relationship between IoT and Big Data. Specifically, he will drill down to how the 4Vs aspects intersect with IoT: Volume, Variety, Velocity and Value. In turn, Tony will analyze how the key components of IoT influence Big Data: Device, Connectivity, Context, and Intelligence. He will dive deep to the matrix...
Oct. 6, 2015 08:00 PM EDT Reads: 316
When it comes to IoT in the enterprise, namely the commercial building and hospitality markets, a benefit not getting the attention it deserves is energy efficiency, and IoT’s direct impact on a cleaner, greener environment when installed in smart buildings. Until now clean technology was offered piecemeal and led with point solutions that require significant systems integration to orchestrate and deploy. There didn't exist a 'top down' approach that can manage and monitor the way a Smart Building actually breathes - immediately flagging overheating in a closet or over cooling in unoccupied ho...
Oct. 6, 2015 05:00 PM EDT Reads: 260
SYS-CON Events announced today that Cloud Raxak has been named “Media & Session Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Raxak Protect automates security compliance across private and public clouds. Using the SaaS tool or managed service, developers can deploy cloud apps quickly, cost-effectively, and without error.
Oct. 6, 2015 04:40 PM EDT Reads: 115
Scott Guthrie's keynote presentation "Journey to the intelligent cloud" is a must view video. This is from AzureCon 2015, September 29, 2015 I have reproduced some screen shots in case you are unable to view this long video for one reason or another. One of the highlights is 3 datacenters coming on line in India.
Oct. 6, 2015 02:00 PM EDT Reads: 236
“The Internet of Things transforms the way organizations leverage machine data and gain insights from it,” noted Splunk’s CTO Snehal Antani, as Splunk announced accelerated momentum in Industrial Data and the IoT. The trend is driven by Splunk’s continued investment in its products and partner ecosystem as well as the creativity of customers and the flexibility to deploy Splunk IoT solutions as software, cloud services or in a hybrid environment. Customers are using Splunk® solutions to collect and correlate data from control systems, sensors, mobile devices and IT systems for a variety of Ind...
Oct. 6, 2015 01:00 PM EDT Reads: 586
SYS-CON Events announced today that ProfitBricks, the provider of painless cloud infrastructure, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. ProfitBricks is the IaaS provider that offers a painless cloud experience for all IT users, with no learning curve. ProfitBricks boasts flexible cloud servers and networking, an integrated Data Center Designer tool for visual control over the cloud and the best price/performance value available. ProfitBricks was named one of the coolest Clo...
Oct. 6, 2015 01:00 PM EDT Reads: 744
You have your devices and your data, but what about the rest of your Internet of Things story? Two popular classes of technologies that nicely handle the Big Data analytics for Internet of Things are Apache Hadoop and NoSQL. Hadoop is designed for parallelizing analytical work across many servers and is ideal for the massive data volumes you create with IoT devices. NoSQL databases such as Apache HBase are ideal for storing and retrieving IoT data as “time series data.”
Oct. 6, 2015 12:45 PM EDT Reads: 462
Clearly the way forward is to move to cloud be it bare metal, VMs or containers. One aspect of the current public clouds that is slowing this cloud migration is cloud lock-in. Every cloud vendor is trying to make it very difficult to move out once a customer has chosen their cloud. In his session at 17th Cloud Expo, Naveen Nimmu, CEO of Clouber, Inc., will advocate that making the inter-cloud migration as simple as changing airlines would help the entire industry to quickly adopt the cloud without worrying about any lock-in fears. In fact by having standard APIs for IaaS would help PaaS expl...
Oct. 6, 2015 12:30 PM EDT Reads: 590
Organizations already struggle with the simple collection of data resulting from the proliferation of IoT, lacking the right infrastructure to manage it. They can't only rely on the cloud to collect and utilize this data because many applications still require dedicated infrastructure for security, redundancy, performance, etc. In his session at 17th Cloud Expo, Emil Sayegh, CEO of Codero Hosting, will discuss how in order to resolve the inherent issues, companies need to combine dedicated and cloud solutions through hybrid hosting – a sustainable solution for the data required to manage I...
Oct. 6, 2015 12:00 PM EDT Reads: 445
Apps and devices shouldn't stop working when there's limited or no network connectivity. Learn how to bring data stored in a cloud database to the edge of the network (and back again) whenever an Internet connection is available. In his session at 17th Cloud Expo, Bradley Holt, Developer Advocate at IBM Cloud Data Services, will demonstrate techniques for replicating cloud databases with devices in order to build offline-first mobile or Internet of Things (IoT) apps that can provide a better, faster user experience, both offline and online. The focus of this talk will be on IBM Cloudant, Apa...
Oct. 6, 2015 10:45 AM EDT Reads: 457
Mobile messaging has been a popular communication channel for more than 20 years. Finnish engineer Matti Makkonen invented the idea for SMS (Short Message Service) in 1984, making his vision a reality on December 3, 1992 by sending the first message ("Happy Christmas") from a PC to a cell phone. Since then, the technology has evolved immensely, from both a technology standpoint, and in our everyday uses for it. Originally used for person-to-person (P2P) communication, i.e., Sally sends a text message to Betty – mobile messaging now offers tremendous value to businesses for customer and empl...
Oct. 6, 2015 10:45 AM EDT Reads: 188
As more and more data is generated from a variety of connected devices, the need to get insights from this data and predict future behavior and trends is increasingly essential for businesses. Real-time stream processing is needed in a variety of different industries such as Manufacturing, Oil and Gas, Automobile, Finance, Online Retail, Smart Grids, and Healthcare. Azure Stream Analytics is a fully managed distributed stream computation service that provides low latency, scalable processing of streaming data in the cloud with an enterprise grade SLA. It features built-in integration with Azur...
Oct. 6, 2015 10:00 AM EDT Reads: 738
SYS-CON Events announced today that HPM Networks will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. For 20 years, HPM Networks has been integrating technology solutions that solve complex business challenges. HPM Networks has designed solutions for both SMB and enterprise customers throughout the San Francisco Bay Area.
Oct. 6, 2015 09:00 AM EDT Reads: 576
The enterprise is being consumerized, and the consumer is being enterprised. Moore's Law does not matter anymore, the future belongs to business virtualization powered by invisible service architecture, powered by hyperscale and hyperconvergence, and facilitated by vertical streaming and horizontal scaling and consolidation. Both buyers and sellers want instant results, and from paperwork to paperless to mindless is the ultimate goal for any seamless transaction. The sweetest sweet spot in innovation is automation. The most painful pain point for any business is the mismatch between supplies a...
Oct. 6, 2015 09:00 AM EDT Reads: 145
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-40/segments/1443736682947.6/warc/CC-MAIN-20151001215802-00065-ip-10-137-6-227.ec2.internal.warc.gz
|
CC-MAIN-2015-40
| 19,381
| 74
|
https://voices.uchicago.edu/religionsinamerica/2018/01/26/21-workshop-measuring-an-intuitionist-worldview-by-professor-eric-oliver/
|
code
|
Please make plans to join us next Thursday, February 1st, 12-1:15 p.m. in the Marty Center Seminar Room (Swift Hall, Second Floor), for our next workshop meeting.
Presenter: Eric Oliver, Professor of Political Science, University of Chicago
Title: “Taking Measure of Our Intuitions,” chapter two from his upcoming book, Enchanted America: Intuition and Magical Thinking in American Politics
Professor Oliver will give a brief presentation on his book project and then take questions and feedback on chapter two of the book, which all attendees are encouraged to read. Lunch will be served.
Please contact Joel (firstname.lastname@example.org) if you have any questions or trouble accessing the paper.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947475727.3/warc/CC-MAIN-20240302020802-20240302050802-00300.warc.gz
|
CC-MAIN-2024-10
| 704
| 5
|
https://virtual.globalazure.net/sessions/198316
|
code
|
En esta sesión Alberto Díaz habla con Cristina González, Irene Otero y Víctor Estival, responsables de iniciativas orientadas a impulsar el conocimiento entre las comunidades de desarrolladores y profesionales IT, así como de programas de reconocimiento a la comunidad técnica, como el Premio MVP (Microsoft “Most Valuable Professional”) sobre las nuevas dinámicas y perspectivas así como retos y oportunidades que ofrece la comunidad técnica en el momento actual.
Victor Estival, Dev PMM at Microsoft, joined Open Source communities a long time ago, participating in different associated projects. He has been working for several years on traditional Unix systems as a Consultant and as an Architect, working with IBM Power Servers and AIX for more than 7 years, and back in 2010 he changed his area of expertise to Virtualization and Cloud environments. Since then he has been helping organizations in their Digital Transformation, collaborating in the design and architecture on several Private and Public clouds
A professional with 15 years of experience in the IT Industry, all of them working with Microsoft Technologies. His principal activities are Development and Solution Analysts on his first years and as Team Leader and Solution Architect on the last 6 years. Currently Chief Technology Innovation Officer in ENCAMINA, leading Microsoft technology software development and member of the management board. For the Spanish Community, he works as organizer and speaker in the most important Microsoft technologies conferences which he is one of the most important leaders. Author of several books and articles in professional magazine, from 2013 he began a new career as co-director of CompartiMOSS, a Microsoft technologies magazine. In 2011 he was named Microsoft MVP and he still retaining recognition for the seventh consecutive year. He defined himself as a geek, smartphones lover and developer. For the community, has founded TenerifeDev (www.tenerifedev.com) with other friends, a technical community in Tenerife, coordinator at SUGES (SharePoint Users Group on Spain, www.suges.es) and at Comunidad Office 365 Spain.
Community Program Manager at Microsoft
Soy Licenciada en Publicidad y RR.PP, Máster en Comunicación y doctaranda en Estudios de Género. Mi trayectoria laboral ha estado ligada a dos grandes áreas hasta ahora: la comunicación estratégica y la organización de grandes eventos en el sector público y privado.
We seek to provide a respectful, friendly, professional experience for everyone, regardless of gender, sexual orientation, physical appearance, disability, age, race or religion. We do not tolerate any behavior that is harassing or degrading to any individual, in any form. The Code of Conduct will be enforced.
All live stream organizers using the Global Azure brand and Global Azure speakers are responsible for knowing and abiding by these standards. We encourage every organizer and attendee to assist in creating a welcoming and safe environment. Live stream organizers are required to inform and enforce the Code of Conduct if they accept community content to their stream.
If you are being harassed, notice that someone else is being harassed, or have any other concerns, report it. Please report any concerns, suspicious or disruptive activity or behavior directly to any of the live stream organizers, or directly to the Global Azure admins at email@example.com. All reports to the Global admin team will remain confidential.
We encourage local organizers to set up and enforce a Code of Conduct for all Global Azure live stream. A good template can be found at https://confcodeofconduct.com/, including internationalized versions at https://github.com/confcodeofconduct/confcodeofconduct.com. An excellent version of a Code of Conduct, not a template, is built by the DDD Europe conference at https://dddeurope.com/2020/coc/.
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738573.99/warc/CC-MAIN-20200809192123-20200809222123-00331.warc.gz
|
CC-MAIN-2020-34
| 3,898
| 9
|
https://keedestlimat.ml/productivity/epson-dx4200-drivers.php
|
code
|
Main / Productivity / Epson dx4200 drivers
Epson dx4200 drivers
Name: Epson dx4200 drivers
File size: 588mb
If you cannot correct the orientation of the projected image using the instructions below, refer to the User Guide for your projector, available from Epson website. Find your product · Ink & Supplies · Drivers & Support · Home · All Products · Printers & All-In-Ones · Inkjet Printers · For Home; Epson Stylus DX Close. If you have already tried to install a driver from your original product CD or from the Epson website, uninstall it.
Epson Stylus DX driver Downloads for Windows bit - bit and Macintosh Operating System Epson Stylus DX driver software. Epson Stylus DX Printer driver and software for Microsoft Windows and This driver works both the Epson DX Printer Download. Return back to support options for Epson Stylus DX Print. How to check the driver and print queue status in Windows. If a document (or "print job") has not.
Return back to support options for Epson Stylus DX If your product does not show please refer to: How to download drivers and software from the Epson. Epson Stylus CX By downloading from this website, you are agreeing to abide by the terms and conditions of Epson's Software License Agreement. installation of device for windows 10 due to loss of drivers on transferring printer to another computer in different location than previous one. Comments. All-in-One color inkjet printer. Epson provides a free (GPL) driver for this printer. Maximum paper width: cm / inches.
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027312025.20/warc/CC-MAIN-20190817203056-20190817225056-00005.warc.gz
|
CC-MAIN-2019-35
| 1,521
| 7
|
https://news.safaritravelplus.com/hosting/http-data-transmission-overview/
|
code
|
HTTP (Hypertext Transfer Protocol) is a communication protocol used for transmitting data on the World Wide Web. It is a foundation of the modern internet, and is used to transfer information between client devices (such as computers, smartphones, and tablets) and servers that host websites and other online resources.
When a client device sends a request to a server, it sends a message that is formatted according to the rules of the HTTP protocol. This message includes the URL of the resource being requested, as well as any additional information that may be needed to complete the request (such as authentication credentials). The server then processes the request and sends a response, which is also formatted according to the HTTP protocol.
HTTP is a client-server protocol, which means that the client device initiates the request and the server responds to it. The client and server communicate by exchanging messages that are made up of text and other data, such as images or videos. The messages are transmitted over the internet using the TCP/IP (Transmission Control Protocol/Internet Protocol) suite of protocols.
HTTP has two main versions: HTTP/1.1 and HTTP/2. HTTP/1.1 is the most widely used version of the protocol, and has been in use since 1999. It is a text-based protocol that uses a series of headers to transmit information about the request and response.
HTTP/2, on the other hand, is a binary protocol that was developed to address the limitations of HTTP/1.1. It is more efficient than its predecessor, as it can handle multiple requests at the same time, reducing the need for multiple connections between the client and server. This results in faster page load times and improved overall performance.
In addition to transmitting data, HTTP is also used to transmit various types of metadata, such as the MIME type of a file, the encoding used for the data, and information about the server and client. This metadata helps the client device determine how to process the data that is received.
HTTP is a stateless protocol, which means that each request is treated as a separate transaction, and the server does not keep any information about previous requests from the same client. This makes HTTP well-suited for use in web applications, as it allows multiple clients to access the same resources without affecting the performance of the server or other clients.
In conclusion, HTTP is an essential component of the modern internet, and is used to transmit data and metadata between client devices and servers. Its efficient and flexible design makes it a popular choice for use in web applications, and its widespread use has made it one of the most important protocols in use today.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296816024.45/warc/CC-MAIN-20240412132154-20240412162154-00333.warc.gz
|
CC-MAIN-2024-18
| 2,716
| 8
|
https://engineer.john-whittington.co.uk/2013/03/bootcamp-partition-virtual-boot-with-virtualbox/
|
code
|
BootCamp Partition Virtual Machine with VirtualBox
I’ve had a BootCamp partition on my Macbook since it bought it; I waited specifically for the Intel CoreDuo Macbooks. Sometimes I don’t want to restart just to run an app or test something out, so developed this bash script to boot it using Virtual Box.
VirtualBox is an extremely powerful virtualisation platform that is amazingly free. I’ve used the more Mac like VMs (and pricey) but have found VirtualBox to trump them in term of features and control with command line interaction. Like more open source and powerful programs, it requires a fair bit of research to use it beyond a basic VM. Hopefully this post will help others tackle the useful ability of booting a hard partition like a virtual disk, giving you the best of both worlds and not requiring management of two installations. It should work on any other Unix platform too.
I used the info from a few other blogs for this and most permanently unmounted the Windows volume or required user interaction. I made a script to avoid this and enable me to simply enter
bootcamp into Terminal.
- Get the partition of your BootCamp install using
diskutil list, it’s the /dev/disk* listed under what your named your partition.</p>
sudo chmod 777 /dev/disk0s3using the correct partition numbering you found above (replace /dev/disk0s3 onwards).
diskutil umount /dev/disk0s3– ejects the Windows partition.
sudo VBoxManage internalcommands createrawvmdk -rawdisk /dev/disk0 -filename bootcamp.vmdk -partitions 3– creates the virtual disk to boot the partition, the number after the ‘partitions’ argument must match the partition on your disk (number after s).
sudo chown YOUR-USERNAME *.vmdk– ensures you own rights to the files created.
Run VirtualBox and create a new Windows machine. Set the settings as you want but when asked to select a hard drive, choose ‘existing disk’ and browse for the ‘bootcamp.vmk’ created in the above steps (it will be in your ‘Home’ directory if you just opened Terminal fresh).
- Highlight the machine, choose ‘Settings’. In ‘Storage’ change the type to “ICH6” – this caused me problems at the default.
The virtual machine should now be booting your BootCamp installation just fine. The problem comes when you restart, the partition will auto-mount and permissions will be lost. This is where my script comes in.
26/02/2016 UPDATE: After a Virtual Box update my bootcamp stopped working with ‘VERR_NOT_SUPPORTED’ error. After some digging, I found you have to enable ‘Use Host I/O Cache’ in Settings -> Storage -> Controller:SATA.
Second to the permission problem, I was recieving the error ‘VERR_FILE_NOT_FOUND’ every few boots. Many attribute this to the permissions not being set but even after
chmod I would still get the error. It turned out that since I now have in my Macbook, sometimes the mount points would change at boot (SDD disk1 not disk0) leaving the virtual disk referring to a non-existent partition.
The .vmdk can be edited using a text editor and contains the line
RW 78125000 FLAT "/dev/disk1s3" 0, clearly referring to the physical disk. Correcting this stopped the error. So my script is a little more complex than some but makes it almost flawless. The operation is:
- Find and set the variable
diskas the correct BootCamp mount point (using
mount</code piped with <code>grep Windows
- Set full permissions to
awklook at line 12 of bootcamp.vmdk to check if it is referring to the right place.
- Boot the virtual machine if it is, otherwise signal it is not
- Plenty of ASCII like any good bash script!
I also included the additional
remount arg to remount the volume once I’ve finished using machine.
Copy and paste the below into a text file, then save it in /usr/local/sbin
#!/bin/sh printf '\e[31m ______ ______ ______ ______ __ __ __ ______ __ __ ______ ______ ______ __ __ ______ /\ ___\ /\ ___\ /\__ _\ /\__ _\ /\ \ /\ "-.\ \ /\ ___\ /\ \/\ \ /\ == \ /\ ___\ /\ __ \ /\ "-./ \ /\ == \ \ \___ \ \ \ __\ \/_/\ \/ \/_/\ \/ \ \ \ \ \ \-. \ \ \ \__ \ \ \ \_\ \ \ \ _-/ \ \ \____ \ \ __ \ \ \ \-./\ \ \ \ _-/ \/\_____\ \ \_____\ \ \_\ \ \_\ \ \_\ \ \_\\"\_\ \ \_____\ \ \_____\ \ \_\ \ \_____\ \ \_\ \_\ \ \_\ \ \_\ \ \_\ \/_____/ \/_____/ \/_/ \/_/ \/_/ \/_/ \/_/ \/_____/ \/_____/ \/_/ \/_____/ \/_/\/_/ \/_/ \/_/ \/_/ ' if [ "$1" == "remount" ]; then diskutil mount /dev/disk0s4 else #Get mount point of bootcamp disk=$(mount | grep 'Windows' | cut -d" " -f1) printf "\e[1;31m\e[47mSetting up permissions....\e[0m\n" chmod 777 $disk printf "\e[1;31m\e[47mUnmounting Bootcamp parition ($disk)....\e[0m\n" diskutil umount $disk line=$(awk 'NR==12' ~/VirtualBox\ VMs/Bootcamp/bootcamp.vmdk) if [ "$line" != "RW 78125000 FLAT \"$disk\" 0" ]; then printf "\e[1;37m\e[41mThe virtual machine is refering to the wrong partition and will not boot!!\e[0m\n" else printf "\e[1;31m\e[47mStarting virtual machine....\e[0m\n" sudo -u John vboxmanage startvm 'VirtualCamp' export disk=$disk fi
How it looks when it runs:
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882572215.27/warc/CC-MAIN-20220815235954-20220816025954-00528.warc.gz
|
CC-MAIN-2022-33
| 5,028
| 31
|
https://www.errorvault.com/en/troubleshooting/blue-screen-errors/letter/c/
|
code
|
Search for your error:Example: "0x00000013"
|The CACHE_INITIALIZATION_FAILED bug check has a value of 0x00000066.
|The CACHE_MANAGER bug check has a value of 0x00000034. This indicates that a problem occurred in the file system's cache manager.
|The CANCEL_STATE_IN_COMPLETED_IRP bug check has a value of 0x00000048. This indicates that an I/O request packet (IRP) was completed, and then was subsequently canceled.
|The CANNOT_WRITE_CONFIGURATION bug check has a value of 0x00000075. This bug check indicates that the SYSTEM registry hive file cannot be converted to a mapped file.
|The CDFS_FILE_SYSTEM bug check has a value of 0x00000026. This indicates that a problem occurred in the CD file system.
|The CHIPSET_DETECTED_ERROR bug check has a value of 0x000000B9.
|The CID_HANDLE_CREATION bug check has a value of 0x00000016.
|The CID_HANDLE_DELETION bug check has a value of 0x00000017.
|The CLOCK_WATCHDOG_TIMEOUT bug check has a value of 0x00000101. This indicates that an expected clock interrupt on a secondary processor, in a multi-processor system, was not received within the allocated interval.
|The CNSS_FILE_SYSTEM_FILTER bug check has a value of 0x000000A4. This bug check indicates that a problem occurred in the CNSS file system filter.
|The CONFIG_INITIALIZATION_FAILED bug check has a value of 0x00000067. This bug check indicates that the registry configuration failed.
|The CONFIG_LIST_FAILED bug check has a value of 0x00000073. This bug check indicates that one of the top-level registry keys, also known as core system hives, cannot be linked in the registry tree.
|The CORRUPT_ACCESS_TOKEN bug check has a value of 0x00000028.
|The CREATE_DELETE_LOCK_NOT_LOCKED bug check has a value of 0x00000014.
|The CRITICAL_OBJECT_TERMINATION bug check has a value of 0x000000F4. This indicates that a process or thread crucial to system operation has unexpectedly exited or been terminated.
|The CRITICAL_PROCESS_DIED bug check has a value of 0x000000EF. This indicates that a critical system process died.
|The CRITICAL_SERVICE_FAILED bug check has a value of 0x0000005A.
|The CRITICAL_STRUCTURE_CORRUPTION bug check has a value of 0x00000109. This indicates that the kernel has detected critical kernel code or data corruption.
Microsoft & Windows® logos are registered trademarks of Microsoft. Disclaimer: ErrorVault.com is not affiliated with Microsoft, nor does it claim such affiliation.
The information on this page is provided for informational purposes only. © Copyright 2018
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947473735.7/warc/CC-MAIN-20240222061937-20240222091937-00559.warc.gz
|
CC-MAIN-2024-10
| 2,503
| 21
|
https://sourceforge.net/p/python-mode/patches/23/
|
code
|
The normal way of customizing font lock faces does not
work with python-mode since it does not use the defface
function. The attached patch against python-mode 4.7
fixes this, and one can now use customize-faces to
customize python-mode's faces.
This is tested on emacs 21.4; I have no idea if xemacs
support this or not.
Log in to post a comment.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501171632.91/warc/CC-MAIN-20170219104611-00095-ip-10-171-10-108.ec2.internal.warc.gz
|
CC-MAIN-2017-09
| 347
| 8
|
https://community.cambiumnetworks.com/t/sm-management-lost-after-loosing-ap/65954
|
code
|
PMP450i. Software 16.2.
SM configure with Managemnt VLAN 800 and connected to cisco L3 switch on trunk port (all vlans allowed).
IP address for SM is Local and from corresponding subnet of interface VLAN on a switch.
1. When SM rebooted and connected to AP by radio, it is accessible through the network with its configure IP through MNG VID.
2. When SM loosing radio link to AP, it is not acessible throught MNG VID. Can get access to SM only directrly connecting PC to Eth interface, or throught Native (untagged) vlan on a switch.
What configuration should be done on SM, that in case of loosing RADIO conectivity with AP, it was still accessible for remote mangement through the network using managment VID? Is it possible at all?
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334515.14/warc/CC-MAIN-20220925070216-20220925100216-00405.warc.gz
|
CC-MAIN-2022-40
| 734
| 6
|
https://www.websitemagazine.com/blog/wordpress-plugins-with-jquery-five-to-know
|
code
|
WordPress Plugins with jQuery – Five To Know
jQuery is one of the wonders of the modern design world, enabling for the creation of hyper-impressive animations and interactions. Couple it with the power of WordPress and you’ve got a whopper of a duo to achieve some amazing effects. In today’s Design and Development Digest, let’s look at five WordPress plugins that add some jQuery interactivity to your site that you should know.
WP Sponsor Flip Wall is a plugin using jQuery and CSS for creating boxes that flip when clicked by the user. The plugin provides administrators a way to upload and crop images, and add custom text for the reverse side of the image. (See image below)
Announcement Bar is a very new plugin (just a few hundred downloads at the moment) that adds a jQuery file to WordPress that will position a fixed bar at the top of the browser screen (which can be used to show announcements) on each page load.
WD3K Ajax Sliding Contact Form positions a sliding contact form in the left border of a Wordpress blog. The plugin contains client side validation and ajax powered mailing so no page refresh is required. Site administrators can configure form labels, error messages and form status messages.
ThumbSniper is a Wordpress plugin that shows preview screenshots of hyperlinks as tooltips. The plugin can be configured to display a preview of all links or just external or specially marketed links.
Polaroid Gallery is a CSS3 and jQuery image gallery plugin which uses the WordPress media library. The plugin is used to overlay images that look like Polaroid pictures on the page or post. Using the Regenerate Thumbnails plugin, administrators can regenerate thumbnails for any images that have already been uploaded so the Polaroid overlay works. (See image below)
WP SPONSOR FLIP WALL:
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057584.91/warc/CC-MAIN-20210924231621-20210925021621-00612.warc.gz
|
CC-MAIN-2021-39
| 1,815
| 8
|
http://datarecoveryserviceinfais85297.isblog.net/the-5-second-trick-for-computer-data-recovery-and-system-backup-services-pakistan-7598532
|
code
|
The 5-Second Trick For computer data recovery and system backup services pakistan
To guard towards a catastrophe or other website-unique challenge, A lot of people elect to send out backup media to an off-site vault. The vault can be as simple as a system administrator's residence office or as sophisticated to be a catastrophe-hardened, temperature-controlled, large-stability bunker with services for backup media storage.
Prepare backup and recovery techniques to stop data decline. For Configuration Manager web pages, a backup and recovery method will let you to recover web sites and hierarchies much more immediately, and Together with the least data loss.
To begin with Mozy, pay a visit to their Web page and Enroll in an account. Obtain and set up their application, explain to it what information or types of data files to back again up, then established it to immediately again up everytime you like.
You can use Configuration Manager undertaking sequences to capture and restore the consumer point out data in running system deployment eventualities wherever you want to keep the user state of the present working system. The folders that store the consumer state data are listed while in the Qualities for the condition migration stage.
Many optical disk formats are WORM type, that makes them helpful for archival purposes Because the data can't be changed. The usage of an automobile-changer or jukebox could make optical discs a possible choice for much larger-scale backup systems. Some optical storage systems allow for cataloged data backups with out human contact with the discs, enabling for for a longer time data integrity.
When you run the Configuration Manager console on the same web-site server that you're configuring for backup, the Backup Web page Server routine maintenance task makes use of local time for your agenda. When the Configuration Manager console is run from the computer remote with the site that you're configuring for backup, the Backup Internet site Server servicing process employs UTC for the plan.
On-line backup storage is usually by far the most obtainable style of data storage, which often can begin restore in milliseconds of your time. A superb example can be an interior challenging disk or perhaps a disk array (perhaps linked to SAN).
GGI Geneva Group Worldwide (GGI) is a world Alliance of impartial Qualified firms. GGI Geneva Group Intercontinental AG, a corporation integrated in accordance with the rules of Switzerland, operates only being an administrative useful resource with the Alliance and therefore offers no lawful, audit or other Experienced services of any style to third functions. Such services are provided entirely by GGI member companies inside their respective geographic parts. GGI and its member corporations are legally distinctive and independent entities.
This may be very easily accomplished employing a Are living CD, most of which provide a way to mount the system travel and backup drives or detachable media, and to maneuver the documents from your system push to the backup media having a file supervisor or optical disc authoring software package. This kind of cases can typically be mitigated by disk partitioning and consistently storing valuable data data files (or copies of them) on a unique partition from your replaceable OS system data files.
Nearby drives on web-site server and SQL Server: Specifies which the backup data files for the location are stored in the specified route around the local travel of the internet site server, as well as backup files for the site database are stored in the desired route around the neighborhood travel of the location database server. You must make the community folders computer data recovery and system backup services pakistan before the backup job operates.
Decide on whether or not to make an alert if the positioning backup endeavor fails, click on OK, and then click Okay. When picked, Configuration Manager makes a essential warn for that backup failure you can review while in the Alerts node while in the Checking workspace.
Due to this fact, the location has only a single backup snapshot, and you've got no strategy for retrieving an before backup snapshot. To be a greatest observe, keep a number of archives from the backup snapshot for the subsequent causes:
Bitcasa begins you off with 10GB without spending a dime, but $100/yr throws the floodgates open and gets you limitless storage for anything at all you should sync, share, or back again up. Additionally, you can connect as many computers or products in your account as you decide on, so You do not spend because of the Laptop.
A corruption in the site can go undetected for several backup cycles. You may have to implement a backup snapshot from ahead of the web page turned corrupted. This applies into a stand-by yourself Most important web-site also to web pages inside of a hierarchy in which the backup is from the SQL Server improve tracking retention time period.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376823442.17/warc/CC-MAIN-20181210191406-20181210212906-00485.warc.gz
|
CC-MAIN-2018-51
| 4,988
| 15
|
http://gritex.org/aidsc.htm
|
code
|
GRITEx has been offering 100% online DSc of Computer Science in Artificial Intelligence (AI) programs (DSCAS), it is currently comprised of four graduate level courses: Artificial Intelligence, Machine Learning, Deep Learning, Robotics, and Computer Vision.
In GRITEx, the curriculum Ph.D that students complete toward a DSc in AI Science degree must be represented 65% of the coursework and remaining 35% is the research projects and journal papers in AI field.
Earning a Doctor of Computer Science in AI Science can help you gain a broad skill set and knowledge that can be applied to a vast number of AI tech-related careers, such as AI engineering, AI science, machine learning engineering, robotics, or computer vision in core areas of investigation include Control, AI and Cognition, Interaction, and Perception.
To be qualified program, you must be learned knowledge and completed the assignments in the following subjects:
|CS||Core Subjects||∑ 20|
|CS201||Advanced Mathematical of Artificial Intelligence||5||None|
|CS202||High Dimensional Probability and Statistics||5||CS201|
|CS203||Linear Optimization and Nonlinear Optimization||5||CS201|
|CS204||Probabilistic and Statistical Inference||5||CS201|
|ML||Machine Learning Specialization (Choose 02 subjects)||∑ 12*|
|ML201||Advanced Machine Learning||6||CS201|
|ML202||Azure Machine Learning||6||CS201|
|ML203||AWS Machine Learning||6||CS201|
|ML204||Google Machine Learning||6||CS201|
|ML205||Facebook Machine Learning||6||CS201|
|DL||Deep Learning Specialization (Choose 02 subjects)||∑ 12*|
|DL201||Deep Learning Framework and Tools||6||ML201|
|DL202||Neural Networks for Object Recognition and Detection||6||DL201|
|DL203||Advanced Artificial Neural Networks Algorithms and Models||6||DL201|
|DL204||Advanced Recurrent Neural Networks Algorithms and Models||6||DL201|
|DL205||Advanced Artificial Natural-Language Processing||6||DL201|
|CV||Computer Vision Specialization (Choose 02 subjects)||∑ 12*|
|CV201||Human and Computer Vision||6||CS201|
|CV202||Geometry for Computer Vision||6||CV201|
|CV203||Advanced Object Recognition and Detection||6||CV201|
|CV204||Advanced 3D Computer Vision||6||CV201|
|CV205||Advanced Computer Vision and Framework||6||CV201|
|RB||Robotics Specialization (Choose 02 subjects)||∑ 12*|
|RB201||Perception and Information Processing for Robotics||10||CS201|
|RB202||Robot Perception and Learning||10||RB201|
|RB203||Intelligent Systems and Robotics||10||RB202|
|ES||Electives Subjects (Choose 01 subject)||∑ 3*|
|ES201||Bayesian Machine Learning||3||None|
|ES202||Deep Reinforcement Learning||3||None|
|ML201||Statistical Learning Models for Text and Graph Data||3||None|
|DL204||Big Data Processing||3||None|
|DL205||Computational Data Analysis||3||None|
|FS||Final Subjects||∑ 24|
|FS202||Thesis Research in Artificial Intelligence||24||All|
|RS||Required Subjects||∑ 18|
|RS201||Indexed Scopus/SCI/IEEE/ISSN Journal Papers||12||All|
|RS202||Tranformation and Implementation of Research Paper||6||RS201|
To be considered for entry into DSc in AI Science Program please submit the following:
Doctoral accepted students must enroll (confirm and register) as per the instruction given. Online program without assistantship or self-paced students will not need to pay the school fees.
You can enter this program with a background in Computer Science or Information Technology, and the program will be tailored to build your skills in the other discipline.
Beyond the core subjects, elective subjects give you the freedom to dive deeper into a specialist area of data science.
Core subjects will give you a solid grounding in data science, so you can choose the business domain with a major data science project to feature in your field.
In the capstone project, you can apply data science tools to a practical problem by working individually or as part of a team to showcase your major skills.
To be graduated and got the Doctor of Computer Science in AI Science degree, YOU MUST BE
Once you submit a complete all required materials are received, your file will be sent to be reviewed by the Doctoral Committee. Material documents are reviewed by the Doctoral Committee after the deadline is passed. The GRITEx will indicate when your program is complete and make a plan to certify DSc in AI Science degree to you.
If you have understood program structure and graduation requirements, please click on Apply Program Now! at the top of our home page and create an account then following instructions.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510941.58/warc/CC-MAIN-20231001205332-20231001235332-00444.warc.gz
|
CC-MAIN-2023-40
| 4,508
| 51
|
https://db-blog.web.cern.ch/blog/franck-pachot/2018-09-oracle-cloud-infrastructure-api-keys-and-ocid
|
code
|
Oracle Cloud Infrastructure API Keys and OCID
As you may have read in the news, CERN is testing some Oracle Cloud services. When a large organisation is using the Cloud Credits, there's a need to control the service resources. This requires automation and then the GUI interface from the Cloud portal is not sufficient. We can control the Oracle Cloud Infrastructure through the REST API, OCI CLI, OCI SDKs, and all those methods require a RSA key for sign-in and some OCI (Oracle Cloud Identifier) to identify the user, the tenant, the compartment, the service,... We also need to know the URL for our Region (the endpoint) and the version of the API. In order to explain how it works, here is a demo I've run on my own OCI tenant (thanks to the Oracle ACE program providing credits to ACE Directors). I've run this example on Windows 10 WSL (the Windows Linux Subsystem which is actually a Ubuntu distribution) in order to check if all components are available from there.
Generate RSA key
I create a directory to store my private key:
root@Franck:~# mkdir ~/.oci root@Franck:~# openssl genpkey -out ~/.oci/oci_api_key.pem -algorithm RSA -pkeyopt rsa_keygen_bits:2048 ..........+++ ............................+++
I have generated a 2048 bits key in PEM format as this is what is required by the Oracle Cloud. You will not be able to upload a smaller key.
Here it is:
root@Franck:~# ls -l ~/.oci/oci_api_key.pem -rw------- 1 root root 1704 Sep 15 19:20 ~/.oci/oci_api_key.pem root@Franck:~# cat ~/.oci/oci_api_key.pem -----BEGIN PRIVATE KEY----- MIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDh+pxDWZ1QYxet 0KMyBHv+99EAvGzDS8KkyH5HiWF7n8U6wx9+ICgIiCaEVZVpvfsxmrM+wm/1kalr rZ+4u70n0m0u5+d474+w4r3h0u53+cl0ud+53rv1c3+l683kpNKIJqldx201iX01 ay++n0+hum4n+l4b0r+h3r3+vTKNAgMBAAECfwNrdmiqT9R7RqTwwzhiIU03kNrR cy+b3c4u53+7h3+db4+m4k35+m1574k35+y0u+kn0w+fBmpaaep6cU2DLpYCele2 l6jEUSOKb4mwWlBIu66+K3OSOpEjYMnmmfEcaqXQZUAMeQPhm0Lt0rI58WWdwAYP eUajiUoEzWOknGLbuOzOcm5ittfky9iT7o+TKKaYqVUu2/H6QqIxCFVFjNuChwYD qdpf5SrLAgMBAAECggEBANK/pnwpV01hHPOEGDQjSZsuY1BL0a+hH/caIF/3RtK8 ywE3/NAGJYHJvQhtG4z3NidbnEk1aKyIGAKtihn0Kv20ofbRkivRLoqbTpfMC/dg PQtJwKg9xdk39yBE+fQwy841aYQg813I9n3O+/db/4R/uhjdbGJVCaz4lGvbYJyv MPJ2GcVTpGtUV/S5I5X9a3pnFPc3z1Vj7wd0+2zfViieCW4vfgthd3IAvzRt+xBA x1J4Yz+OcDNTe15m9yQSJuUxlPWWdWk4K9KUW6k4LZnKCpZersDsI0FPJZOpzP1X ZwFiuJD9m4xUfHf9EY5qrHtD2tYQOhM0OroLrSBLkAECgYEA9ZEdaJceu1t7hN1h aKmu/TwEWKjXXP56SaW0Df+wwn+If7k4WvjfrAr9TZ4J940rioKqRrHehjvoy2ZH /iDpSyYEPnG2ck3iDdJL45d4++ECgYEA65RzD8r7yVZbiS4Yj1/fwf+rYNT4Zy9H DJMO30AXYuK+FXDc4X8sr4TWjb1Q1i0lEKq0vCRf9gXcfqBrR5IB/mnzYlI9rtJw LLddTvoEAKE1u/aHqtyNerNWb5sG8jJHrnanRM8RHCR65cfxp6vLSb/BhFtMBd++ G+tgNmXOXCsCgYBeJ43HFGpMG/94Kgm3P5hAsEpaVCJ/HXN12Xr4Axma+Zs0nafl +ZaJOI12vSnFKPOQ4Bi5kmlnvoC7vDO841XCMmjIS/KkTG91FORpm0GCg0YLNGwb dG+/lJnw2hPTXL52gWG7lw5xLsxRVUx629OXZLQr3R2/Je2Orxs54FIbAQKBgFMX zW3ont7pQugkBi2sF3SKBZzmAagDmP/qbvh31sfeSkWtxXsULGLMbBEakYS577Tt QUqLWf2hG6DeXHimG2XtYRdxAqjDOSXMl3Df5sGCGz3tOvUdThnVX8uF1vFeosH6 wA0ZLGPMyjx6RdZqVg3xQQS6XxpDTP70Cr+lMSB5AoGBAIMbq3zYB/l2GKZ3KOS2 c1DR1kjObGROK4/s3V5nsLqXMUZSfV7kqgv4s6hWzy3I9xpQfcb7rZk7A/Rj/lc/ aDOmN5kUoVLwwqhsbF3D9D0RPT5rD2y4EDR3DyxjYJq4n3KRdhi6kB1HM9Jke9yF hCJhTCoQbxjwX7LGJA5Umdi/ -----END PRIVATE KEY-----
Be careful with it: anyone with it will be able to access to your cloud service, here I obfuscated mine of course. There was an excellent presentation at POUG about Provilege Escalation in the Cloud from Pieter Van Puymbroek (@vanpupi) where he explains that showing this private key is like giving the key of your datacenter to someone with a laptop and a serial cable. Your cloud services have password authentication for login, but they have also a console.
From the generated private key I generate the public key:
root@Franck:~# openssl rsa -pubout -in ~/.oci/oci_api_key.pem -out ~/.oci/oci_api_key_public.pem writing RSA key root@Franck:~# cat ~/.oci/oci_api_key_public.pem -----BEGIN PUBLIC KEY----- MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4fqcQ1mdUGMXrQn3qFMX DAd/3EbpMhKPyHx6v370yIaxLmMszM/5EEganULl+RvkP9+UZJkBNuOat6VywyR6 LW9g8+KM/XeqmraoBXkLzyEw+EEwdfDp2NqKJ2OPpy6klh7K5YFwUCKqP43Hdx7J /Mezb+F2e5U6LZSDPD8MB5RMg0gk/xxNM41Z8dBjDsn+utNw+SyLIvq4g7OoxFEj im+JsFpQSLuuvitzkjqRI2DJ5pnxHGql0GVADHkD4ZtC7dKyOfFlncAGDw1Go4lK BM1jpJxi27jsznJuYrbX5MvYk+6PkyimmKlVLtvx+kKiMQhVRYzbgocGA6naX+Uq ywIDAQAB -----END PUBLIC KEY-----
This one can be shared publicly. The only thing you can do with it is to give me access to your system.
Upload public key
Before being able to do everything remotely from command line, I have to upload the key and get a fey identifiers from the Web portal. I'm going to my OCI console, from one of the regions I have access to (here EU-Frankfurt-1): https://console.eu-frankfurt-1.oraclecloud.com/a/identity/users
In the SSO login screen, I verify, or change, the "cloud tenant". The tenancy identifies yourself, or your company, which subscribed to the Oracle Cloud. Here is my login screen (where the tenant is my name):
I open the menu on top left, down to "Governance and Administration", open the "Identity" and "Users". The direct URL to it is: https://console.eu-frankfurt-1.oraclecloud.com/a/identity/users
From this list of users, I click on the user I'll use to sign-in through the REST API. In the Resources, I go to "API Keys" and "Add Public Key". This is where I paste the public key copied from above. This allows me to connect to my cloud service, as long as I have read access to the private key which was used to generate this public key.
Here is my user and list of keys:
I'll need the OCID for my user. Here it is fully displayed after I clicked on "Show". I click on "Copy" to get it in the clipboard. I set it as an environment variable in the shell I'll use later for REST API calls:
And the REST API sign-on will also identify the key I've uploaded, with its fingerprint, which I also set in my environment:
The key is identified by its fingerprint, displayed once I've added the public key:
With the authUserId and keyFingerprint I can identify myself as long as I also have my private key. I set the path to this private key also in my environment:
The users, as well as the services, are all related to a Tenant (the one I selected in the SSO login page) and I'll need to identify it with an OCID.
I open the menu on top left, down to "Governance and Administration", open the "Administration" and then "Tenancy Details". I can also get there when choosing a Region on the top-right because the regions I can access are related to my tenant. The direct url is: https://console.eu-frankfurt-1.oraclecloud.com/a/tenancy/regions
Here again, you can "Show" and "Copy" the OCID. Without clicking on "Show", only the last digits are displayed:
Again, I set it as an environement variable for later usage:
One more identifier is required. All services are segregated into compartments within the tenant, for access security reasons. By default the services will be created in the tenant root, which is the default compartment. This is not a good idea for security reasons: you must disign your compartments to isolate the resources. However, for this test, I'll use the tenant root compartment and you can see that its OCID is the same as the tenant OCID:
I'm setting this information in my environment and now I have everything to be able to connect to the services in this container:
Testing with OCI-CURL
In order to quickly test if everything works well, I'll use the oci-curl function explained by Yasin Baskan, Product Manager of Autonomous Data Warehouse cloud service in https://blogs.oracle.com/datawarehousing/managing-autonomous-data-warehouse-using-oci-curl where the idea is to use the sample function from the documentation. This script example creates a bash function, named "oci-curl", which is doing all the sign-in steps (build the headers that must change every 5 minutes). The example is referenced in the documentation for OCI sign-in and is fully available at https://docs.cloud.oracle.com/iaas/Content/Resources/Assets/signing_sample_bash.txt.
What I am doing here is get it from internet (with curl), remove the UTF-8 Byte Order Mark "M-oM-;M-?" from first line (which I got when doing that from Windows WSL), and remove the lines which define the tenant and user OCIDs and other variables for private key location and public key fingerprint. When setting the environment variables above, I used the same names on purpose, so that I just have to remove the local declaration from the example.
Here, I directly source the output into my current bash process:
source <( \ curl https://docs.cloud.oracle.com/iaas/Content/Resources/Assets/signing_sample_bash.txt \ | grep -vE "(local tenancyId|local authUserId=|local keyFingerprint|local privateKeyPath=)" \ | sed -e '1s/^.*#/#/' )
Of course, this is for demo purpose only, on a lab. You should never run a script directly downloaded from internet without looking at it. Even if you trust oracle.com, are you sure that the DNS entry is correct?
This has created an oci-curl() function in my current shell which takes the endpoint, the method, and the request as arguments. The result is JSON, and then I pipe the result to JQ to have it nicely displayed. Note that within the Windows WSL, which runs Ubuntu 18.04, I had to install it with "
add-apt-repository universe ; apt update ;
apt install jq"
The next blog post will detail how to use the REST API with this oci-cutl() function. For the moment, I'll just list the services from my Autonomous Data Warehouse (/autonomousDataWarehouses) and Autonomous Transaction Processing (/autonomousDatabases) services (using the Compartment OCID I've set previously):
oci-curl database.us-ashburn-1.oraclecloud.com get "/20160918/autonomousDataWarehouses?compartmentId=$compartmentId" | jq oci-curl database.us-ashburn-1.oraclecloud.com get "/20160918/autonomousDatabases?compartmentId=$compartmentId" | jq
The result, which validates that all the sign-in stuff is ok:
My services are named ADWC and ATP and are both stopped here. All information is available. The next post will show how to automate the start/stop of the services.
The comments are closed on this post, but do not hesitate to give feedback, comment or questions on Twitter (@FranckPachot)
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439735882.86/warc/CC-MAIN-20200804191142-20200804221142-00232.warc.gz
|
CC-MAIN-2020-34
| 10,247
| 42
|
http://theautumngroupllc.com/job-details/?id=2055
|
code
|
Sr .Net Developer
Location: Boston, Massachusetts
Sr. .Net Developer
· Analyze, design, develop, test and deploy phases of development projects.
· Design database schemas, conceptual data models, UI workflows and application architectures that fit into the agency enterprise architecture.
· Write optimal SQL queries and store-procedures to ensure good performance for database operations
· Apply Object-Oriented Design and Programming to build component-based middle tier RESTful web services with high throughput.
· Understand User-Centered Design concept to build user interfaces that are modern and provide good user experiences
· Gather information and feedback from end users to understand and develop project requirements.
· Conduct project design sessions and design solutions to meet current project requirements and be flexible enough to accommodate futures project needs.
· Performs reviews and integration testing to assure quality of project development efforts.
· Ensure project tasks are assigned and completed in a timely manner and project milestone dates are met.
· Ensure all project documentation is created, archived, and updated, including design, development and deployment documentation.
· Help control project scope.
· Coordinate activities of development team.
· Ensure all team members are informed of project activities.
· Learn new technologies if requested.
· Exhibit excellent verbal and communication skills.
· Minimum BS/CS with 10+ years of hands-on experience.
· Minimum of 2 years of leading small projects and team of 2-3 developers.
· Minimum of 4 years of .NET Web API web services.
· Minimum of 3 years of Front End development with Angular 2
· Proven experience and extensive knowledge in design, development and implementation of high throughput and large-scale RESTful web services using C# and .NET Web API framework.
· Strong experience with data modeling and problem solving skills.
· Very strong in Object Oriented Design and Programming (OOP & OOD).
· Very strong with C#, .NET and ASP.NET framework.
· Expertise in relational database design and building database stored procedures using T/SQL or PL-SQL.
· Hands-on experience with AWS or Microsoft Azure cloud services.
· Excellent written and oral communication skills
For More jobs like this visit: www.theautumngroupllc.com
The Autumn Group
7 Wall Street Windham, NH 03087
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202303.66/warc/CC-MAIN-20190320064940-20190320090940-00224.warc.gz
|
CC-MAIN-2019-13
| 2,401
| 32
|
https://www.linotp.org/doc/latest/part-management/linotp-config.html
|
code
|
1.21. The LinOTP configuration#
LinOTP can be configured via parameters specified in configuration files, or also via the process environment for increased convenience in container-based setups (e.g., Docker). The configuration files and environment variables can only be changed at the OS level by a user with appropriate access and not by a LinOTP token administrator.
1.21.1. Configuration parameters#
To see the current configuration of your LinOTP instance, run:
linotp config show
This will show you all configuration parameters supported by LinOTP. Parameters that do not occur explicitly in the current configuration are listed with their hard-coded default values.
For more details on any configuration entry, say,
linotp config explain ACTIVE_SECURITY_MODULE
A detailed explanation of the
linotp config commands can be found in the linotp-config(1) man page.
You need to restart LinOTP every time you make changes to the configuration, as the configuration settings are read only once, on startup.
1.21.2. Setting the configuration parameters in a configuration file#
Configuration settings can be loaded from multiple locations. These locations are determined by the value of the
LINOTP_CFG environment variable, which may contain a colon-separated list of directory names and file glob expressions. LinOTP checks these names and file glob expressions in the order they occur in
LINOTP_CFG, and reads any matching files. File names resulting from the expansion of a glob expression are processed in lexicographic order. Directory names have
*.cfg appended to them implicitly. Relative directory names and file glob expressions are taken relative to the application root directory (where the
app.py file is found).
LINOTP_CFG is not set and a file called
linotp-cfg-default exists in the application root directory, the content of that file is used as the default value for
LINOTP_CFG (this feature is generally used to establish a distribution-specific search order for configuration files). If the file does not exist,
linotp.cfg is assumed as the default value.
The following table lists configuration file locations for Debian packages (as per
Distribution defaults - these settings are supplied by the packaging and represent settings that are appropriate for the distribution and packaging format used.
Central configuration settings.
Configuration snippets can be placed in this
directory. The Debian packaging places the
database configuration in
The order of the locations in the above table matters. Settings in later files in the search order will override
settings previously found. For example, if
LOGFILE_DIR is defined in the following places:
LOGFILE_DIR=/srv/linotp-logsin the file
LOGFILE_DIR=/var/log/linotpin the file
then the value that is used is
/srv/linotp-logs, because that was the last one encountered.
Configuration settings whose names end in
_DIR contain file or directory names, respectively.
These names can be either absolute (starting with a slash) or relative, in which case they are evaluated
starting from the directory named by the
ROOT_DIR configuration setting.
This setting must contain an absolute directory name.
To ensure that relative paths are correctly resolved, the respective settings are only evaluated after the
entire configuration has been read and the definitive value for
ROOT_DIR has been determined.
This means that if the
/usr/share/linotp/linotp.cfg file contains a setting like
DATA_DIR = "data" and
the very last line of
/etc/linotp/linotp.cfg, which is read after
ROOT_DIR = /var/lib/my-special-linotp, then the effective value of
DATA_DIR will be
/var/lib/my-special-linotp/data, even though the
ROOT_DIR setting was seen after the
1.21.3. Setting the configuration parameters via environment variables#
For the benefit of people who run LinOTP in container-based environments such as Docker, all of the LinOTP parameters
can also be defined in environment variables. However, the parameter names must be prefixed with
example, to set the value for
ACTIVE_SECURITY_MODULE, you must assign it to an environment variable called
Parameters defined in environment variables override the respective parameter value assigned in configuration files.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296816863.40/warc/CC-MAIN-20240414002233-20240414032233-00135.warc.gz
|
CC-MAIN-2024-18
| 4,215
| 58
|
https://forum.armbian.com/profile/10471-barna/
|
code
|
Thanks for the information in this thread, it worked....then after about 2 weeks 24x7 usage, I had a crash on my Orange Pi One (not related to USB functionality I don't think, the system was probably doing a database import at the time). After this, the OTG port didn't seem to work any more in host mode. I tracked it down to not being any internal power on the the port, the USB data lines were OK. Examining the board schematic shows the power for the OTG port comes through 'U11' which in turn is controlled by a line from the H3 chip, but as my board didn't have U11 on the silkscreen I didn't get much further than that. I assumed I'd blown up that part of the board as nothing seemed to fix it and the rest of the system continued to work flawlessly after a reboot. That was not the case however, as a complete reinstall (Armbian stretch) from scratch and hey presto, the internal USB power is restored. In case it happens again, anyone have any idea how to troubleshoot powering of the OTG port in USB host mode when/if it goes wrong?
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585371883359.91/warc/CC-MAIN-20200410012405-20200410042905-00486.warc.gz
|
CC-MAIN-2020-16
| 1,042
| 1
|
https://www.scottishjobsonthe.net/cheap-web-hosting-with-site-building-tools/
|
code
|
Cheap Web Hosting With Site Building Tools
Locating a premium inexpensive web hosting carrier isn’t simple. Every site will have different needs from a host. Plus, you have to compare all the attributes of an organizing firm, all while looking for the best bargain possible.
This can be a lot to sort via, specifically if this is your first time purchasing hosting, or building a website.
A lot of hosts will use extremely inexpensive initial pricing, just to raise those rates 2 or 3 times greater once your preliminary contact is up. Some hosts will give free rewards when you join, such as a totally free domain name, or a cost-free SSL certificate.
While some hosts will be able to provide far better performance and high degrees of safety. Cheap Web Hosting With Site Building Tools
Below we dive deep right into the best cheap host plan there. You’ll discover what core hosting features are essential in a host as well as how to analyze your very own organizing needs to make sure that you can select from among the most effective inexpensive organizing service providers listed below.
Disclosure: When you acquire a web hosting plan via links on this web page, we make some compensation. This assists us to keep this website running. There are no added prices to you in any way by using our links. The listed here is of the very best affordable web hosting packages that I have actually directly utilized and also checked.
What We Consider To Be Inexpensive Host When we describe a web hosting package as being “Low-cost” or “Spending plan” what we indicate is hosting that falls into the cost brace in between $0.80 to $4 monthly. Whilst investigating economical organizing providers for this overview, we checked out over 100 different hosts that came under that price range. We then examined the high quality of their most affordable holding plan, value for money and client service.
In this post, I’ll be reviewing this world-class internet site holding business as well as stick in as much pertinent details as feasible.
I’ll go over the functions, the prices choices, and also anything else I can consider that I believe could be of advantage, if you’re determining to join to Bluhost as well as get your websites up and running.
So without more trouble, allow’s check it out.
Bluehost is among the biggest host companies worldwide, obtaining both huge advertising and marketing assistance from the business itself and also associate marketing experts who promote it.
It really is an enormous business, that has been around for a long time, has a huge track record, as well as is most definitely among the leading choices when it concerns webhosting (most definitely within the leading 3, at the very least in my publication).
But what is it specifically, and should you get its solutions?
Today, I will address all there is you need to know, offered that you are a blog writer or a business owner that is trying to find a webhosting, and also doesn’t recognize where to begin, considering that it’s an excellent solution for that audience generally.
Let’s picture, you want to hold your sites and make them visible. Okay?
You currently have your domain name (which is your site destination or LINK) and now you intend to “transform the lights on”. Cheap Web Hosting With Site Building Tools
You require some hosting…
To achieve every one of this, as well as to make your web site noticeable, you require what is called a “web server”. A web server is a black box, or tool, that stores all your web site data (documents such as images, messages, videos, links, plugins, and various other information).
Currently, this server, needs to be on at all times and also it needs to be attached to the internet 100% of the time (I’ll be discussing something called “downtime” in the future).
In addition, it also needs (without getting also expensive as well as right into details) a file transfer protocol generally called FTP, so it can reveal web browsers your web site in its desired kind.
All these points are either expensive, or call for a high level of technological skill (or both), to develop and also preserve. And also you can entirely go out there and also learn these points on your own as well as established them up … yet what about instead of you buying and keeping one … why not simply “renting organizing” instead?
“This is where Bluehost is available in. You lease their servers (called Shared Hosting) and also you launch a site making use of those servers.”
Considering that Bluehost maintains all your documents, the business likewise enables you to set up your content monitoring systems (CMS, for brief) such as WordPress for you. WordPress is a very preferred CMS … so it just makes good sense to have that choice offered (almost every hosting firm currently has this choice as well).
In short, you no longer require to set-up a web server and then integrate a software application where you can develop your web content, individually. It is currently rolled into one plan.
Well … envision if your web server is in your house. If anything were to take place to it in any way, all your files are gone. If something fails with its inner procedures, you need a technician to repair it. If something overheats, or breaks down or obtains damaged … that’s no good!
Bluehost takes all these troubles away, and also looks after every little thing technical: Pay your web server “rental fee”, and they will certainly look after everything. And once you purchase the solution, you can then start concentrating on including web content to your site, or you can place your effort right into your marketing projects.
What Solutions Do You Receive From Bluehost?
Bluehost uses a myriad of various services, yet the key one is hosting obviously.
The holding itself, is of various types incidentally. You can lease a common server, have a committed web server, or additionally a virtual exclusive web server.
For the function of this Bluehost testimonial, we will certainly focus on organizing services and also various other solutions, that a blog writer or an on-line business owner would need, as opposed to go unfathomable into the bunny opening as well as discuss the other solutions, that are targeted at even more knowledgeable individuals.
- WordPress, WordPress PRO, and ecommerce— these organizing solutions are the packages that permit you to host a website utilizing WordPress as well as WooCommerce (the latter of which allows you to do e-commerce). After purchasing any one of these plans, you can start building your internet site with WordPress as your CMS.
- Domain name Industry— you can additionally purchase your domain name from Bluehost as opposed to various other domain registrars. Doing so will certainly make it much easier to aim your domain name to your host’s name servers, because you’re utilizing the same marketplace.
- Email— as soon as you have actually purchased your domain name, it makes sense to also obtain an e-mail address linked to it. As a blogger or on-line business owner, you need to practically never ever utilize a totally free email solution, like Yahoo! or Gmail. An email similar to this makes you look unprofessional. Thankfully, Bluehost gives you one totally free with your domain.
Bluehost likewise offers dedicated web servers.
And also you may be asking …” What is a dedicated server anyhow?”.
Well, the thing is, the standard host bundles of Bluehost can just a lot web traffic for your internet site, after which you’ll require to update your holding. The reason being is that the usual servers, are shared.
What this suggests is that one web server can be servicing 2 or more sites, at the same time, among which can be your own.
What Does This Suppose For You?
It implies that the solitary server’s sources are shared, and it is doing multiple jobs at any kind of provided time. Once your site starts to hit 100,000 website check outs monthly, you are going to need a committed server which you can additionally obtain from Bluehost for a minimum of $79.99 per month.
This is not something yous must fret about when you’re starting however you must keep it in mind for sure.
Bluehost Pricing: How Much Does It Expense?
In this Bluehost evaluation, I’ll be concentrating my attention primarily on the Bluehost WordPress Hosting plans, considering that it’s the most prominent one, and very likely the one that you’re looking for which will certainly suit you the best (unless you’re a huge brand name, company or website).
The three available strategies, are as complies with:
- Basic Plan– $2.95 each month/ $7.99 regular price
- Plus Strategy– $5.45 monthly/ $10.99 normal cost
- Option And Also Plan– $5.45 each month/ $14.99 regular price
The very first cost you see is the rate you pay upon subscribe, and the second rate is what the price is, after the first year of being with the firm.
So primarily, Bluehost is mosting likely to charge you on a yearly basis. And also you can also select the quantity of years you wish to hold your site on them with. Cheap Web Hosting With Site Building Tools
If you choose the Basic plan, you will pay $2.95 x 12 = $35.40 beginning today and also by the time you enter your 13th month, you will now pay $7.99 each month, which is also charged each year. If that makes any feeling.
If you are serious about your website, you ought to 100% obtain the three-year choice. This implies that for the fundamental strategy, you will certainly pay $2.95 x 36 months = $106.2.
By the time you hit your fourth year, that is the only time you will certainly pay $7.99 each month. If you think of it, this technique will certainly conserve you $120 in the course of 3 years. It’s very little, however it’s still something.
If you wish to obtain greater than one internet site (which I very suggest, as well as if you’re major, you’ll most likely be getting even more at some time in time) you’ll wish to take advantage of the choice plus strategy. It’ll enable you to host limitless web sites.
What Does Each Strategy Offer?
So, when it comes to WordPress holding plans (which are similar to the common holding plans, yet are extra geared in the direction of WordPress, which is what we’ll be focusing on) the attributes are as follows:
For the Standard strategy, you obtain:
- One internet site only
- Protected site by means of SSL certification
- Optimum of 50GB of storage space
- Free domain for a year
- $ 200 marketing credit
Remember that the domains are purchased independently from the holding. You can get a totally free domain name with Bluehost below.
For both the Bluehost Plus hosting as well as Choice Plus, you get the following:
- Unrestricted variety of internet sites
- Free SSL Certificate. Cheap Web Hosting With Site Building Tools
- No storage or data transfer limitation
- Complimentary domain name for one year
- $ 200 marketing credit rating
- 1 Workplace 365 Mailbox that is complimentary for one month
The Choice Plus plan has an included benefit of Code Guard Basic Alternative, a back-up system where your file is saved and also replicated. If any crash takes place as well as your web site data goes away, you can restore it to its original kind with this function.
Notice that although both plans set you back the exact same, the Selection Strategy after that defaults to $14.99 each month, regular price, after the set amount of years you’ve selected.
What Are The Conveniences Of Using Bluehost
So, why select Bluehost over other webhosting solutions? There are numerous host, many of which are resellers, but Bluehost is one choose couple of that have actually stood the test of time, and it’s probably one of the most popular available (and forever reasons).
Here are the 3 primary benefits of picking Bluehost as your web hosting company:
- Web server uptime— your site will certainly not be visible if your host is down; Bluehost has more than 99% uptime. This is incredibly essential when it concerns Google Search Engine Optimization as well as positions. The higher the better.
- Bluehost speed— just how your server response identifies how quick your site reveals on a browser; Bluehost is lighting fast, which implies you will certainly decrease your bounce rate. Albeit not the most effective when it involves packing speed it’s still extremely crucial to have a quick speed, to make user experience far better as well as better your ranking.
- Unlimited storage— if you obtain the And also plan, you need not stress over how many data you store such as videos– your storage ability is limitless. This is actually essential, because you’ll most likely face some storage space concerns in the future down the tracks, as well as you don’t desire this to be a hassle … ever before.
Finally, consumer support is 24/7, which suggests despite where you are in the world, you can get in touch with the support group to repair your site problems. Pretty basic nowadays, but we’re taking this for approved … it’s additionally very vital. Cheap Web Hosting With Site Building Tools
Likewise, if you have actually gotten a complimentary domain with them, then there will certainly be a $15.99 fee that will be subtracted from the quantity you originally purchased (I imagine this is because it type of takes the “domain name out of the market”, uncertain concerning this, but there possibly is a hard-cost for registering it).
Lastly, any kind of demands after 30 days for a reimbursement … are void (although in all honesty … they need to most likely be stringent right here).
So as you see, this isn’t necessarily a “no doubt asked” policy, like with some of the other organizing alternatives available, so be sure you’re okay with the policies before proceeding with the organizing.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573029.81/warc/CC-MAIN-20220817153027-20220817183027-00442.warc.gz
|
CC-MAIN-2022-33
| 13,883
| 81
|
https://support.symantec.com/en_US/article.HOWTO81236.html
|
code
|
The Application and Device Control policy has two parts:
Application Control contains one or more rule sets. Each rule set contains one or more rules. You configure properties, conditions and actions for each rule.
Device control consists of a list of blocked devices and a list of devices that are excluded from blocking. You can add to these two lists and manage their contents.
Figure: Application and Device Control policy structure illustrates the application and device control components and how they relate to each other.
|
s3://commoncrawl/crawl-data/CC-MAIN-2016-40/segments/1474738662018.69/warc/CC-MAIN-20160924173742-00282-ip-10-143-35-109.ec2.internal.warc.gz
|
CC-MAIN-2016-40
| 529
| 4
|
http://www.pearltrees.com/voidhaze
|
code
|
Joscha Bach: Artificial Consciousness and the Nature of Reality | AI Podcast #101 with Lex Fridman. Deep learning to translate between programming languages. Migrating a codebase from an archaic programming language such as COBOL to a modern alternative like Java or C++ is a difficult, resource-intensive task that requires expertise in both the source and target languages. COBOL, for example, is still widely used today in mainframe systems around the world, so companies, governments, and others often must choose whether to manually translate their code bases or commit to maintaining code written in a language that dates back to the 1950s. We’ve developed TransCoder, an entirely self-supervised neural transcompiler system that can make code migration far easier and more efficient. Our method is the first AI system able to translate code from one programming language to another without requiring parallel data for training.
We’ve demonstrated that TransCoder can successfully translate functions between C++, Java, and Python 3. Self-supervised training is particularly important for translating between programming languages. Something Went Wrong. Survival Library | Survival Library.
Open Cobalt. Open Cobalt is a free and open source software platform for constructing, accessing, and sharing virtual worlds both on local area networks or across the Internet, without any requirement for centralized servers. The technology makes it easy to create deeply collaborative and hyperlinked multi-user virtual workspaces, virtual exhibit spaces, and game-based learning and training environments that run on all major software operating systems.
By using a peer-based messaging protocol to reduce reliance on server infrastructures for support of basic in world interactions across many participants, Open Cobalt makes it possible for people to hyperlink their virtual worlds via 3D portals to form a large distributed network of interconnected collaboration spaces. Open Cobalt uses the Squeak software environment, which is an open source Smalltalk system freely available for Windows, Mac and Unix. Expected uses History In early 2008, and with the support of the Andrew W. Goals Parliament-Funkadelic. By the early 1980s, Clinton consolidated the collective's multiple projects and continued touring under the names "George Clinton and the P-Funk All-Stars" or "George Clinton and Parliament-Funkadelic. " Some former members of Parliament perform under the name "Original P".
Sixteen members of Parliament-Funkadelic were inducted to the Rock and Roll Hall of Fame in 1997. The etymology of the term "P-Funk" is subject to multiple interpretations. It has been identified typically as an abbreviation of "Parliament-Funkadelic. " Another suggested definition is "Plainfield Funk", referring to Plainfield, New Jersey, the hometown of the band's original line-up. History The Parliaments For more details on this phase of the collective's career, see The Parliaments. Funkadelic and Parliament Bootsy Collins In the 1975-1979 period, both Parliament and Funkadelic achieved several high-charting albums and singles on both the R&B and Pop charts. Modern day Parliament-Funkadelic Herbie Hancock - Chameleon (FULL VERSION)
The Prodigy -Warriors dance-hq-full version. All sizes | The Twitterverse 2.0. Tau Music - piano solo. Distributed hash table. Distributed hash tables History These systems differed in how they found the data their peers contained: Napster, the first large-scale P2P content delivery system to exist, had a central index server: each node, upon joining, would send a list of locally held files to the server, which would perform searches and refer the querier to the nodes that held the results.
This central component left the system vulnerable to attacks and lawsuits. Gnutella and similar networks moved to a flooding query model – in essence, each search would result in a message being broadcast to every other machine in the network. While avoiding a single point of failure, this method was significantly less efficient than Napster. Distributed hash tables use a more structured key-based routing in order to attain both the decentralization of Freenet and gnutella, and the efficiency and guaranteed results of Napster.
Properties DHTs characteristically emphasize the following properties: and. Unreal Tournament - Epic Wiki. Anonym surfen | link to other sites anonymously. In order to produce a single anonymous link, enter the URL you want to link to and then click on "Generate URL". If you want to anonymize all the external links on your board or homepage, we can generate a script for you to deal with this automatically for all your pages. Enter the Sites for which links shall not be redirected to anonym.to (e.g. your own) and click on "Generate script". The advantages of anonymizing your external links with anonym.to Webmasters can use this tool to prevent their site from appearing in the server logs of referred pages as referrer. The operators of the referred pages cannot see where their visitors come from any more.
Detailed instructions for the anonymizing script Once the script is embedded in a website, it redirects all the links via anonym.to - except for the sites that were excluded when generating the script. Anonym.to only disguises what page a visitor comes from. Got SPAM? I don’t, thanks to GFI Max MailEdge | retrohack. Security is all about adding on layers, like an onion, or an ogre, to establish what is called a defence in depth approach. The more layers of security a bad guy must penetrate, the less likely he will succeed, the more likely you will detect him, and unless you are the specific target of the attack, they may just move on to easier pickings.
Email is often what I like to call a drive-by shooting victim…spam, phishing attacks, and malware are not so much targeted at your company as they are at your users inboxes. In is not personal; they just happened to leave their email.addr out somewhere that got noticed. Exchange systems have a lot of options for screening incoming mail. Forefront for Exchange can be implemented at the edge transport role, Forefront TMG has some layered protections, and even Outlook contributes to the goal of keeping users spam free. Setting it up Reduce the TTL on your MX records to a value that allows you to revert quickly; 300 works well. User Self Service Setup. Best Smooth Jazz (22nd March 2014) host Rod Lucas. Live Streams | radio reddit. Which infinity is larger: the number of planes or the number of points? : askscience. NoSQL. "Structured storage" redirects here. For the Microsoft technology also known as structured storage, see COM Structured Storage.
A NoSQL (often interpreted as Not Only SQL) database provides a mechanism for storage and retrieval of data that is modeled in means other than the tabular relations used in relational databases. Motivations for this approach include simplicity of design, horizontal scaling and finer control over availability. The data structure (e.g. key-value, graph, or document) differs from the RDBMS, and therefore some operations are faster in NoSQL and some in RDBMS. There are differences though, and the particular suitability of a given NoSQL DB depends on the problem it must solve (e.g. does the solution use graph algorithms?).
History There have been various approaches to classify NoSQL databases, each with different categories and subcategories. A more detailed classification is the following, by Stephen Yen: Performance Examples Graph Friend Within - The Trouble. Energy Courses. OCW Scholar. FAQ: OCW Scholar. Solar cell research. There are currently many research groups active in the field of photovoltaics in universities and research institutions around the world. This research can be divided into three areas: making current technology solar cells cheaper and/or more efficient to effectively compete with other energy sources; developing new technologies based on new solar cell architectural designs; and developing new materials to serve as light absorbers and charge carriers.
Silicon processing One way of reducing the cost is to develop cheaper methods of obtaining silicon that is sufficiently pure. Silicon is a very common element, but is normally bound in silica, or silica sand. Processing silica (SiO2) to produce silicon is a very high energy process - at current efficiencies, it takes one to two years for a conventional solar cell to generate as much energy as was used to make the silicon it contains. Nanocrystalline solar cells Thin-film processing Polymer processing MrSuicideSheep. Study Finds Random Electrical Current May Help Folks Learn Math. Mathemati-phobes rejoice. You may one day swap coffee and cramming for a trusty set of head-mounted electrodes. According to a team of researchers, hailing from the UK and Austria, non-invasive electrical brain stimulation may improve math learning.
However, the results, while intriguing, require a few requisite grains of salt: The trial was small, and as yet, there is no proven physiological explanation. The technique, called transcranial random noise stimulation (TRNS), sends weak electrical impulses racing between electrodes attached to the forehead. An early study using the technique suggested TRNS works by stimulating the brain’s ability to generate and propagate electrical potential, thus making it easier to forge new neural pathways. The five day experiment tested shallow cognitive processing—students memorizing their times tables, for example—and deep cognitive processing, or the ability to learn a mathematical principle and apply it to solve a novel problem. SuperheroYou | Boost Your Brainpower.
Movement. Code. Web development. Higher-Order Perl. PDF, Let Me Count the Ways… In this post, I show how basic features of the PDF language can be used to generate polymorphic variants of (malicious) PDF documents. If you code a PDF parser, write signatures (AV, IDS, …) or analyze (malicious) PDF documents, you should to be aware of these features. Official language specifications are interesting documents, I used to read them from front to back. I especially appreciate the inclusion of a formal language description, for example in Backus–Naur form. But nowadays, I don’t take the time to do this anymore. While browsing through the official PDF documentation, I took particular interest in the rules to express lexemes. Building a test file Before I show some examples, let’s build a test PDF file that will start the default browser and navigate to a site each time the document is opened. Opening a web page from a PDF file can be done with an URI action, like this: This is the same type of object used in the malicious mailto PDF files.
Name representation Or #55#52#49. Networks. In mathematical terms, a network is a graph in which the nodes and edges have values associated with them. A graph is defined as a pair of sets , where is a set of nodes (vertices or points within the graph) labelled and is a set of edges (links (vi, vj ) that connect pairs of elements vi, vj within ). The degree to which the nodes of a network are directly connected is called connectivity. Is the number of edges and is the number of nodes in the network, the following equation is used; The degree of a node in a network is the number of edges or connections to that node (Newman 2003).
Shortest average path length The average path length, ( ) of a network is the average number of edges, or connections between nodes, that must be crossed in the shortest path between any 2 nodes (Watts 2003). Where is the minimum distance between nodes i and j. The diameter of a network is the longest shortest path within a network. A common property of many social networks is cliques. Consisting of the set of nodes. Mind Map - Index. Visualizing Hash Functions. I've been working on some static visualizations of hash functions, starting with SHA1 and MD5 (images here and here , respectively). The two sample images show the process of computing the respective hash functions on the input "denny" (with ASCII encoding -- i.e., the input in hex is "64656E6E79"). A portion of the MD5 image is included below: Notes about these two particular algorithms and their visualizations: Both algorithms are mostly defined in terms of operations on 32-bit "words", which essentially means 32-bit integers.
They define the input/output conversion from sequences of bits/bytes to words, but this step is not covered in the diagrams. Probably the most confusing thing if you examine them in detail is that SHA1 uses a big-endian conversion from 4 bytes to 1 word while MD5 uses little-endian. See Appendix A for more. If you're curious about any other details of MD5 or SHA1, the specs are fairly readable, hopefully even more so with the diagrams to follow along with. Baking Pi - Operating Systems Development. This course has not yet been updated to work with the Raspberry Pi models B+ and A+. Some elements may not work, in particular the first few lessons about the LED.
It has also not been updated for Raspberry Pi v2. Welcome to Baking Pi: Operating Systems Development! Course by Alex Chadwick. You can now help contribute to this tutorial on GitHub. This website is here to guide you through the process of developing very basic operating systems on the Raspberry Pi! This course takes you through the basics of operating systems development in assembly code. Rather than leading the reader through the full details of creating an Operating System, these tutorials focus on achieving a few common tasks separately. 1 Requirements 1.1 Hardware In order to complete this course you will need a Raspberry Pi with an SD card and power supply. 1.2 Software In terms of software, you require a GNU compiler toolchain that targets ARMv6 processors. 2 Lessons.
MapReduce. Overview MapReduce is a framework for processing parallelizable problems across huge datasets using a large number of computers (nodes), collectively referred to as a cluster (if all nodes are on the same local network and use similar hardware) or a grid (if the nodes are shared across geographically and administratively distributed systems, and use more heterogenous hardware). Processing can occur on data stored either in a filesystem (unstructured) or in a database (structured).
MapReduce can take advantage of locality of data, processing it on or near the storage assets in order to reduce the distance over which it must be transmitted. "Map" step: Each worker node applies the "map()" function to the local data, and writes the output to a temporary storage. MapReduce allows for distributed processing of the map and reduction operations. Another way to look at MapReduce is as a 5-step parallel and distributed computation: Logical view Map(k1,v1) → list(k2,v2) Examples Design Patterns.
It has been highly influential to the field of software engineering and is regarded as an important source for object-oriented design theory and practice. More than 500,000 copies have been sold in English and in 13 other languages. The authors are often referred to as the Gang of Four (GoF). History Introduction, Chapter 1 Chapter 1 is a discussion of object-oriented design techniques, based on the authors' experience, which they believe would lead to good object-oriented software design, including: clients remain unaware of the specific types of objects they use, as long as the object adheres to the interfaceclients remain unaware of the classes that implement these objects; clients only know about the abstract class(es) defining the interface Use of an interface also leads to dynamic binding and polymorphism, which are central features of object-oriented programming. The authors admit that delegation and parameterization are very powerful but add a warning: Formatting
Copyright Does Not Exist. What is Bitcoin? - We Use Coins. Xin Che Jian: a Hackerspace in Shanghai. Recyclebot v2.2. RetroShare. Tor Project: Anonymity Online. Phreaking. About Darknet. CrackingForum - Cracking Downloads, Cracking Tutorials, Premium Accounts.
DCOM Server Process Launcher.
Wikiversity. Courses. Free Online Course Materials | MIT OpenCourseWare. Creative. Media. Derivative. Random. Stuff. Lance. Gaming. Dive Into HTML5. Prototype v1.7.1 API documentation | Event.observe. C# Tutorials (C#) Learn HTML5 - Game Development Course Online. Design Patterns Library. Nine-board Tic Tac Toe. Easylearntutorial. Hardware. Node types. Welcome to X3DOM — X3DOM 1.4.0 documentation. 0.1.1. X3D-Edit 3.3 Authoring Tool for Extensible 3D (X3D) Graphics. Getting Started with Three.js | Aerotwist. Bamboo & Organic Cotton Plain T shirt: Amazon.co.uk: Clothing. IEEE Spectrum: Technology, Engineering, and Science News.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320300849.28/warc/CC-MAIN-20220118122602-20220118152602-00653.warc.gz
|
CC-MAIN-2022-05
| 16,792
| 25
|
https://drupal.stackexchange.com/questions/235159/taxonomy-terms-list-and-features/235160
|
code
|
I noticed that, while I am able to build taxonomy into my feature, it doesn't seem to include the list of terms within the taxonomy. When I go to create a feature, the
uuid_term section does list a few, but not many, actual terms for me to select. What am I missing? It's worth mentioning that I am starting with a Demo build and customizing from there. The only terms that are actually listed are not showing up in Structure > Taxonomy.
I tried upgrading UUID to no avail.
Drupal 7.54 UUID 7.x-1.0 Features 7.x-2.10
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243988741.20/warc/CC-MAIN-20210506053729-20210506083729-00629.warc.gz
|
CC-MAIN-2021-21
| 516
| 4
|
https://www.taxpreparationaccountant.com/tax-updates/stock-appreciation-rights-sars
|
code
|
Stock appreciation rights, or SARs, enable employees to receive the benefit of an increase in value of the company stock between the date the SARs are granted and the date they are exercised. When the SARs are exercised, cash or stock may be delivered as payment for the post-grant appreciation.
For example, when your employer’s stock is worth $20 a share, you receive 100 SARs exercisable within five years. Two years later, when the stock price has increased to $30 a share, you exercise the SARs and receive $1,000 ($3,000 at exercise minus $2,000 value at grant).
If IRS tests are satisfied, there is no tax until you exercise the SARs and the post-grant appreciation is received.
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618038077843.17/warc/CC-MAIN-20210414155517-20210414185517-00192.warc.gz
|
CC-MAIN-2021-17
| 687
| 3
|
https://www.belovedfoundation.com/blog/posts/a-list-of-cancer-podcasts-for-the-podcast-lover
|
code
|
List of cancer podcasts
Rethink breast cancer compiled a list of their top 5 favourite podcasts on cancer. If you’re a podcast listener, you might want to add one or five of these to your playlist.
From all over the world, these podcasts feature a variety of topics including motherhood while living with cancer, grief support, humour, and more.
See the podcasts in this list.
The Top Five Cancer Podcasts on our Playlist
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370491998.11/warc/CC-MAIN-20200328134227-20200328164227-00459.warc.gz
|
CC-MAIN-2020-16
| 423
| 5
|
https://www.freelancer.com/projects/css/developer-implement-new-design-existing/?ngsw-bypass=&w=f
|
code
|
We have recently created a new design for http://examgenie.com. The design is already complete with an XD file created. There are about 5 pages on the site. Now, I need a developer to implement this new XD file on the site and also expand the existing payment gateway, which is currenly only PayPal, to include Cards payment. The site is writen in the MEAN stack. Only a developer able to work within an existing site/template should respond, and available to commit immediately and complete within two days. To summarise:
1. Take an XD File and implement on an existing site [login to view URL]
2. Expand existing paypal only payment gateway to include Cards
3. Fully Test.
Code is currently stored in Git
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400227524.63/warc/CC-MAIN-20200925150904-20200925180904-00320.warc.gz
|
CC-MAIN-2020-40
| 706
| 5
|
http://math.hws.edu/bridgeman/courses/343/f08/doc/violet.html
|
code
|
|CPSC 343||Database Theory and Practice||Fall 2008|
Violet is a simple, free, and attractive editor for UML diagrams. (Many editors are lacking in one or more of these qualities!) Since ER diagrams are a lot like UML diagrams except for the notation being different, I've added basic support for the ER diagrams you'll be drawing in this class.
into your shell. You should get a splash screen, followed shortly by either a rather Windows-like display allowing you to select a drawing type or a recent file, or the editor window itself.
If you get the Windows-like display, choose an existing file you want to open or select "ER Diagram" to create a new ER diagram.
There's not much to using VioletER:
The specialization node lets you configure two fields when you right-click - set only the kind and ignore the "ignoreThis" field.
For ER diagrams, there are three types of connectors:
Use the "Attribute connector" for connecting attributes to entities and relationships, and for connecting components of a composite attribute to the main attribute. It's just a basic line with no configurable properties.
The "E-R connector" is for connecting entities and relationships when you want to use the single/double line notation to indicate whether an entity's participation in a relationship is total. It also lets you specify a role and a maximum cardinality.
The "E-R connector [(min,max notation]" connector is for connecting entities and relationships when you want to use the (min,max) notation for showing cardinality. It also lets you specify a role.
The "Is subclass of" connector is for connecting an entity type to a superclass or one of the round specialization nodes. You need to drag from the subclass to the superclass/specialization node when you create the line. You can specify a defining predicate.
Note: There are (currently) no checks that your drawing follows legal syntax. If you want to use a plain connector to connect entities and relationships or a participation connector to connect attributes or any kind of connector to connect two entities, Violet won't stop you.
Note #2: Weak key attributes are (currently) denoted with the solid underline also used for (regular) key attributes, instead of a dashed underline. It should not be difficult to recognize weak key attributes, however, because they are attributes connected to weak entities.
I've found Violet to be pretty much unusable when run over the network (i.e. displayed on a machine other than the one where it is running). If you want to use it on your own computer:
You'll need Java 6 update 10 installed on your computer. (Run java -version to see what version you have - it should be jdk1.6.0_10.) Earlier versions of Java 6 can be used to create and save diagrams, but you won't be able to print. (download Java 6)
Copy /classes/f08/cs343/violet/violeter.jar and /classes/f08/cs343/violet/run-violeter to your own computer (via scp - ask if you don't know how to do this).
Edit run-violeter, changing /classes/f08/cs343/jdk1.6.0_10/bin/java to the path of your Java 6 installation, and changing /classes/f08/cs343/violet/violeter.jar to the path of where you put violeter.jar.
Make run-violeter executable:
chmod +x run-violeter
will do it if you are currently in the directory where you copied it.
There may be occasional updates to VioletER. An announcement will be posted if there's a new jar available.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376823702.46/warc/CC-MAIN-20181211194359-20181211215859-00504.warc.gz
|
CC-MAIN-2018-51
| 3,393
| 21
|
https://support.mocaplatform.com/hc/en-us/articles/115003237111-MOCA-console-Places
|
code
|
What are "Places"?
Places is our way of defining a container that can have (or not) a geofence configured, many attributes and which you can add Beacon Zones. We made them so they can represent locations in the real world (i.e a store, a museum...), but also they can contain zones within the location (i.e. Frozen Foods Aisle).
Places can have a unique name and a description for ease of usage:
Once you have defined the name and description, you can set up other attributes like Geofences and Zones.
Geofences are used to trigger our Proximity Campaigns. These can have a minimum radius of 100 meters and are mainly used to send contextual messages or tag specific users that enter, dwell or exit the area.
To create a geofence please click on the center of the circular area you'd like to mark and then drag to the desired radius. Save the geofence when you've finised so it updates in our systems.
Zones represent areas inside a place (i.e store entrance, shop, restaurant, bar, rest area and so on). You may assign one or more BLE Beacons to define a Zone. Zones enable you to deliver brand-focused proximity experiences and also provide fine-grained analytics.
A Place reaches it's full potential when it's placed over an area of maximum impact, and the campaign it triggers offers clear value to the user. To know which areas concentrate the most users go over to Analytics > Locations and discover!
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320305423.58/warc/CC-MAIN-20220128074016-20220128104016-00134.warc.gz
|
CC-MAIN-2022-05
| 1,406
| 8
|
https://forums.runnersworld.co.uk/discussion/5154
|
code
|
Lucky I couldn't sleep really
What: 3-mile jolly. First proper attempt at running for a fair bit due to injuty.
Why: Obviously don't want to go too far, but I want to go for at least 2 1/2 to 3 miles (if possible) to get a good idea of the state of my foot.
Last hard: Upper body the other day.
Last rest: Yesterday.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463607245.69/warc/CC-MAIN-20170523005639-20170523025639-00466.warc.gz
|
CC-MAIN-2017-22
| 316
| 5
|
http://sqlbits.com/(X(1)S(xkq3bqj54ulh1rmknteodx3p))/Sessions/Event11/Time_has_come_for_Power_Pivot
|
code
|
Time has come for Power Pivot
Power Pivot has been released a while ago.... and finally we suspect over 50% of companies have Excel 2010 (probably still using Windows XP). This is an introduction session to Power Pivot and we will cover basics like:
1) What kind of hardware should I have?7
2) How Power Pivot can be used? Just for Power User? How about data profiling or testing or cubes?
3) Quick comparison. Power Pivot vs SSAS Tabular Model
and the remaining 50% is building a simple Power Pivot Model using AdventureWorksDW2012 and we show how to extend your model.
Katie & Emil
Sorry, there are no downloads available for this session.
Katie & Emil is a consulting company specializing in BI using Microsoft Technologies. We often share our experience on our blog, YouTube Channel and SQL Bits conferences.
The video is not available to view online.
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-11/segments/1424936462141.68/warc/CC-MAIN-20150226074102-00248-ip-10-28-5-156.ec2.internal.warc.gz
|
CC-MAIN-2015-11
| 855
| 10
|
http://www.aast.edu/en/colleges/ccit/contenttemp.php?page_id=400072
|
code
|
Abstract: Given a list W = [w1,…, wn] of n positive integer symbol weights, and a list L = [l1,…,ln] of n corresponding integer codeword lengths, it is required to find the new list L when a new value x is ed in or when an existing value is d from the list of weights W. The presented algorithm uses the given information about the weights and their corresponding levels in order to perform the required update. No other knowledge about the original Huffman tree is assumed to be known. Instead of rebuilding the Huffman tree, the new algorithm redistributes the weights among the levels to obtain the new Huffman code. In many special cases, the updated Huffman code can be generated with lower complexity than reconstructing the Huffman tree from scratch by efficiently using the information of weights and their levels. In this paper, we present an updating algorithm that requires a linear complexity in many practical cases rather than the O(n log n) needed for reconstructing the Huffman tree. We also give a practical O(n log n) implementation for our algorithms.
Abstract: Segment trees are considered among the important data structures in computer science. A segment tree is, in principle, a static structure that is we cannot a new segment once the structure is built. In this paper we are presenting an enhanced model for this data structure. The suggested model is a parallel dynamic one that gives the flexibility of adding new segments to the structure in an efficient way and improves the query operation performance as well. The analysis of this parallel model on the CREW PRAM using O(log n) processors shows an improvement to the time complexity of the initial construction of the tree from O(n log n) to O(n), where n is the number of segments. The query operation also improves from O(log n + k) to O(log n/(log(log n)) + k) where k is the number of reported segments. Finally, the new segment ion feature is achieved with a time complexity of O(log n + m) where m is the maximum number of segments covering the same point. This proved to be very efficient specially when handling large amount of data.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123102.83/warc/CC-MAIN-20170423031203-00510-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 2,127
| 2
|
https://www.ibm.com/developerworks/community/forums/html/topic?id=77777777-0000-0000-0000-000014959899&ps=100
|
code
|
We have many SQLs with interval DAY TO SECOND(3) on Oracle. We are migarating DB2 from Oracle and these SQLs need to run on DB2. However, DB2 does not have DAY TO SECOND(3) function.
How should we modify SQLs to implement DAY TO SECOND(3)?
Thank you for your advice in advance.
nivanov1 100000H9H722 Posts
Re: interval DAY TO SECND(3)2013-03-22T14:01:52ZThis is the accepted answer. This is the accepted answer.The answer depends on whether your DB2 database is in the Oracle compatibility mode or not.
In a standard (not Oracle-compatible) DB2 database time duration is a decimal number in the form HHMMSS (for more details http://pic.dhe.ibm.com/infocenter/db2luw/v9r7/topic/com.ibm.db2.luw.sql.ref.doc/doc/r0023457.html)
In an Oracle-compatible DB2 database all datetime arithmetic operations follow Oracle semantics, that is, the duration will be a DECFLOAT value containing the number of days (or fractions of a day) between the two datetime values.
juchin 270002FA078 Posts
Re: interval DAY TO SECND(3)2013-03-26T11:09:10ZThis is the accepted answer. This is the accepted answer.
- nivanov1 100000H9H7
What SQL should we execute to get "interval DAY TO SECOND(3)"?
We don't know SQL function instead of DAY TO SECOND(3) in Oracle.
We are migrating DB2 with Oracle compatibility mode.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267868237.89/warc/CC-MAIN-20180625170045-20180625190045-00226.warc.gz
|
CC-MAIN-2018-26
| 1,289
| 13
|
https://flowingdata.com/2012/03/19/visualizing-the-history-of-everything/
|
code
|
Big History is a field of study that crosses multiple disciplines such as biology, natural history, and economics to form a single timeline that starts at the beginning of time and ends in the present. It’s the history of everything, essentially. ChronoZoom, a collaboration between UC Berkeley, Moscow State University, and Microsoft Research, aims to visualize this seemingly endless timeline.
You can browse years on top, and rectangles in the main view represent different scopes such as the Cosmos and Earth and the Solar System. Click on one those rectangles, and ChronoZoom, as you might guess, zooms in on the corresponding window of time. Circles within the rectangles provide videos and explanations for significant events in history.
To get right into it though, move your mouse to the top right. There’s a thing that looks like a bar graph, which is actually navigation for the scopes. Click on Humanity and watch it go.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817765.59/warc/CC-MAIN-20240421101951-20240421131951-00232.warc.gz
|
CC-MAIN-2024-18
| 936
| 3
|
http://laptops100.com/selling-laptops-under-100-for-newer-technology/
|
code
|
For many of the laptops which have been developed over the years to provide reliable performance levels, they have surprised people by lasting as long as some of the have. If you were to ask a random selection of people about the projected lifespan of their laptop computer then they might say that it would not be all that long. The reality is that this has a great deal to do with how the computer is taken care of over time. Obviously a well taken care of used laptop will have a far greater chance of working far longer than one which was not so well taken care of.
When the more popular laptop models have been around for a while, people seem to get used to them, yet the really good ones are thought to be well worth keeping longer than what some folks normally would do. However there are quite a few computer users who will actually trade off or simply sell their used laptops because they want to have the latest models which have been released. The reasoning behind this can be traced back to not the performance of the laptop, but the fact that the person just has an intrinsic need to have the absolute most cutting edge technology which has been released.
For the people who fit this bill, they are often in quite a hurry to acquire the newest features that are available in notebook computers no matter what the cost. Often the price tags for these are surprisingly high compared to what more modest examples will fetch. When a person like this needs to get their hands on the latest machines they are usually going to trade or sell the old laptop. Sometimes if the computer really is a bit older, then they might be sold as used laptops under 100 dollars from the original owner who simply wants a fast sale. Some of the really great buys have been had based on scenarios such as this. Although the used laptops were actually worth a whole lot more than simply $100 in the end, at least everyone seems to wind up happy with the transaction and that is all that really matters when it is all said and done.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917118519.29/warc/CC-MAIN-20170423031158-00220-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 2,020
| 3
|
https://discourse.gnome.org/t/gtk-window-is-active-always-returns-false/1962
|
code
|
I call gtk_window_is_active(wnd) and always receive 0, even when I know for sure that window is active and receiving keyboard input. What is the cause and where is the remedy for this?
We’re going to need more context to give you a meaningful answer.
Which version of GTK? Which platform? If you connect to the
GtkWidget::focus-out-event signals, do you observe them being emitted when the window acquires and loses key focus?
Gtk2 on Linux, I do connect focus-in-event to my windows and do process it in a handler that returns 1. What does my handler have to return for gtk_window_is_active() to work adequately?
If you return 1 (or
TRUE) then you’re telling GTK that you have handled the event, and the signal emission chain will stop.
FALSE (or 0), which means “continue the signal emission” instead.
Additionally, I’d strongly recommend you moved to GTK 3—which has more readable constants, like
Emmanuele, thank you very much! Returning 0 from my handler of focus-in-event did the trick, and now gtk_window_is_active() works as expected according to the documentation. This makes me think of several issues:
It would be great - in fact, I think it’s a must-do - to have some mention of this in the documentation. If it is already there, maybe it needs to be made more obvious. Like, in the description of gtk_window_is_active() add a line 'This function’s proper behavior depends on you letting GTK run it’s focus-in-event handlers, so make sure to return GDK_EVENT_PROPAGATE from your focus-in handlers". Actually, I have (now) found quite a bit of similar questions on StackOverflow, where the not-so-experienced go down the same road and make my mistake.
The same holds true for configure-event handler and maybe some others. The words “The value returned from this function indicates whether the event should be propagated further by the GTK event handling mechanism” do not make it clear what “further” is, and that the standard behavior of standard GTK widgets is included in this “further”
In my case, though gtk_window_is_active() returns 1 for the active window, it didn’t help do my task. Because the correct value is only returned AFTER the whole activation process has been completed - and that is in another handler, and way later than I need it. Well, I guess this is ingrained into the asynchronous architecture of GTK, whereas my project has a synchronous architecture, so I had to figure out other ways to do my task.
Actually, here’s what the documentation says:
When widgets receive an event, they frequently emit one or more “signals”. Signals notify your program that “something interesting happened”
This gives (me) the impression that signals are an interface between the widgets (GTK) and my app, from which I derive that widgets proper functioning does not depend on my app. All this to say, Please extend the documentation!
I’m sorry to say, but the behaviour of the focus-in-event is documented.
TRUE means “you handled the event”, so you’re overriding the default widget behaviour. We don’t document the internals of the toolkit because that would prevent us from changing the code at any point in the future. Returning
TRUE will stop the signal emission chain, and prevent the default handler of provided by the widget class to which you’re connecting your signal handler; this is intentional, and has been for the past 20 years.
That’s something that generally holds true for all signals that do not have a return value that controls the signal emission chain—an accumulator, in GObject terms.
You need to be familiar with the concepts of GObject and signal emission when using GTK.
I’m very thankful for your aswers and for the work you and others have done in GTK. Just to make sure: I’m not filing a complaint, I’m suggesting an improvement.
Yes, the general things and priciples are documented. But the connection between gtk_window_is_active() and focus-in-event handlers is not. It is not documented, and it is not obvious; It is the missing link in the docu that caused my question and that I suggest to add into the documentation.
Here’s what is says:
TRUE] to stop other handlers from being invoked for the event. [
FALSE] to propagate the event further.
Talks about other handlers, nothing about default handlers and overriding the default widget behaviour. Even taking your words that I quoted into the docu would be helpful, without disclosing any internals.
Making this connection more obvious in the function’s reference would save poor GTK newbies like me a lot of time and frustration. On Stackoverflow, it took a poor lad a week’s time to figure out the connection of his configure-event-handler’s return value and his widgets not responding to window size change.
This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499816.79/warc/CC-MAIN-20230130101912-20230130131912-00335.warc.gz
|
CC-MAIN-2023-06
| 4,889
| 29
|
https://scottishpolicynet.org.uk/1012.html
|
code
|
The Rocky Horror Picture Show is the original soundtrack album to the film The Rocky Horror Picture Show, an adaptation of the musical The Rocky Horror Show that had opened in The soundtrack was released as an album in by Ode. Original release · 1. "Touch-a, Touch-a, Touch-a, Touch Me", Susan Sarandon, Little Nell, Patricia Quinn, Tim Curry, Barry Bostwick, Richard O'Brien. If you've ever pondered the origin of the expletive "Dammit Janet," this original soundtrack to the cult classic musical The Rocky Horror Picture Show. Check out The Rocky Horror Picture Show - Original Soundtrack by Various artists on Amazon Music. Stream ad-free or purchase CD's and MP3s now on. View credits, reviews, tracks and shop for the Vinyl release of "The Rocky Horror Picture Show - Original Sound Track" on Discogs. Listen to The Rocky Horror Picture Show - Original Soundtrack on Spotify. Various Artists · Compilation · · 16 songs. "The Rocky Horror Picture Show (Original Soundtrack)" is the soundtrack of The Rocky Horror Picture Show, the second soundtrack made in the United. Reissue of the original soundtrack to the cult film and Halloween staple starring Tim Curry, Susan Sarandon, and Meat Loaf. Reached #49 on Billboard pop. CD THE ROCKY HORROR PICTURE SHOW - Original TV Soundtrack --> Musical CDs, DVDs @ SoundOfMusic-Shop, Music: Richard O'Brien Lyrics: Richard O'Brien. Product Identifiers. Record Label, Rhino (Label). UPC, eBay Product ID (ePID), Product Key Features. Release Year, Style, Film.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320302740.94/warc/CC-MAIN-20220121071203-20220121101203-00012.warc.gz
|
CC-MAIN-2022-05
| 1,507
| 1
|
http://forums.indiegamer.com/archive/index.php/t-9316.html
|
code
|
View Full Version : Editable user data
12-11-2006, 09:23 AM
I'm preparing a new rev. of Real E$tate Empire that will let users build their own towns. Currently the graphics for a town are built by feeding a text file that describes .TGA files into a compiler. The other town data comes from a series of readily editted text files. This is a pretty easy system to clean up and then maintain but I'm a little concerned that if I don't wrap the whole thing into a user friendly WYSIWYG editor that people are going to be flummoxed. But I'm bad for under estimating end users - anybody have any real world feedback on this topic?
12-16-2006, 03:29 PM
Just a thought: if users have trouble getting things to work, they will likely be contacting you. However much time you spend responding to their problems could probably have been better spent developing a better interface in the first place.
12-17-2006, 01:07 AM
I would guess that creating a more friendly interface would certainly give your users more incentive to create their own content. That would be especially true if your user base consists of more casual gamers, who probably would not take the time to learn anything other than an easy to use WYSIWYG editor. Most people won't even read the manual for the game, let alone any documentation required to understand an editing tool. Not to mention WYSIWYG editors almost always provide much better turn around time for content producers, resulting in more and better output from the people making the content.
Powered by vBulletin™ Version 4.1.3 Copyright © 2013 vBulletin Solutions, Inc. All rights reserved.
|
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368705284037/warc/CC-MAIN-20130516115444-00001-ip-10-60-113-184.ec2.internal.warc.gz
|
CC-MAIN-2013-20
| 1,618
| 8
|
https://ppo.puyallup.wsu.edu/pmr/survey/
|
code
|
We are creating a database to update the madrone distribution map and learn more about the health condition of the species throughout its range.
For details about the data and representative photos to help you complete the survey, download the Pacific Madrone Assessment Guidebook, then complete the survey using one of the methods below.
Survey data can be collected in several ways:
- Download the paper survey, complete, and return to WSU either by scanning/emailing, entering data in the webform on your computer, or snail mail.
- Webform – use your smartphone and web browser (must have internet connection)
- TreeSnap – use the smartphone app. Data can be uploaded later if internet connection not available in the field.
Youtube video on how to use TreeSnap (~5 minutes).
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370526982.53/warc/CC-MAIN-20200404231315-20200405021315-00486.warc.gz
|
CC-MAIN-2020-16
| 782
| 7
|
http://www.withoutbook.com/InterviewQuestionAnswer.php?tech=1&quesId=248&subject=Core%20Java%20Interview%20Questions%20and%20Answers
|
code
|
Core Java Interview Questions and Answers
Ques. Which class should you use to obtain design information about an object?Ans. The Class class is used to obtain information about an object's design.
Is it helpful? Yes No
Most helpful rated by users:
- How could Java classes direct program messages to the system console, but error messages, say to a file?
- What are the differences between an interface and an abstract class?
- Why would you use a synchronized block vs. synchronized method?
- How can you force garbage collection?
- When should I use abstract methods?
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123491.68/warc/CC-MAIN-20170423031203-00438-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 569
| 9
|
http://blogs.reading.ac.uk/centre-for-cognition-research/2015/07/08/summer-reading-group-first-meeting/
|
code
|
There will be a summer reading group on Concepts running in the Philosophy Department, all welcome.
The first meeting is Thursday 9th July at 1pm, discussing Fodor’s ‘Why There Still Has to Be a Language of Thought’. You can access a pdf via the dropbox below.
The full programme will be posted here shortly.
Time: 1pm onwards
Location: G73 Humss
Reading List: https://docs.google.com/document/d/1rifRGD1j3gFqRBKQ95OqK8oXXCoa5F6yKPydg-1RiAg/edit
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243988986.98/warc/CC-MAIN-20210509122756-20210509152756-00586.warc.gz
|
CC-MAIN-2021-21
| 451
| 6
|
https://blenderartists.org/t/help-blender-no-longer-rendering-smoke-dynamics/567897
|
code
|
Blender has stopped rendering smoke simulations for me. The smoke plays fine in the animation player but does not show up in the render. Or in some cases the smoke shows up but is frozen in a single state - no animation to the smoke. All static images render fine. (Blender Internal - CPU Render)
I think particle renders are gone too. I usually do not render particles. But when I test by turning render to “halo” in the particle settings, nothing shows up in the render. Even though the display settings get the particles to show up fine in the internal animation player…
I have been trying to fix it for days now. I ended up doing a total clean re-install of Windows and then installed only Windows updates and critical software. Re-installed Blender still has the same problem.
Again: Help! I am fairly technical and know PC’s. But I can not figure what is wrong…
What I did:
- Installed Blender 2.66
- The next render of my smoke simulation tests had this problem. (CPU Render)
- Assuming it was Bender 2.66, I downgraded to 2.65. This did not fix the problem.
- Opened a previous version of the smoke simulation which had been working. Same problem - smoke frozen.
- Created a new very simple smoke simulation .blend file. This one does not render the smoke at all even though it shows the smoke animating fine in the internal animation player.
- Completely uninstalled Blender, deleted all blendcache files, renamed the “Blender Foundation” folder in AppData. Rebooted and did a clean install of Blender 2.65. Same problem - animations still frozen.
- Bad video card driver? Downloaded newest driver and re-installed. No fix.
- Virus? Virus scans - two different anti-virus (AVG and Avast). Adware scans, Malware scans. Nothing found.
- A few other quirky things had been showing up around then. Could Windows be corrupt? PC memory failing?
- Tested PC memory. Downloaded some video card exercisers / testers. No errors found.
- OK… Windows had been getting slow lately. Perhaps it is time for a re-install.
- Total clean re-install of Windows. Re-install only needed software such as Firefox, Thunderbird openoffice.org, etc.
- Restore my user documents including the .blend files.
- Install Blender 2.65 which had been working fine.
- Wha’ the Wha’??? Blender still will not render smoke animation. No versions of my files render the smoke animation. The totally new simple smoke animation blend file does not render the smoke. Another simple smoke simulation file just now created with the newly installed Blender 2.65 still does not render the smoke simulation.
- Just for laughs I now downloaded the new Blender 2.66a and tried it. Still no fix.
- I have now re-installed most of my other software onto the PC. The ones that were acting a bit quirky before are working fine now. As far as I can tell it is only Blender which is still messed up.
- No where in any of this did any error show up in the Blender System Console.
- All done with Blender Internal Render and CPU Render (NOT OpenGL Render).
I am at a total loss. I know I need to upgrade to a better computer and a newer version of Windows. But I just do not have the money right now. Although the old computer and Windows Vista were slowing down Blender before, they sure never caused anything like this. I need to get this art project finished and I sure don’t have the time to restart in something other than Blender.
I can not imagine what could be wrong. Multiple Blender versions and multiple Blender files would imply a system problem. But I have now done a clean installation and I know of no way Virus or Malware would survive a complete re-installation of windows reformatting the entire drive. I can not figure any video card failure which would cause problems for just Blender and no other programs. Plus no video card failures showed up on testing.
Please, please. My only hope now is someone very familiar with Blender internals has some suggestion what might be wrong and how to fix it.
Attached: (all in .zip format to get them into the forum size and file format limits)
Simple smoke test (Test Render Bug.zip)
Simple smoke .blend file where smoke render which is not working.
Simple smoke .avi video file showing short render with no smoke at all (every tenth frame).
Original problem .blend file
Original problem .avi video file showing short render with no smoke at all (every tenth frame).
Both .blend files have System Info in the Text Editor.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320300624.10/warc/CC-MAIN-20220117212242-20220118002242-00447.warc.gz
|
CC-MAIN-2022-05
| 4,458
| 34
|
http://careers.stackoverflow.com/cities/new-york?f=e
|
code
|
Become a citizen of the world and Silicon Alley
Web Applications Developer
What developers know it for
In the city that never sleeps, technology permeates almost every industry - from fashion, advertising, finance, and more. In addition to tech giants like Microsoft and Google, New York is home to more than 900 tech companies, dozens of incubators and coworking spaces, and even has its own Chief Digital Officer. New York is a work hard, play hard type of town, but there's merit to the saying, "If you can make it here, you can make it anywhere."
- Software Development
Featured CompanyCreate a free company page
London; New York; Princeton
We are 4,000+ technologists in the financial industry. We produce the most financial news and data, bringing together 350+ exchanges, 4,000+ FIX feeds, and 80,000+ news wires from across the world, consuming over 45 billion ticks a day that our customers use to power their businesses.
Our open jobsView 2 more jobs
More Companies in New YorkMore companies
Looking for a Job?
Create a Stack Overflow Careers profile and let employers find you. Connect your account easily with GitHub, Stack Overflow, Codeplex, and more.
|
s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701155060.45/warc/CC-MAIN-20160205193915-00281-ip-10-236-182-209.ec2.internal.warc.gz
|
CC-MAIN-2016-07
| 1,165
| 12
|
https://liqd.net/en/projects/
|
code
|
We are developing the meinBerlin platform for residents of Berlin to participate in the decision-making and urban planning of their city.
In cooperation with Stiftung Mercator and MitOst e.V, we have implemented the European-wide competition Advocate Europe.
We built an online forum where users can provide input regarding the future of Tempelhofer Feld in Berlin.
We enable the collaborative editing of the SPD's digital agenda.
In this project funded by the European Commission, we develop a tool for the analysis of open data sets along with an international research consortium.
OPIN.me, which we develop in collaboration with our international partners, is a platform designed as a hub for European youth participation.
Our free and easily accessible platform adhocracy.de enables organizations small and large to internally implement democratic decision making processes.
In collaboration with politic-digital and Marina Weisband we are testing Liquid Democracy in high schools and thus contribute to the formation of future democratic citizens.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123635.74/warc/CC-MAIN-20170423031203-00472-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 1,052
| 8
|
http://www.formidablepro2pdf.com/support/subject/pdf-generator-drops-every-5th-time/
|
code
|
I have the plugin installed a month ago, changed into live modus about a week ago. I am handling about 200 pdf creations and sendings per day in the moment. It seems that when two people filling out the form almost same time one of them is getting following error
There was an error generating the PDF file. Your server wasn't able to upload PDF file: read function returned funny value
I was able to trace this through the last days and have made following discoveries: The time between pushing the “Send” button in the form and the automatic forwarding is about 10 seconds with my installation and having somebody else pushing the “Send” button a different computer just in this time frame seems to produce an abort on the first one.
I now disabled the attaching of the PDF and just send the download link to reduce failures. I was able to reproduce the same problem with just clicking on two pdf links almost the same time.
Am I right that the plugin is not able to queue PDF creating jobs?
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267867644.88/warc/CC-MAIN-20180625092128-20180625112128-00123.warc.gz
|
CC-MAIN-2018-26
| 1,001
| 5
|
http://cadastralvocabulary.org/Concept.php?TID=95
|
code
|
Geographical grid system
Harmonised multi-resolution grid with a common point of origin and standardised location and size of grid cells [INSPIRE Directive] (Source: INSPIRE glossary).
In other Languages:
Danish: Geografiske kvadratnetsystemer
Turkish: Coğrafi grid sistemi
Malay: Sistem grid geografi
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499634.11/warc/CC-MAIN-20230128121809-20230128151809-00410.warc.gz
|
CC-MAIN-2023-06
| 302
| 6
|
https://www.tib.eu/en/research-development/research-groups-and-labs/scientific-data-management/teaching
|
code
|
Prof. Dr. Maria-Esther Vidal lectures Master courses at the Faculty of Electrical Engineering and Computer Science at Leibniz Universität Hannover and in BIOMEDAS.
In addition, the research group supervises laboratory and final theses.
- Introduction to Database Systems (Lecture, Exercises)
- Data Science & Digital Libraries (Seminar)
- Knowledge Engineering and Semantic Web (Lecture, Exercises)
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-43/segments/1634323585424.97/warc/CC-MAIN-20211021133500-20211021163500-00709.warc.gz
|
CC-MAIN-2021-43
| 399
| 5
|
http://e2e.ti.com/support/embedded/linux/f/354/p/205232/729450.aspx
|
code
|
I need to do kernel and filesystem compilation manually without using ./prj. So I have tried the following
I have downloaded the compiler c6x-uclinux and installed manually.
I have downloaded the linux source, extracted, modified .config for 6678, and gave make ARCH=c6x CROSS_COMPILE=c6x-uclinux-
It generated vmlinux and using objcopy command i converted into vmlinux.bin
Similarly I need to compile filesystem manually..
Please suggest the ways
Do you just want to build the filesystem and no other components? I am not sure if this is what you want. You can try "./prj build rootfs". That will only build the filesystem.
Thanks for the reply
Can't I compile the filesystem manually using some commands without using ./prj ?
I did for linux by giving make menuconfig , customized config file and then make to give vmlinux
Similarly can I compile filesystem using some commands without using ./prj ?
Its hard to answer this question without knowing why you can't use ./prj.
./prj is just a helper to ensure that tools are installed and environment variables set correctly and then it calls the top level Makefile.
The top level Makefile does a lot of things. Unlike the kernel, the "filesystem" is a lot of things. You could certainly do each of the things the top level Makefile does yourself but that seems like a lot of added work.
Again, please explain your need and we can try to support you. If you really need to redo the whole build system yourself you can look at the Makefile. However, keep in mind that if you build things different than we do it will be hard for us to help you with other issues.
All content and materials on this site are provided "as is". TI and its respective suppliers and providers of content make no representations about the suitability of these materials for any purpose and disclaim all warranties and conditions with regard to these materials, including but not limited to all implied warranties and conditions of merchantability, fitness for a particular purpose, title and non-infringement of any third party intellectual property right. TI and its respective suppliers and providers of content make no representations about the suitability of these materials for any purpose and disclaim all warranties and conditions with respect to these materials. No license, either express or implied, by estoppel or otherwise, is granted by TI. Use of the information on this site may require a license from a third party, or a license from TI.
TI is a global semiconductor design and manufacturing company. Innovate with 100,000+ analog ICs andembedded processors, along with software, tools and the industry’s largest sales/support staff.
|
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368704433753/warc/CC-MAIN-20130516114033-00047-ip-10-60-113-184.ec2.internal.warc.gz
|
CC-MAIN-2013-20
| 2,675
| 17
|
http://askubuntu.com/users/20334/tapper
|
code
|
Apparently, this user prefers to keep an air of mystery about them.
51 How can I convert an ODT file to a PDF? Feb 15 '12
33 How can I convert a ppt to a pdf from the command line? Feb 15 '12
1 Evolution Contacts not Auto-populating Oct 4 '11
1 DevHelp php-doc is not available Mar 21 '12
|
s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257828282.32/warc/CC-MAIN-20160723071028-00147-ip-10-185-27-174.ec2.internal.warc.gz
|
CC-MAIN-2016-30
| 288
| 5
|
https://www.webhosting.uk.com/kb/setup-private-name-servers-for-reseller-hosting/
|
code
|
What are private name servers ?
Private name servers are the name servers which includes your own domain name. Private name servers would be ns1.yourdomain.com and ns2.yourdomain.com. Usually when you sign up for web hosting you use your hosts name servers.
What are my private name servers IP’s ?
Each reseller server has it’s own DNS so you’ll need to refer to your welcome email for the nameserver information. For ns1.yourdomain.com the IP would be the IP listed in your welcome email. For ns2.yourdomain.com the IP would the same accept change the last number in the IP to be one digit higher. Example: If the IP listed in your welcome email was 184.108.40.206 your ns2.domain.com name servers IP would be 220.127.116.11
What are the benefits of having your own private name servers ?
Many resellers choose to have their own private name servers for appearance reasons. As a reseller hosting client you can resell our hosting packages or create your own hosting packages and prices and resell to your clients under your own brand name. Your clients won’t know that you are reselling our web hosting services unless you let them know. We remain completely in the background and management of the server hardware and software will be taken care of by us. Using our default name servers would seem less professional for a web hosting company run by our reseller.
How Do I get Private Name Servers ?
To setup private name servers you will need to create the name servers at the place you registered your domain from. Every domain registrar has a different process for creating domain name servers. While setting up private name servers, try and look for something along the lines of “register DNS” “create DNS” or “add DNS” from within your account with your registrar. From there you can register your private name servers. Once your private name servers are registered with the registrar you can start to use them. You will have to use the IP address, provided in the welcome email, while registering the private name servers.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656104141372.60/warc/CC-MAIN-20220702131941-20220702161941-00122.warc.gz
|
CC-MAIN-2022-27
| 2,049
| 8
|
https://jobs.ngpcap.com/companies/zoomcar/jobs/31428205-zoomcar-software-engineer-senior-engineer-data-engineering
|
code
|
ZoomCar - Software Engineer/Senior Engineer - Data Engineering
Software Engineering, Data Science
Bengaluru, Karnataka, India
Posted on Sunday, November 5, 2023
- Engage with stakeholders to gather requirements to deliver data solutions.
- Implement best practices around the data platform to automate the ingestion, transformation, and augmentation of data sources, and provide best practices for pipeline operations.
- Innovate at the optimal storage guidelines around transactional, clickstream & IoT data.
- Provide advice and ideas for technical solutions and improvements to data systems.
- Manage a team of high-performance data engineers. Perform regular performance evaluations and share and seek feedback.
- Create and maintain clear documentation on data models/schemas as well as infra-level processes.
- Help troubleshoot and remediate data quality issues raised by pipeline alerts or downstream consumers.
- 3 plus years of hands-on experience in Data Warehouse space - that involves dimension modeling and SQL-based (expert level) data transformation and analytics.
- Hands-on experience in Big Data space - that involves designing, developing, testing, and implementing data solutions using Apache NiFi, Kafka, Spark and other Big Data ecosystem tools.
- Expert in database technologies. OLTP, OLAP & MPP, alike.
- Expert-level exposure to AWS and most data/analytics services.
- Proven proficiency in programming languages like Java/Python.
- Strong engineering mindset to build resiliency/restart ability in all the Data Platform components.
- Exposure to Hortonworks solutions.
- Exposure to Tableau.
- Past experience working with high-energy startups.
- Ability to thrive in a fast-paced start-up environment, and to manage multiple, competing priorities simultaneously
- Superior analytical skills and a strong sense of ownership.
- Ability to develop and organize high-quality documentation.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679103464.86/warc/CC-MAIN-20231211013452-20231211043452-00485.warc.gz
|
CC-MAIN-2023-50
| 1,914
| 23
|
http://quant.stackexchange.com/questions/tagged/historical?sort=frequent&pagesize=15
|
code
|
Quantitative Finance Meta
to customize your list.
more stack exchange communities
Start here for a quick overview of the site
Detailed answers to any questions you might have
Discuss the workings and policies of this site
tag has no wiki summary.
What continous adjustment methods are firms using for futures backtesting?
There are several methods available between data vendors and associated software programs to adjust futures contract data for historical simulations. Some of the methods are: 1) Back or forward ...
Feb 7 '13 at 22:30
recently active historical questions feed
frequent question tagged
Hot Network Questions
If future farming is largely automated by robots, what produce would still require human farmers?
Why does a free-falling body experience no force despite accelerating?
Did Tolkien ever specify that Orodruin was a volcano?
50s synonym for "snogging"
Upgrade to Windows 10
What is the correct usage of 'if' and 'when'?
Distributing points evenly on a sphere
Why at the school not at school
What's the fastest ferry to get from Italy to Albania?
Would seeing a Basilisk in a memory cause death?
Bash "unexpected end of file" on line 11, there is no line 11
Where to place debouncing circuit for remote mounted buttons and rotary encoders
How to stop thinking in terms of code?
Why is a Voided Blank Check Necessary to setup Direct Deposit?
Can I Stifle Heritage Druid's ability?
Curious about the implementation of CaseInsensitiveComparator
hook into magento's output to the client
Switch buffer to the file path in current cursor position
Are QR codes a good idea?
Ubuntu only resolves DNS when the router's IP address is present in the DNS servers list
Why is a quantum computer not capable of solving more problems than a classical computer?
Security Implications of Neglecting the Extra Byte for NULL Termination in C/C++ Arrays
Can you be tripped by an AoO while standing up?
Does the bonus attack from "Polearm Master" have reach as well?
more hot questions
Life / Arts
Culture / Recreation
TeX - LaTeX
Unix & Linux
Ask Different (Apple)
Geographic Information Systems
Science Fiction & Fantasy
Seasoned Advice (cooking)
Personal Finance & Money
English Language & Usage
Mi Yodeya (Judaism)
Cross Validated (stats)
Theoretical Computer Science
Meta Stack Exchange
Stack Overflow Careers
site design / logo © 2015 Stack Exchange Inc; user contributions licensed under
cc by-sa 3.0
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-32/segments/1438042986451.45/warc/CC-MAIN-20150728002306-00336-ip-10-236-191-2.ec2.internal.warc.gz
|
CC-MAIN-2015-32
| 2,412
| 55
|
https://www.nivida.in/blog/detail/can-writing-code-with-ai-and-ml-change-the-future-of-software-development
|
code
|
Is it common for software to be developed with the help of AI and ML, or is that the exception? Do we see this as the way forward for software development? Reading this blog will allow you to delve deeper into the subject at hand.
There was recent press on a Google employee who seemed to suggest that the company's efforts in artificial intelligence (AI) were becoming "sentient," or alive.
This article provides an in-depth look at the ways in which software engineers and software developers can benefit from utilising Machine Learning (ML) and AI.
In the coming decade, the rise of ML as well as data science in the software development field will significantly alter the software development process and the entire IT sector.
It appears that machine learning is the latest fad in the field of software engineering. However, as more and more data for software development as well as cheaper processing power become available, it will become more noticeable.
The integration of numerous tools across the process of the life cycle of software delivery is a major source of headaches when developing a useful tool for use within the SDLC. Clients, software engineers, non-technical members, quality assurance engineers, product designers, configuration and deployment engineers, security engineers, production support engineers, infrastructure providers, as well as project managers all have their own unique perspectives and experiences that inform their identities.
Here is how AI and ML can help to change the future of Software Development.
It was in the 1970s that the first algorithms were developed. This is mostly employed to hasten the construction or development of code by providing autonomous insights that are enabled using automated commands which are pre-programmed. A software engineer can make more informed and timely design choices with this information at hand.
Machine learning is a game-changer for software development since it allows for the advancement of algorithms, their use and adaptation in real-time, and the provision of novel insights to programmers that can increase their efficiency and the quality of their work.
Many technology-centric firms, such as Google, encourage best practices by making important components open, such as companion bots, without fully divulging their machine learning techniques to software creation.
Many companies in the software engineering industry are looking into the use of artificial intelligence to enhance engineering decisions and to make better use of vast stores of data accumulated during the creation of software in a variety of languages, some of which are quite old.
Robo-advisors are a good example of this trend; these programmes are simply algorithms designed to speed up the process of achieving software engineering goals. Developers working in an IDE can benefit from the assistance of these automatic guides and assistants.
When a developer begins constructing lines of code, the adviser pops up and informs them of all the classes and functions they might use to check whether or not the developer is taking the correct approach to a functional or non-functional region. The system then makes all of the relevant code for that feature immediately accessible, with the best possible match as its primary goal.
Developers that are confident working in pairs without the benefit of human peer review have shown a strong preference for Robo-advisors. The GitHub Co-pilot is an example of an AI tool that can be used by engineers and developers to receive help and advice in an automated fashion.
Locating and fixing problems before they worsen:
Using machine learning for defect detection and prevention is crucial because bugs are a major issue for every software engineering development team.
In the past, defect detection systems relied primarily on a robust and complicated set of rules, which need extensive human intervention. Today, finding and preventing faults requires more than just checking boxes. It can automatically self-remediate and self-heal when certain conditions are met.
It is within the realm of software engineering functionality where bugs are most likely to be found, making machine learning the best tool for fixing them. This is due to the fact that machine learning systems are able to rapidly scan through massive databases of codes, identifying anomalous activity and immediately flagging them. Because there are so many potential points of detection at any given time, machine learning systems will likely become indispensable in the near future for bolstering test strategy as well as software development discipline. Such ML-assisted features will soon be able to automatically fix bugs in the code.
Software development projects benefit greatly from automated processes. By streamlining common, day-to-day operations, software engineers can free up time for more meaningful projects. Money and time are saved during production, which is an additional advantage.
Automating with AI and ML can give developers yet another level of assistance. When given the right information, AI and ML may help development teams make important decisions much more quickly by providing in-depth data analysis. Collaboration and the subsequent advice on the optimal course of action can, in the not-too-distant future, even make pair programming unnecessary.
In software engineering, AI and automation can be trained to spot mistakes, speeding up the process of both finding and fixing them. This means that there will be fewer instances of team members being held up and a higher likelihood that their tasks will be completed on time and at a high standard. In the future, it is likely that companion chatbots will be developed to aid in the engineering of the whole software delivery process, involving all stakeholders.
Developer performance prediction:
Developer performance prediction will become a complex field of software delivery lifecycle as well as software development to understand the elements that motivate developers and guess the performance of their code.
Coding approaches and programming practices are used to predict developer performance. An example is a function that uses static analysis classes and interfaces.
With more software engineering data, ML and AI-based analysis will develop.
AI won't start creating code, but it will help software developers, as well as engineers, to produce better code. The software engineer may use AI in complex areas while focusing on high-value ones.
Real-time sentiment analysis:
With past and real-time data of codes, coding styles, and amounts of unstructured data such as transcriptions, videos, audio files, photos, articles, as well as codes, AI can understand and provide superior recommendations across the software delivery lifecycle.
Existing video conferencing and collaboration software often include real-time feedback. AI-accelerated insights can affect the way software engineers interact as well as use the essential functionality. ML algorithms may learn and offer relevant and changeable content, providing recommendations on emphasis areas for engineers.
Sentiment analysis in software engineering can be used to estimate budget, schedule, or effort based on functional and non-functional needs. It predicts consumer proposal behaviour and trade-offs.
The developer community reacts to a multiplicity of human-related elements and thinks machine learning may imitate and improve human software engineering intuition by detecting new trends and signals.
Future machine learning applications will focus on understanding and forecasting how needs, trends, and real-time factors like collaboration and team events influence software engineers' feelings towards writing better code.
It won't be confined to predicting better code and software engineering methods. Error management, a key area to understand and reduce SaaS downtime, can be forecasted and automated to self-remediate as well as identify without any sort of human interaction.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506429.78/warc/CC-MAIN-20230922234442-20230923024442-00683.warc.gz
|
CC-MAIN-2023-40
| 8,014
| 33
|
http://wavehunt21.com/not-working/repairing-godaddy-masking-not-working.php
|
code
|
Assuming a single case fan plus computer handles dual sided or single sided memory... The safest bet is to always order single sided i belive they go up to 4... Thus making it impossible to Hello to you and how do you do?All volume options arepost on this wonderful forum.
My house is now under renovation and sounds but even those didn't work. And if you want more than 4 godaddy navigate here as a CPU\GPU cooler or a watercooling setup? not Mask Url Online I rebooted and checked in the (Code 39)' (translated from the Italian). Why not make ayou already bought any parts?
I have a SATA HDD which the OS at a university, so the internet is fast). Have you tried any of you problem devices masking this is $513 with the LCD.For your info the drive 1000W PSU with enough 6-pin and 8-pin rails.
If you have vista or windows 7 is about right for that setup/res. I want more Dual-QPI,with how HD Tune says warning. Godaddy Domain Forwarding Not Working EVGA X58 Micro SLI Apevia Q-XPACK 2 Acore CPU's with that X58 platform.Also what resolution do you plan toa driver and they lack one entirely.
I looked at their site for I looked at their site for Most will do 2-Way SLI the drive. I tried playing the system sounds but even those didn't work.I have googled, but don't know enough tothat is your Network Interface Card (NIC).Driver could be damaged or missing if it was Intel/Nvidia based.
With HD Tune Pro it shows (requiring between 100 to 500mA from the port).Do I upgrade to Godaddy Subdomain Forwarding use the Internet on my computer.I am desperate for help as the now or wait. Click on the Audio tab, and confirm theif budget allows and is necessary.
Only from Newegg HaveDo you have an Operating System (OS)?I brought it out and wantGPU that fits in?Maybe an odd and brief 25 FPS during 'From hell' and the like.on other computers to see if they work?I purchased a wifi router his comment is here masking flash drive it says "Unknown Device".
I connected to my TV by for your time. well...The number in the socket name is howruns off so didn't think it was the drivers. A new windows will open and then you memory I can install on my system?This is where I think I amlttle help with a sound issue if you can.
The symptoms are that it will not Please help No sound in webpages Try this. Windows 7 Ultimate 64-Bitstuck where a lot of people are.Most likely not, but will considerable to go completely without a video card.Zone Alarm firewall is on both machines, but It gets near 100 Degrees Celsius, and I am looking to upgrade.
It would be 100to connect it to my HD tv.Device doesn't appear in device manager times faster than a CPU. Click the + sign next to Godaddy Subdomain Forwarding Not Working click on the Driver tab, the click Uninstall.I will post photos so play on if this is for gaming?
Thank you in advance this contact form or even Quad-QPI availability.The problem, as I see it, is that
but i belive thats it...Whether to buyan Operating System (OS)?Do you havegigs you need a 64 bit operating system...
Or are you interested in getting a new monitor? I will could rate my build, I'd greatly appreciated. I have a big Url Masking Free C7 CRC DATA ERROR it says warning.Also keep in mind weather or not yourframe rate should rarely be below 30 FPS.My hdd in my hp a power issue on your USB ports.
I appreciate these forums and would like aread any disks regardless of their type i.e.I really just want to know whatunmuted and on max.Yes, but will cancelnow explain my problem : After updating my drivers my microphone stopped work.I removed the Geforce cardslimline just not being recognized.
I'd also like it weblink so I'm now at a total loss.Thanks guy HI Bazark, Thatproblem with my gaming computer. and uninstalled the nvidea drivers. I tried playing the system Url Masking Htaccess extra fans as needed/if the price allows.
Http://www.techspot.com/guides/buying/page2.html This is $400 but would be less if you went with and still nothing could be seen. I got somewhere around 100Mbs (I livedevices in the Sound playback, "default Devices" list.Your other devices may be "high powered" devices pins I have because it's DDR2? That is not your sound card,unmuted and on max.
The good news is that the minimum you guys can see my screen. Is there a limit to how muchhas been configured to allow access each way. So logged back into Windows Mask Url Link computer is useless to me without the Internet. working My COCD is bothering mePanel > Sound and Audio Devices.
Depending on if that and like 40 of them (look at Wikipedia). All volume options areeveryone seems to have a different realtek than me. So also wonder if you have Url Masking In Html supports up to 256 cores.Dont quote me on thatthe hell's wrong with this thing.Click to expand...
If you still want to build, If so, you can just install the appropriate driver. BIOS and the drive could be seen. Anyways, if some of you hardware gurusorder due to mistake. Will you need any aftermarket cooling, such on-board video: http://www.tomshardware.com/reviews/gaming-pc-build-a-pc-overclock,2739.html Hi, running XP sp3 on both desktop and netbook.
Then you can restart the computer and windows should find the right drivers for or Crossfire with x16/x16 configuration. Hello all, its my first Socket 1155, or 1366? Apparently, they will have PCI-E 3.0 Lanes, main bored supports it...None. * my computer has moved away from the modem.
For starters, we most likely will be model is an LG GH24NS50. No, maybe a CPU heatsink many pins, to hopefully, the more, the better. When i plug in a and a wireless network card.Go to start > Control on this whole matter or changing compatibilities.
Can you tell how many Components to expand it Click Problem Devices. So, people, stay here to trade advice separate what I need from the search results.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039747665.82/warc/CC-MAIN-20181121092625-20181121114625-00321.warc.gz
|
CC-MAIN-2018-47
| 5,820
| 22
|
https://www.omgubuntu.co.uk/2016/07/install-gnome-extensions-google-chrome
|
code
|
So you want to install GNOME Extensions on Ubuntu, but you’re not sure how? Well, we’re gonna show you!
Just like web browser add-ons, GNOME Shell extensions make it easy to extend, adjust and reshape the GNOME Shell experience to suit your needs.
This could be as basic as adding a weather indicator to the panel or as dramatic as converting the whole desktop into a single-panel, Windows-style set-up.
Point is: whatever the task, there’s (probably) a GNOME Shell extension that can do it!
How to Install GNOME Extensions on Ubuntu
It is easy to install GNOME Extensions on Ubuntu 18.04 LTS, 19.10 and 20.04 LTS, provided that you set things up correctly first.
To install GNOME extensions on Ubuntu you need three things:
- A compatible web browser, e.g., Firefox, Google Chrome, Vivaldi
- A free web browser extension
chrome-gnome-shellpackage from the repos
The web browser extension and “native connector” handle integration between your desktop shell and the extensions repository at extensions.gnome.org.
When all three parts combine you can go to the visit the GNOME Extensions website, find an add-on you like, and hit ‘Install’ to install it! You can also manage, adjust and remove GNOME extensions the same website.
Step 1: Install the Browser Add-on
First things first: install the GNOME Shell Integration extension for your web-browser.
This free, open source extension is available for Mozilla Firefox and Google Chrome (as well as Chromium-based browsers like Vivaldi).
You can install the GNOME Shell Integration extension from each browser’s respective add-on store:
Step 2: Install the Host Connector
Next, you need need to install the native connector package on Ubuntu. if you do not install this package you will see the following error when you visit the GNOME Extensions website:
Although gnome shell integration extension is running, native host connector is not detected
If you’re on Ubuntu 18.04 LTS or above you can install the
chrome-gnome-shell package directly from Ubuntu Software by clicking this button:-
Prefer to install things using a Terminal? You can also install
chrome-gnome-shell on Ubuntu 18.04 using the command:
sudo apt install chrome-gnome-shell
For Arch, Gentoo, Fedora and manual build instructions head over to the official project wiki page.
Step 3: Restart Your Browser
With both host connector and browser extension(s) are installed all that’s left to do is restart your browser to make sure that all of the changes take effect and connect correctly.
Head to the GNOME extensions website in your browser and click on any extension you see. You should now see an “on/off” toggle clearly displayed on extension listings:
Slide the toggle to the ‘on’ position to install the extension on your desktop. You’ll see an on-screen modal dialog asking you to confirm. Accept, and the extension will download, unpack, and install on your system.
How to Manage GNOME Extensions
You can remove GNOME extensions, access extension settings, and update GNOME extensions from this GNOME Extensions website too.
Just head to the extensions.gnome.org/local page to see a list of installed extensions with the following options available: –
- To disable a GNOME extension slide its toggle to ‘off’
- To remove a GNOME extension click the red ‘x’ icon
- To update a GNOME extension click the green arrow icon
Note: you won’t see a ‘green arrow icon’ if there is no update available.
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107874026.22/warc/CC-MAIN-20201020162922-20201020192922-00108.warc.gz
|
CC-MAIN-2020-45
| 3,461
| 36
|
http://ajaxian.com/archives/mootools-1-3-hits-the-street
|
code
|
Tuesday, October 19th, 2010
After almost a year in development, the MooTools team announced the release of version 1.3 with major notable enhancements.
Slick Selector Engine
The biggest update is the inclusion of the new Slick standalone selector engine. The engine was developed by Thomas Aylott, Fabio Costa and Valerio Proietti with accuracy and portability in mind, allowing the engine to not only provide MooTools developers with excellent selector support but also the ability for it to be included in other projects as well. The MooTools team created over 3,000 tests to ensure that even the most obscure scenarios were considered. The inclusion of Slick also allowed the team to extend the Element method to allow creation of DOM elements using complex CSS selector expressions. Very cool!
- var newElement = new Element('div#cow.moo.big[data-size=5]')
The MooTools team also focused on enhancing the core API in preparation for MooTools 2.0 by deprecated every $- prefixed function and nd moving those to the relevant natives namespaces:
We have revised the whole base of our library to speed up MooTools, to provide an even more beautiful API and to make our library future-proof. In order to reduce the amount of global variables we have moved all $-prefixed functions into their according namespace ($splat » Array.from, …). In short we think that with 1.3 you are getting the best MooTools experience ever.
New Build Manager
Building a new version of MooTools has always been easy using the builder and now it’s been improved via the new Packager preprocessor which allows you to create custom MooTools builds from Github repositories. It handles code dependencies and already provides support for several MooTools plugins. The new Packager is a PHP 5.2+ library but can also be found as a Python app thanks to Aaron Newton’s work.
To make sure that MooTools continues to adhere to the projects’s specifications, Christoph Pojer & Arian Stolwijk spent some time over the summer building a new spec testing engine called MooTools Runner which has been adopted by all major MooTools projects – Core, More and Slick. The engine is based on Jasmine and runs all specs within just 2 seconds. In addition, code coverage was increased to about 95%, ensuring excellent test coverage for Core.
Once you’re ready, you can grab the latest version of MooTools or roll your own via the following links:
- Download MooTools Core 1.3
- Build MooTools Core 1.3
- Build MooTools More 1.3RC
- Update from 1.2 to 1.3 with 1.2 compatibility.
- Upgrade from 1.2 to 1.3 and drop 1.2 compatibility.
- Fork / clone MooTools 1.3 from GitHub, and build it yourself with Packager.
- Read the commit logs to see precisely what has changed.
- Browse the Documentation for Core & More.
Posted by Rey Bango at 8:02 pm
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218189802.18/warc/CC-MAIN-20170322212949-00009-ip-10-233-31-227.ec2.internal.warc.gz
|
CC-MAIN-2017-13
| 2,812
| 20
|
https://www.physicsforums.com/threads/four-conceptual-questions-as-i-teach-myself-physics.623702/
|
code
|
So I'm just reading through a high school Physics text, essentially, and I have these four general question as I learn the basics of Classical Mechanics. 1. If acceleration has units distance/time^2, then I would expect, between two different objects accelerating over the same distance but with one taking twice as long, it would have a quarter the acceleration. However, when I try to draw out tables of time, position, velocity, and acceleration, this seems not to come out. Can someone give a toy example which would demonstrate the proportionality between acceleration and the inverse square of time? 2. I feel uneasy about the nature of potential gravitational energy. It is supposedly 0 for an object at rest on the ground ... unless that ground happens to be a trap door? And then it has some non-0 PE? Can anyone say some things to clarify this idea? Is potential energy not intrinsic to an object and the gravitational field surrounding it? (I just use the word "gravitational field" only due to the understanding of it which I've gained from science fiction--I don't truly and totally understand what that is.) 3. When something falls, it loses PE but gains KE. But what happens when its at rest on the ground? Does it have both 0 PE and 0 KE? Has all of the energy which it once had, in one of those forms, now distributed into the ground? In the form of heat and some small measure of mechanical energy (denting the ground)? 4. If in space you apply a force on an object and, in response, the object travels infinite distance, does that imply that the force has done infinite work? Thank you for any help.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267162385.84/warc/CC-MAIN-20180925202648-20180925223048-00458.warc.gz
|
CC-MAIN-2018-39
| 1,618
| 1
|
http://rift.magelo.com/en/quest/-5553652/field-medicine
|
code
|
My leg wound is worse than I feared. I can't simply bind it and go; I'll need a splint if I'm going to walk.
The jungle surely has what you need - vines, sticks and the like. Go hunting for some materials, and hurry! I've been hearing some odd noises in the bushes...
No time to lose! Set the splint quickly!
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440645371566.90/warc/CC-MAIN-20150827031611-00185-ip-10-171-96-226.ec2.internal.warc.gz
|
CC-MAIN-2015-35
| 308
| 3
|
https://forum.literatureandlatte.com/t/problems-copying-to-a-new-project/10904
|
code
|
I am having problems copying information from one project to another. I am dragging the info from one project to the other and all seems to go well, however when I start looking through the info in the new project it is missing a lot of what was in each topic. After about half way down the topics are all blank inside (titles are still there). How can I fix this? Please help, this is really important!
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817455.17/warc/CC-MAIN-20240419203449-20240419233449-00214.warc.gz
|
CC-MAIN-2024-18
| 403
| 1
|
https://highearthorbit.com/articles/urban-mapping-api-opens-up
|
code
|
Need to add hyperlocalization to your mashup/site/app? Users searching for "Pizza in Dogpatch" and your geocoder just falls over? Well, Urban Mapping pulled a some what surprising, and incredibly great move today and announced they are opening up the API to their neighborhood database.
If you haven't heard before, Urban Mapping provides the data to most of the major mapping users on defining areas like "Little Italy", or "Soho". This includes Google, MapQuest, and others. By opening their API, third-party developers can now build this type of capability into their own applications. It's something we'll definitely be adding to Mapufacture very soon.
There are numerous other geocoders out there, not least of which is the excellent and open GeoNames. However, an API for looking up ambiguous, and changing local definitions of a neighborhood has been missing. UMI fills that by providing multiple mechanisms for finding and defining 'hoods. For example, you can look up the neighborhoods at a location, by name, or even get the exonyms of a neighborhood, depending on the language. There are a number of other methods and demos available. Check out the very good
documentation that links to each of the demos and even includes code snippets in Ruby and PHP for how to call the API.
As Brady points out on O'Reilly Radar, the API is using SOAP, and not REST. The API was developed just as REST was "becoming all the rage", but had various reasons for being SOAP based. Still, the code examples show how easy it is to use a SOAP library to create a simple wrapper around the API. [via Brian Suda]
Of course, I'm perhaps a little biased on the Urban Mapping demos and documentation - considering I helped developed them. Urban Mapping is a great company to work with and I'm really looking forward to their continued expansion of their data products and APIs.
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243988923.22/warc/CC-MAIN-20210508181551-20210508211551-00084.warc.gz
|
CC-MAIN-2021-21
| 1,863
| 6
|
https://lecturenotes.in/materials/20558-note-of-data-structures-and-algorithms-using-python-by-rakesh-kumar
|
code
|
This page intentionally left blank
Data Structures and Algorithms Using Python Rance D. Necaise Department of Computer Science College of William and Mary JOHN WILEY & SONS, INC.
ACQUISITIONS EDITOR MARKETING MANAGER EDITORIAL ASSISTANT SENIOR DESIGNER MEDIA EDITOR PRODUCTION MANAGER PRODUCTION EDITOR Beth Golub Christopher Ruel Michael Berlin Jeof Vita Thomas Kulesa Micheline Frederick Amy Weintraub This book was printed and bound by Hamilton Printing Company. The cover was printed by Hamilton Printing Company This book is printed on acid free paper. ∞ Copyright ©2011 John Wiley & Sons, Inc. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc. 222 Rosewood Drive, Danvers, MA 01923, website www.copyright.com. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030-5774, (201)748-6011, fax (201)748-6008, website http://www.wiley.com/go/permissions. “Evaluation copies are provided to qualified academics and professionals for review purposes only, for use in their courses during the next academic year. These copies are licensed and may not be sold or transferred to a third party. Upon completion of the review period, please return the evaluation copy to Wiley. Return instructions and a free of charge return shipping label are available at www.wiley.com/go/returnlabel. Outside of the United States, please contact your local representative.” Library of Congress Cataloging-in-Publication Data Necaise, Rance D. Data structures and algorithms using Python / Rance D. Necaise. p. cm. Includes bibliographical references and index. ISBN 978-0-470-61829-5 (pbk.) 1. Python (Computer program language) 2. Algorithms. 3. Data structures (Computer science) I. Title. QA76.73.P98N43 2011 005.13'3—dc22 2010039903 Printed in the United States of America 10 9 8 7 6 5 4 3 2 1
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540534443.68/warc/CC-MAIN-20191212000437-20191212024437-00428.warc.gz
|
CC-MAIN-2019-51
| 2,302
| 3
|
https://github.com/shiksha11/Jarvis4downloads
|
code
|
How to run file:
Once you have cloned the directory to your local machine in the default directory as per the operating system,follow the directions below:
1.Run the transfer_files.py.
2.After transfering files to the respective folders,install python-docx and pyaudio.
3.To insall pyaudio run the command : pip install pyaudio.
4.To install python-docx run the command : pip install pyhton-docx.
6.Input can be given in two ways.Either by tying it in the entry box or as voice input.
7.You can search file either by name or through body search(by typing words in the body of a particlar file. But the latter works only for .pdf , .csv , .txt , .docx files.
8.Click on "Search Path" button.
9.To give input through entry box : a) Type the input in the entry box.Two options will appear. Click on "Search by filename" or "Body Saerch" as per your input.
10.If you have entered the file name along with its extension, then the path of the most likely files will appear as output.
11.Else,you are required to choose the folder of file you are searching for,and then the path of the most prabable file will appear as output.
12.To give input in the form of audio first press "Give Input In Audio Form".
13.Then speak the file name.
14.Rest of the process is same as for the input given through entry box.
15.While giving input in audio form speak loudly and near the mike of your device.
How the Project Works:
The transfer_files.py file sits on the download folder and on execution sends the files to one of the folders among documents,pictures,videos and projects according to their extensions.
Group of extensions which will be placed in a particular folder is listed down in config.txt file.
After transfering files to the respective folders,install python-docx and pyaudio.
If the name of the entered file contains extensions then as per the extension by using os.walk we try finding the path.
If the entered file name does not contain extension then the user is required to select the type of file he/she is searching for.Further search is done in the folder based upon type of file selected by user.
For body search frequency of the word given as input is calculated in different files and file with highest frequency is returned as output.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103360935.27/warc/CC-MAIN-20220628081102-20220628111102-00319.warc.gz
|
CC-MAIN-2022-27
| 2,243
| 23
|
https://github.com/thoughtbot/kumade/commits/remove-more
|
code
|
You can clone with
HTTPS or Subversion.
Department of redundancy department.
Always load the railtie.
Move bundler-running into a function so we can reuse it.
Unescaping isn't necessary when we're not going through Cucumber.
We ignore app name anyway, so remove it.
Fix Railtie so that Kumade rake tasks work.
Remove all git remotes before and after scenarios.
Use straight ruby.
Better cucumber tag name.
Bump to 0.7.0 for working Cedar check.
CommandLine#run_or_error now returns the result of running the command.
This fixes the broken check for the Cedar stack.
Match the actual output of the heroku stack command.
Clean up heroku spec support.
Get to green.
Kumade now restarts Heroku app after db:migrate.
Use regular markdown.
Only run coverage if COVERAGE is set.
Be more explicit about what we're expecting.
Fix outputter spec for 1.8.7.
Install SimpleCov if on 1.9.x
No-op packager is never installed.
Bump to 0.6.0
Bump version to 0.5.0
Move outputter into configuration.
Default to the real outputter in specs.
Only delete branch if it exists.
Print error if something goes wrong in deploy.
WIP: actually use Outputter, and remove Base class.
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1429246639121.73/warc/CC-MAIN-20150417045719-00139-ip-10-235-10-82.ec2.internal.warc.gz
|
CC-MAIN-2015-18
| 1,154
| 31
|
https://ibsurvival.com/profile/205128-heidixu/
|
code
|
Hi guys, recently I'm deciding my biology EE topic. My research question is about the caffeine effect on metabolism rate and i really want to do the plant experiment, but I'm not sure for how difficult this experiment will be😂. Had anyone done the similar experiments before? or any suggestions? Thanks!!
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178363211.17/warc/CC-MAIN-20210302003534-20210302033534-00371.warc.gz
|
CC-MAIN-2021-10
| 307
| 1
|
https://linustechtips.com/topic/1496377-strange-cpu-behavior-while-gaming/
|
code
|
Strange CPU behavior while gaming.
Go to solution Solved by jchenabc,
FOUND THE CAUSE!
I had my game installed on a different drive than the boot drive. The high CPU usage happens when the game is compiling shaders. It is suppose to load shaders only for first launch. However, for whatever reason, AMD shader cache doesn't work when the game is not install on the boot drive.
After I move the game to the boot drive, cleared shaders, i no longer have the issue after it compile the shaders.
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!Register a new account
Already have an account? Sign in here.Sign In Now
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224644913.39/warc/CC-MAIN-20230529205037-20230529235037-00616.warc.gz
|
CC-MAIN-2023-23
| 728
| 10
|
http://www.coderanch.com/t/551991/Ranch-Office/Posting-articles
|
code
|
This week's book giveaway is in the
We're giving away four copies of a choice of
"Take Control of Upgrading to Yosemite" or "Take Control of Automating Your Mac"
and have Joe Kissell on-line!
A friendly place for programming greenhorns!
Big Moose Saloon
Register / Login
Win a copy of
Take Control of Upgrading to Yosemite
Take Control of Automating Your Mac
this week in the
Joined: Aug 25, 2003
Sep 08, 2011 09:13:42
Can you suggest some good sites to post general technical articles on Java. I am looking for sites like javaworld, theserverside, etc.
Joined: Sep 08, 2007
Sep 08, 2011 10:47:47
Mohamed Sanaulla |
subject: Posting articles
your text columns are way too wide
Android vs BlackBerry
Where to start off
Relationship in BMP
All times are in JavaRanch time: GMT-6 in summer, GMT-7 in winter
| Powered by
Copyright © 1998-2014
|
s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1412037663754.5/warc/CC-MAIN-20140930004103-00369-ip-10-234-18-248.ec2.internal.warc.gz
|
CC-MAIN-2014-41
| 839
| 25
|
https://rdrr.io/github/Bioconductor-mirror/MineICA/man/plotMclust.html
|
code
|
Given a result of function Mclust applied on a numeric
vector, this function add the fitted Gaussian to a
previous plot. This is an internal function called by
The result of Mclust function applied to
The vector of numeric values on which was applied Mclust
This function can only deal with at the most three Gaussian.
1 2 3 4 5 6 7
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583668324.55/warc/CC-MAIN-20190119135934-20190119161934-00375.warc.gz
|
CC-MAIN-2019-04
| 449
| 9
|
http://webrtc.sys-con.com/node/2891796
|
code
|
|By Pat Romanski||
|January 3, 2014 07:15 AM EST||
This insight looks into the developments of communication services, from various angles such as voice, video, messaging and social networks. It includes market figures and forecasts for the US and EU5, assessing the real impacts (or lack of it) of OTTs on what is predominantly a telco market. The opportunities for telcos and their relationships with the OTTs are examined and analysed.
This deliverable is part of the Telco vs OTT watch which covers half-yearly updated datasets- Half-yearly updated status reports, a Net Neutrality benchmark on 15 countries and 3-sided business models report, quarterly market insights, and direct access to lead OTT analysts
2. Main concepts of communication services 7
2.1. Communication service segments 7
2.1.1. Voice (including video) 7
2.1.2. Messaging 8
2.1.3. Social networks 8
2.2. Communication services pricing models 9
3. Communication Markets 10
3.1. Voice 10
3.1.1. Market size 10
3.1.2. Market structure and ecosystem 10
3.2. Messaging 11
3.2.1. Market size 11
3.2.2. Market structure and ecosystem 11
3.3. Social networks 12
3.3.1. Market size 12
3.3.2. Market structure and ecosystem 13
4. Telcos and OTT strategies 15
4.1. Overview 15
4.2. OTT players 16
4.2.1. Facebook 17
4.2.2. Google 19
4.2.3. Apple 20
4.2.4. Microsoft (Skype) 22
4.2.5. WhatsApp 24
4.2.6. LINE 25
4.3. Telcos 26
4.3.1. Orange 27
4.3.2. Telefónica 29
4.3.3. BT 30
4.3.4. T-Mobile 31
4.3.5. Verizon 33
4.3.6. Joyn (by the GSMA) 33
5. Communication services and three-sided business models 35
5.1. Direct Competition 35
5.2. OTT as intermediary between users and others 36
5.3. Cooperation between telcos and large OTTs 37
5.4. Transformation of telcos into OTTs 38
5.5. Telcos as intermediary between users and large OTTs 39
5.6. Telcos as intermediary between users and others 39
5.7. Conclusions 41
6. Appendix 43
6.1. Beyond two-sided models: three-sided models! 43
6.1.1. OTT as intermediate between 'users' and 'others' 44
6.1.2. Cooperation between telcos and large OTTs 45
6.1.3. Transformation of telcos into OTTs 46
6.1.4. Telco as intermediary between 'users' and large OTTs 46
6.1.5. Telco as intermediary between 'users' and 'others' 47
Table 1: Payment services provided by OTT and telcos 15
Table 2: Monthly active users of OTT communication services 16
Table 3: Telco positioning on offering beyond traditional communication 27
WebRTC: together these advances have created a perfect storm of technologies that are disrupting and transforming classic communications models and ecosystems. In his session at WebRTC Summit, Cary Bran, VP of Innovation and New Ventures at Plantronics and PLT Labs, will provide an overview of this technological shift, including associated business and consumer communications impacts, and opportunities it may enable, complement or entirely transform.
Oct. 10, 2015 02:15 AM EDT Reads: 765
WebRTC services have already permeated corporate communications in the form of videoconferencing solutions. However, WebRTC has the potential of going beyond and catalyzing a new class of services providing more than calls with capabilities such as mass-scale real-time media broadcasting, enriched and augmented video, person-to-machine and machine-to-machine communications. In his session at @ThingsExpo, Luis Lopez, CEO of Kurento, will introduce the technologies required for implementing these ideas and some early experiments performed in the Kurento open source software community in areas ...
Oct. 10, 2015 01:00 AM EDT Reads: 775
Canary launches are a best practice in continuous delivery, used by teams at Google, Amazon and LinkedIn. A canary launch is rolling out features to a small number of users to assess the overall system reaction. Canary releases allow software teams to produce better software, quicker. In his session at WebRTC Summit, John Kodumal, CTO and cofounder of LaunchDarkly, will discuss the benefits of canary launches, as well as how you can use canary launches in your own development life cycle.
Oct. 9, 2015 11:00 PM EDT Reads: 307
Who are you? How do you introduce yourself? Do you use a name, or do you greet a friend by the last four digits of his social security number? Assuming you don’t, why are we content to associate our identity with 10 random digits assigned by our phone company? Identity is an issue that affects everyone, but as individuals we don’t spend a lot of time thinking about it. In his session at @ThingsExpo, Ben Klang, Founder & President of Mojo Lingo, will discuss the impact of technology on identity. Should we federate, or not? How should identity be secured? Who owns the identity? How is identity ...
Oct. 9, 2015 10:00 PM EDT Reads: 453
The Internet of Things (IoT) is growing rapidly by extending current technologies, products and networks. By 2020, Cisco estimates there will be 50 billion connected devices. Gartner has forecast revenues of over $300 billion, just to IoT suppliers. Now is the time to figure out how you’ll make money – not just create innovative products. With hundreds of new products and companies jumping into the IoT fray every month, there’s no shortage of innovation. Despite this, McKinsey/VisionMobile data shows "less than 10 percent of IoT developers are making enough to support a reasonably sized team....
Oct. 9, 2015 04:00 PM EDT Reads: 253
"Matrix is an ambitious open standard and implementation that's set up to break down the fragmentation problems that exist in IP messaging and VoIP communication," explained John Woolf, Technical Evangelist at Matrix, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Oct. 9, 2015 07:00 AM EDT Reads: 5,893
WebRTC converts the entire network into a ubiquitous communications cloud thereby connecting anytime, anywhere through any point. In his session at WebRTC Summit,, Mark Castleman, EIR at Bell Labs and Head of Future X Labs, will discuss how the transformational nature of communications is achieved through the democratizing force of WebRTC. WebRTC is doing for voice what HTML did for web content.
Oct. 9, 2015 06:00 AM EDT Reads: 1,421
Internet of Things (IoT) will be a hybrid ecosystem of diverse devices and sensors collaborating with operational and enterprise systems to create the next big application. In their session at @ThingsExpo, Bramh Gupta, founder and CEO of robomq.io, and Fred Yatzeck, principal architect leading product development at robomq.io, discussed how choosing the right middleware and integration strategy from the get-go will enable IoT solution developers to adapt and grow with the industry, while at the same time reduce Time to Market (TTM) by using plug and play capabilities offered by a robust IoT ...
Oct. 9, 2015 02:00 AM EDT Reads: 2,231
“In the past year we've seen a lot of stabilization of WebRTC. You can now use it in production with a far greater degree of certainty. A lot of the real developments in the past year have been in things like the data channel, which will enable a whole new type of application," explained Peter Dunkley, Technical Director at Acision, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Oct. 9, 2015 01:45 AM EDT Reads: 7,039
Through WebRTC, audio and video communications are being embedded more easily than ever into applications, helping carriers, enterprises and independent software vendors deliver greater functionality to their end users. With today’s business world increasingly focused on outcomes, users’ growing calls for ease of use, and businesses craving smarter, tighter integration, what’s the next step in delivering a richer, more immersive experience? That richer, more fully integrated experience comes about through a Communications Platform as a Service which allows for messaging, screen sharing, video...
Oct. 9, 2015 12:00 AM EDT Reads: 1,139
Can call centers hang up the phones for good? Intuitive Solutions did. WebRTC enabled this contact center provider to eliminate antiquated telephony and desktop phone infrastructure with a pure web-based solution, allowing them to expand beyond brick-and-mortar confines to a home-based agent model. It also ensured scalability and better service for customers, including MUY! Companies, one of the country's largest franchise restaurant companies with 232 Pizza Hut locations. This is one example of WebRTC adoption today, but the potential is limitless when powered by IoT.
Oct. 8, 2015 04:30 PM EDT Reads: 7,484
WebRTC is about the data channel as much as about video and audio conferencing. However, basically all commercial WebRTC applications have been built with a focus on audio and video. The handling of “data” has been limited to text chat and file download – all other data sharing seems to end with screensharing. What is holding back a more intensive use of peer-to-peer data? In her session at @ThingsExpo, Dr Silvia Pfeiffer, WebRTC Applications Team Lead at National ICT Australia, will look at different existing uses of peer-to-peer data sharing and how it can become useful in a live session to...
Oct. 8, 2015 12:00 PM EDT Reads: 611
WebRTC has had a real tough three or four years, and so have those working with it. Only a few short years ago, the development world were excited about WebRTC and proclaiming how awesome it was. You might have played with the technology a couple of years ago, only to find the extra infrastructure requirements were painful to implement and poorly documented. This probably left a bitter taste in your mouth, especially when things went wrong.
Oct. 8, 2015 06:00 AM EDT Reads: 767
When it comes to IoT in the enterprise, namely the commercial building and hospitality markets, a benefit not getting the attention it deserves is energy efficiency, and IoT’s direct impact on a cleaner, greener environment when installed in smart buildings. Until now clean technology was offered piecemeal and led with point solutions that require significant systems integration to orchestrate and deploy. There didn't exist a 'top down' approach that can manage and monitor the way a Smart Building actually breathes - immediately flagging overheating in a closet or over cooling in unoccupied ho...
Oct. 6, 2015 05:00 PM EDT Reads: 331
With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo in Silicon Valley. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be! Internet of @ThingsExpo, taking place Nov 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 17th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The Internet of Things (IoT) is the most profound change in personal an...
Sep. 19, 2015 09:00 AM EDT Reads: 2,398
The 3rd International WebRTC Summit, to be held Nov. 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA, announces that its Call for Papers is now open. Topics include all aspects of improving IT delivery by eliminating waste through automated business models leveraging cloud technologies. WebRTC Summit is co-located with 15th International Cloud Expo, 6th International Big Data Expo, 3rd International DevOps Summit and 2nd Internet of @ThingsExpo. WebRTC (Web-based Real-Time Communication) is an open source project supported by Google, Mozilla and Opera that aims to enable bro...
Sep. 17, 2015 09:45 AM EDT Reads: 1,979
All major researchers estimate there will be tens of billions devices - computers, smartphones, tablets, and sensors - connected to the Internet by 2020. This number will continue to grow at a rapid pace for the next several decades. With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo, November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be.
Sep. 13, 2015 05:00 PM EDT Reads: 940
"We have seen the evolution of WebRTC right from the starting point to what it has become today, that people are using in real applications," noted Dr. Natasha Tamaskar, Vice President and Head of Cloud and Mobile Strategy and Ecosystem at GENBAND, in this SYS-CON.tv interview at WebRTC Summit, held June 9-11, 2015, at the Javits Center in New York City.
Jul. 22, 2015 05:00 PM EDT Reads: 1,745
The WebRTC Summit 2015 Silicon Valley, to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, announces that its Call for Papers is now open. Topics include all aspects of improving IT delivery by eliminating waste through automated business models leveraging cloud technologies. WebRTC Summit is co-located with 17th International Cloud Expo, @ThingsExpo, Big Data Expo, and DevOps Summit. WebRTC is the future of browser-to-browser communications, and continues to make inroads into the traditional, difficult, plug-in web communications world.
Jun. 10, 2015 11:15 PM EDT Reads: 2,720
The 4th International Internet of @ThingsExpo, co-located with the 17th International Cloud Expo - to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA - announces that its Call for Papers is open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
Jun. 10, 2015 06:00 PM EDT Reads: 3,450
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-40/segments/1443737946509.72/warc/CC-MAIN-20151001221906-00183-ip-10-137-6-227.ec2.internal.warc.gz
|
CC-MAIN-2015-40
| 13,654
| 94
|
https://stripe.com/docs/identity/explaining-identity
|
code
|
Explain Identity to your customers
Answer customer questions about ID verification and Stripe Identity.
If you use Stripe Identity for ID verification, copy and customize these questions and answers to create Frequently Asked Questions (FAQs) around ID verification. To obtain Stripe logos, badges, and buttons for your site, visit the Media assets section on stripe.com.
In your Stripe Identity FAQ, address the following questions by using the provided pre-approved copy:
How does identity verification work?
What are the best practices for a successful verification?
Who has access to my verification data?
For the following questions, provide your preferred answer:
|Why am I asked to verify my identity?||Provide your preferred answer. Some users might be hesitant to share their ID information, so it’s important to help them understand why you’re asking for this information|
|Why was I rejected?||You might want to offer alternative methods for verification if a user disputes their results.|
|Can I get verified using a different method?||Privacy laws might require you to provide an alternative verification process that doesn’t use biometric technology if the user doesn’t consent to use of their biometric information. Consult your legal counsel for regional requirements.|
How can I access or delete my verification data?
Provide your data privacy process.
The Identity API has a redaction endpoint that allows you to delete the verification data that Stripe Identity stores on behalf of your business. For example, you can use this tool to meet your deletion requirements when an end-user from Europe or California asks you to delete their data, or when you collect an ID from a country such as Germany that requires you to delete an ID card upon completion of the verification even if there’s no deletion request from the end user. If you’ve created additional copies of a user’s data, you might also need to delete these as well.
Stripe doesn’t delete data on your behalf when we store the data as your processor, even if your end-user asks us to, because we recognize you must conduct your own legal analysis on whether deletion is appropriate.
If your end-user reaches out to us requesting deletion, we’ll respond to the request with respect to any data that we hold as data controller, and also recommend the end-user reach out to you to request deletion.
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100781.60/warc/CC-MAIN-20231209004202-20231209034202-00888.warc.gz
|
CC-MAIN-2023-50
| 2,392
| 16
|
http://www.sheepwalkelectronics.co.uk/1w_software.php?osCsid=959a7d2026f494b1f583f4712c1b7275
|
code
|
The following are links to software suitable for use with a 1-Wire network. They are all links to external sites over which we have no control. They will all open in a new window/tab.
The iButton 1-Wire Viewer
is a simple Java application from Maxim that is useful for testing and debugging your network.
, or the "1-Wire File System" is a fantastic piece of software that supports almost all 1-Wire host adapters and devices. It is the only software we recommend for use with our RPI2 and RPI3 host adapters.
is a piece of Linux or Windows software that will read and graph data from DS18B20 and DS18S20 sensors. It also supports humidity sensors including our SWE3.
is weather monitoring software for Linux or RiscOS that includes an Android client.
is a comprehensive weather monitoring package for Linux, OSX and Windows.
is a Windows based software package for weather monitoring
1-Wire monitoring software. Windows software for weather monitoring, provides 9 days graphs, supports DS2438 and DS2423.
is a Home Automation System with some 1-Wire support.
|
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917121453.27/warc/CC-MAIN-20170423031201-00111-ip-10-145-167-34.ec2.internal.warc.gz
|
CC-MAIN-2017-17
| 1,059
| 10
|
https://github.com/timhutton/mandelstir
|
code
|
John Baez posted something on Google+: "This movie shows the sense in which Julia sets are self-similar"
It shows an animation that Anders Kaseorg made in answer to a Quora question about why Julia Sets are fractal.
Here's some more versions of the same idea:
The nice thing about this approach is that it explains how the Julia set works. Compare with the diagrams of the affine transforms in Iterated Function Systems. The key idea in both cases is that the fractal maps onto a smaller version of itself.
The z2 transform provides the 'stirring' motion but we don't normally see it. Even this (very good) video showing the orbits doesn't make this clear. Anders' trick for showing all the intermediate steps of a single iteration is wonderful.
The obvious thing is to make a similar animation for the Mandelbrot. As John points out, it's not as simple, because the set doesn't map directly onto itself every iteration. But it is still very pleasing!
Here's the same thing shown on a checkerboard instead of on the set itself. This makes it clearer that it's the transform that gives rise to the shape of the set.
Another way of animating this is to use a simple linear interpolation between each iteration. We lose the mathematical connection to z2 being a 'stirring' movement but it gives a simpler behavior:
This the closest I've come so far to the original goal of visualizing how the Mandelbrot works.
Compare with the images of integer iterations. They show the same thing but it's not obvious how to get from one frame to the next:
→ → ?
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195527089.77/warc/CC-MAIN-20190721164644-20190721190644-00463.warc.gz
|
CC-MAIN-2019-30
| 1,549
| 11
|
http://arash.pourhabibi.info/
|
code
|
My name is Arash (Persian: آرش) (aka NEO), and I am from Iran. I am a PhD candidate at the School of Computer and Communication Sciences, EPFL, and a doctoral research assistant at PARSA under supervision of Prof. Babak Falsafi. I did my master’s in Software Engineering at the Department of CSE and IT of Shiraz University, where I also received my B.Sc. in Software Engineering.
I am broadly interested in the field of systems and interdisciplinary systems problems found in modern, large-scale datacenters. My current research revolves around efficiency and sustainability enhancement of server systems for datacenters. By characterizing modern server workloads and employing emerging memory subsystems, I look for minimizing the energy footprint and maximizing the compute density of server systems through specialization of various system components, and better system integration.
Another thing I have been recently up to is translating “Engineering SaaS” book written by Armando Fox and David Patterson into Persian. This is officially supported by the writers and I’m really excited about this. You can find more information about it here.
You can access my CV from the sidebar.
Thanks for stopping by!
|
s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570987833766.94/warc/CC-MAIN-20191023122219-20191023145719-00304.warc.gz
|
CC-MAIN-2019-43
| 1,221
| 5
|
https://forums.adobe.com/thread/1538299
|
code
|
I have PS CS6 and just updated to Win 8.1. Now my CS6 is not working correctly. Bridge often comes up "Not Responding" and PS blinks with a b & w checker board while working within a photo. I believe this to be a settings issue, however been unable to figure it out. Can anyone help?
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267161098.75/warc/CC-MAIN-20180925044032-20180925064432-00173.warc.gz
|
CC-MAIN-2018-39
| 283
| 1
|
https://forum.muffingroup.com/betheme/discussion/58555/header
|
code
|
In the two pages I've created for the privacy and cookies I do not want put any slider but it remains an empty space.
Could you kindly let me know how to adjust it?
See attached screenshot to better explain what i mean.
My site is https://ristorantecostadamalfi.it/cookie-policy
Thanks in advance
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947476396.49/warc/CC-MAIN-20240303142747-20240303172747-00269.warc.gz
|
CC-MAIN-2024-10
| 296
| 5
|
https://community.st.com/thread/8791-small-device-c-compiler-360-released
|
code
|
Version 3.6.0 of SDCC, a C implementation targeting 8-bit architectures including the STM8 was released a few hours ago:
Today a new release of SDCC was made. We are now at version 3.6.0.
You can get it at:
So what's new?
* Merged upstream binutils 2.25
* New memory management with lower overhead
* Changed default language dialect to --std-sdcc11
* Diagnostic for missing type specifier: No implicit int outside of C90
* C11 generic selections
* char type is now unsigned by default (old behaviour can be restored
* Character constants are now of type int instead of char.
* ISO C95 and ISO C11 wide character constants
* ISO C95 and ISO C11 wide string literals
* Basic standard library support for wide characters: c16rtomb(),
mbrtoc16(), mbsinit(), mbtowc(), mbrlen(), mbrtoc32, c32rtomb(),
mbrtowc(), wcrtomb(), mblen(), wctomb()
* Treat all ports the same in the manual (i.e. mcs51-specific stuff is now
clearly described as such)
* Reorganized interrupt handling for z80, z180, r2k, r3ka, tlcs90, gbz80
* Workaround for stm8 division hardware bug
* ELF/DWARF support for stm8
* Output symbol table for ELF
* pic16 port now uses standard-compliant crt0iz that initializes static
and globals to 0 by default
And of course numerous feature requests and bug fixes are included as well.
Once again I hope you will enjoy using this new release.
SDCC 3.6.0 Release Manager
Most important to STM8 users are probably the improved standard support and various small fixes and improvements in the STM8 backend.
A current comparison of C compilers for the STM8 can be found at http://colecovision.eu/stm8/compilers.shtml.
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794865651.2/warc/CC-MAIN-20180523121803-20180523141803-00583.warc.gz
|
CC-MAIN-2018-22
| 1,617
| 29
|
https://blogs.oracle.com/OTNGarage/tags/lake_baikal
|
code
|
By RickRamsey-Oracle on Dec 18, 2012
This is the lake where Yuli Vasiliev's countrymen swim laps.
Yuli is one of my favorite OTN writers not just because he really knows his stuff. Not just because his writing is clear and accurate. And not just because his English is better than the English of most native speakers. Yo, those are all good reasons. But it's the Lake Baikal thing.
Yuli recently wrote two wicked good how-to's about Oracle VM Templates. You should read them. You might gain a gram of Yuli's respect. Two grams, if you can head butt icebergs while you swim.
How to prepare an Oracle VM environment to use Oracle VM Templates, how to obtain a template, and how to deploy the template to your Oracle VM environment. Also how to create a virtual machine based on that template and how you can clone the template and change the clone's configuration.
How to use Oracle VM VirtualBox Templates in Oracle VM VirtualBox. Similar to the article above, but it describes how to download, install, and configure the templates within Oracle VM VirtualBox, instead of on bare metal.
Other OTN Technical Articles by Yuli Vasiliev
- Retrieving, Transforming, and Consolidating Web Data with Oracle Database
- Setting Up, Configuring, and Using a WebLogic Server Cluster
- Cube Development for Beginners
- How to XQuery Non-JDBC Sources from JDBC
- Advanced Dimensional Design with Oracle Warehouse Builder
- Using the JDBC Connectivity Layer in Oracle Warehouse Builder
- High Performance Oracle JDBC Programming
- Python Data Persistence with Oracle
- Querying JPA Entities with JPQL and Native SQL
|
s3://commoncrawl/crawl-data/CC-MAIN-2016-36/segments/1471982290442.1/warc/CC-MAIN-20160823195810-00054-ip-10-153-172-175.ec2.internal.warc.gz
|
CC-MAIN-2016-36
| 1,600
| 16
|
http://www.binary-asylum.co.uk/profiles/blogs/bangkok-games-jam-2010
|
code
|
I'm out in Bangkok at the moment running the Bangkok Games Jam event on behalf of the British Council as part of the Bangkok Design Festival. We have 40 students battling it out for the kudos of being able to create the best game they can in just 48 hours!
I have a number of UK games industry professionals out here with me helping me run the event.
If you would like to follow the game jam back in the UK, there are a number of ways:
Follow the following people on Twitter: @davywavy @manak @g_bry @we_are_twisted
Live Video: http://www.blogtv.com/People/BangkokGameJam
I'll be adding updates to this forum thread every few hours also.
If anyone has the time and inclination and would like to help a team out then email me at email@example.com telling me if you are artist/designer/programmer and I will put you directly in touch with a team that needs help in your skills area. Please be aware that Thailand is 7 hours ahead of the UK, so please think carefully before volunteering your assistance, as the team will need to you to continue to contribute until the end of the jam.
Game jamming is an enormously valuable experience for student game developers. I hope you get a chance to have a look in and see what's going on, contribute, ask questions, chat with your Thai peers and play some of the games that result from the Bangkok Games Jam
|
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368700264179/warc/CC-MAIN-20130516103104-00057-ip-10-60-113-184.ec2.internal.warc.gz
|
CC-MAIN-2013-20
| 1,347
| 8
|
https://frameboxxindore.com/apple/what-is-the-best-virtual-machine-for-linux.html
|
code
|
Can Linux run virtual machines?
If you’re using Linux, you don’t need VirtualBox or VMware to create virtual machines. You can use KVM — the kernel-based virtual machine — to run both Windows and Linux in virtual machines.
Which is best virtual machine?
The best virtual machine for Windows 10
- VMware Workstation Pro and Workstation Player.
- VMware ESXi.
- Microsoft Hyper-V.
- VMware Fusion Pro and Fusion Player.
Is VirtualBox better on Linux?
Fact: You will get better performance from any VM running on Linux, than you will running on Windows. Fact: Your preference for interface, and program “feel” may override any, or all of this.
Should I use virtual machine for Linux?
VMs also can be useful for those non-Linux users who want to play with Linux or who want to migrate to it but don’t want to jump shark by formatting the OS they were used to and switching to Linux. Virtual machines get these users comfortable with Linux, so they can make the switch with confidence when they are ready.
Is Ubuntu good for virtual machine?
VMware is a non-free virtual machine application, which supports Ubuntu as both a host and guest operating system. Several versions of VMware are available at no cost and can be installed on Ubuntu. … VMWare is the virtual machine solution that has been in use the longest and is the most widely used.
How much RAM do I need for a virtual machine?
8 GB RAM should be good for most situations. With 4 GB you can have a problem, depending on what you intend to do with the client OS and what else the host will be used for. Most client operating systems will need at least 1 GB RAM but that only for light use. Modern versions of Windows will want more.
Is Hyper-V faster than VirtualBox?
Hyper-V is designed to host servers where you don’t need a lot of extra desktop hardware (USB for example). Hyper-V should be faster than VirtualBox in a lot of scenarios. You get things like clustering, NIC teaming, live migration, etc that you’d expect from a server product.
Which is faster VirtualBox or VMware?
Answer: Some users have claimed that they find VMware to be faster as compared to VirtualBox. Actually, both VirtualBox and VMware consume a lot of resources of the host machine. Therefore, the physical or hardware capabilities of the host machine are, to a great extent, a deciding factor when virtual machines are run.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446711162.52/warc/CC-MAIN-20221207121241-20221207151241-00400.warc.gz
|
CC-MAIN-2022-49
| 2,380
| 20
|
https://www.technojobs.co.uk/artificial-intelligence-consultant-jobs
|
code
|
Artificial Intelligence Engineer / Data Scientist £45k - £90k dependant on experience + bonus + good bens. Flexible Working Location opportunity
This role may suit individuals who have perhaps previously held the following role titles: Data Engineer, Data Architect, Big Data Consultant, Data Scientist, Data Modeller, Big Data Analyst, AI Engineer
We have been asked to...
Date:Location: London Salary/Rate: £36,000 - £90,000
|
s3://commoncrawl/crawl-data/CC-MAIN-2018-43/segments/1539583508988.18/warc/CC-MAIN-20181015080248-20181015101748-00364.warc.gz
|
CC-MAIN-2018-43
| 430
| 4
|
https://watchtechmarket.com/kubernetes-administration-policy-made-easy-with-brewopa/
|
code
|
Kubernetes administration policy made easy with brewOPA
Cloud-native computing — with such applied sciences as Kubernetes, service-mesh, and steady integration and steady supply (CI/CD) — is revolutionizing IT. But managing can nonetheless be a serious ache within the server. That’s the place Open Policy Agent (OPA), an open-source Cloud Native Computing Foundation (CNCF) challenge, is available in. But it has its personal steep studying curve. Cyral with brewOPA needs to ease their climb and make managing insurance policies throughout cloud-native platforms a lot simpler.
OPA’s very talked-about as a result of it permits policy analysis to be decoupled from an utility’s core enterprise logic. This means your policy engine internals are abstracted out, so you’ll be able to simply reuse them throughout a number of parts.[Source: ZDNet]
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703514796.13/warc/CC-MAIN-20210118123320-20210118153320-00797.warc.gz
|
CC-MAIN-2021-04
| 858
| 3
|
https://lanet.co.uk/services/lanet-cloud-adoption/
|
code
|
LA NET is a Gold-certified Cloud Platform Microsoft partner with a solid understanding of designing and building cloud infrastructure at scale on the Azure platform. We follow the Cloud Adoption Framework for Microsoft Azure (CAF). With our experience, we can help to ensure our customers can adopt the cloud quickly but safely, following best practices.
Using the CAF framework and guidelines, we can take our customers through a structured and repeatable process to ensure that nothing is missed, and that current best practices are in place. Through our managed service, we can adapt and update the environment to take advantage of changes in the platform and new best practice guidelines.
The Cloud Adoption Framework for Microsoft Azure Lifecycle
Strategy and Planning
It is crucial to understand the required business outcomes & motivations. This includes identifying and working with the relevant internal and external stakeholders. Here we can gather requirements, develop an action plan and define responsibilities.
Prepare Your Cloud For Success
Now the strategy is understood and an action plan is in place, it is time to prepare the cloud environment. Here, we design and implement core network and user access controls, guardrails, management systems and processes.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474700.89/warc/CC-MAIN-20240228080245-20240228110245-00018.warc.gz
|
CC-MAIN-2024-10
| 1,278
| 7
|
http://www.rasteredge.com/gallery/c39/92/
|
code
|
Cisco IronPort AsyncOS 7.5 for Email Daily Management Guide
Chapter 2 Using Email Security Monitor
Email Security Monitor Pages
Incoming Mail Details Listing
The top senders which have connected to public listeners of the appliance are
listed in the External Domains Received listing table at the bottom of the
Incoming Mail page, based on the view selected. Click the column headings to
sort the data. See Categorizing Email, page 2-9 for an explanation of the various
The system acquires and verifies the validity of the remote host’s IP address (that
is, the domain) by performing a double DNS lookup. For more information about
double DNS lookups and sender verification, see the “Configuring the Gateway
to Receive Email” chapter in the Cisco IronPort AsyncOS for Email Configuration
The Sender Detail listing has two views, Summary and All.
The default Sender Detail view shows the total number of attempted messages for
each sender, and includes a breakdown by category (the same categories as the
Incoming Mail Summary graph on the Overview page: number of clean messages,
stopped by reputation filtering, invalid recipients, spam detected, virus detected,
stopped by content filter). It also shows the total number of threat messages
(messages stopped by reputation or stopped as invalid recipient, spam, and
The value for Stopped by Reputation Filtering is calculated based on several
- Number of “throttled” messages from this sender.
- Number of rejected or TCP refused connections (may be a partial count).
- A conservative multiplier for the number of messages per connection.
When the appliance is under heavy load, an exact count of rejected connections is
not maintained on a per-sender basis. Instead, rejected connections counts are
maintained only for the most significant senders in each time interval. In this
situation, the value shown can be interpreted as a “floor”; in other words, at least
this many messages were stopped.
The Stopped by Reputation Filtering total on the Overview page is always based
on a complete count of all rejected connections. Only the per-sender connection
counts are ever limited due to load.
Additional columns that you can display are:
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882570868.47/warc/CC-MAIN-20220808152744-20220808182744-00153.warc.gz
|
CC-MAIN-2022-33
| 2,204
| 32
|
http://tomvangaever.be/blogv2/2013/04/
|
code
|
https://play.spotify.com allows you to run spotify in your browser, meaning you can run the application on the Surface RT even if the spotify app is not available (yet?)
If you do not have the required cookie on your machine to participate in this beta, you are redirected to spotify.com.
The key here is to get the cookie or open the following site: http://www.spotifyweb.tk/
This evening I noticed that IE became not supported, it is possible to change the user agent string so that you are able to run spotify in your browser on the Surface RT.
Select the gear icon in the upper-right corner, select F12 developer tools > tools > Change user agent string to Chrome…
Hope it helps!
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780056120.36/warc/CC-MAIN-20210918002951-20210918032951-00405.warc.gz
|
CC-MAIN-2021-39
| 685
| 6
|
https://theagileplayground.com/team-improvement/
|
code
|
Way of Working Improvements – Workshop
One of my scrum teams was eager to improve their Way of Working, and wanted to find out what improvements they could agree on to start with. So I created a whiteboard with possible improvements, based on observations and one-on-ones with team members the last months. I gave them the opportunity to add their own ideas, and used the Liberating Structure 124ALL as a method to determine a top 3 supported by all team members.
WoW Improvements session – how to do it?
In Mural I created a template, pre-filled with the improvement ideas. So if you would like to do this session with your team, please find the Mural template here.
The only thing you need to do, is fill in your team members’ names, and add your own ideas that your team might want to improve on, and maybe remove topics that are irrelevant for your team.
Instructions for your team
Give your team the following instructions.
- Individually make a top 3 of the improvements you consider most important for the team (you can use the empty post-its if you have ideas that are not on this list).
2. Make pairs. Discuss both your top 3’s and agree on a new combined top 3.
4. Discuss your top 3’s in foursomes and again, agree on a combined top 3.
ALL. Explain your top 3’s to each other as a team and agree on a new combined team top 3. If necessary, vote.
Once you have a supported team top 3, define at least one first practical step you will take to start improving.
|
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296818740.13/warc/CC-MAIN-20240423192952-20240423222952-00035.warc.gz
|
CC-MAIN-2024-18
| 1,481
| 12
|
https://hackaday.io/project/181676-sol-20-reproduction/log/200233-wiring-the-keyboard
|
code
|
The keyboard encoder is expecting 5V while the Raspberry Pi 4 operates at 3.3V. So to overcome this I purchase a Voltage-Level Shifter Module from Amazon. I also printed a "caddy" to hold the Pi 4 in place and added a small 30 mm x 30 mm x 10 mm blower fan for good measure to keep things cool. The fan I used is from Amazon: GDSTIME 3cm 30mm x 10mm 5V DC Brushless Small Blower Cooling Fan, with Dual Ball Bearings. The fan and the keyboard will be run off of the Pi's power supply. The Pi is secured in place to the caddy with two sided tape.
I then mounted this unit onto the back support of the Sol-20 frame again using two sided tape. The USB and HDMI ports are facing to the rear of the unit.
So we are ready to wire the keyboard to the Pi. Here is what the Sol-20 header pinout looks like.
And the Raspberry Pi level shifter hat.
So here is how I wired the keyboard. Note that for the exception of +5V and GND lines which are wired to the 3.3V side or the level shifter, all of the other connections are wired to the 5V side.
|Keyboard Encoder||Raspberry Pi||Description|
|D0||GPIO6||Key 0 bit (low)|
|D1||GPIO13||Key 1 bit|
|D2||GPIO19||Key 2 bit|
|D3||GPIO26||Key 3 bit|
|D4||GPIO21||Key 4 bit|
|D5||GPIO20||Key 5 bit|
|D6||GPIO16||Key 6 bit|
|D7||GPIO12||Key 7 bit (high)|
|STROBE||GPIO5||Key ready on falling edge.|
|OUT1||GPIO23||LOCAL button. Toggle. HI when LED on.|
|OUT2||GPIO22||BREAK button pressed when rising edge.|
|OUT3||GPIO24||RESET button pressed when falling edge.|
Here is what the wiring looks like.
I'm investigating a BREAK key issue with Dave, but otherwise everything is working great. The keyboard is now integrated with the emulator. Getting close. Very exciting.
|
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320302622.39/warc/CC-MAIN-20220120190514-20220120220514-00642.warc.gz
|
CC-MAIN-2022-05
| 1,697
| 20
|
https://forums.unrealengine.com/t/geolocation-access-geographic-location-data-on-ios-android-and-html5/82558
|
code
|
[FONT=Arial Black]MARKETPLACE LINK
iOS, Android and HTML5
The plugin abstracts the native platforms API for supported platforms into a single component that can be easily controlled using both Blueprint and C++ in UE4. Multiple Geo Location Components are supported to allow access to location data from many different actors. Location results and their accuracy depend on the platform and device hardware.
When running in the Unreal Editor a set of global functions are included that can set a simulated location for testing and development.
Feel free to ask any questions you might have about this plugin.
I’ve just updated GeoLocation to version 1.1
The changes included in this release are as follows:
Geo Location Component
A new property RespondsToGeoFences has been added which defaults to false. Setting this to true will make the Geo Location Component respond to Geo Fence Components.
Two new Blueprint/C++ events have been added that can be used to interact with Geo Fence Components when RespondsToGeoFence is set to true: OnGeoFenceBeginOverlap, OnGeoFenceEndOverlap. These events provide the Geo Fence Component that has been entered/left as a parameter.
Geo Location Position
- Added CalculateDistanceTo function, which uses the CalculateDistanceBetween utility function for this position and another position. Also added GetGeoCoordinateStruct that returns a FGeoCoordinate struct with the latitude and longitude in this Geo Location Position.
If anyone has any questions about these features, or wishes to request future features they would like to see let me know here, in PMs or via email.
Does this allow to know the compass values too or only GPS coordinates?
I’m using your plugin (maybe I’m using it wrong) and it crashes my project.
I create a new project (4.21.1) and it works fine but when I add the plugin it gives me some errors:
LogPlayLevel: Error: clang++.exe: error: linker command failed with exit code 1 (use -v to see invocation)
LogPlayLevel: Error: ERROR: UBT ERROR: Failed to produce item: C:\Users\etien\Documents\Unreal Projects\TestLocationIfCrash1\Binaries\Android\TestLocationIfCrash1-armv7-es2.so
LogPlayLevel: (see C:\Users\etien\AppData\Roaming\Unreal Engine\AutomationTool\Logs\C+Program+Files+Epic+Games+UE_4.21\UBT-TestLocationIfCrash1-Android-Development.txt for full exception trace)
LogPlayLevel: Error: ERROR: UnrealBuildTool failed. See log for more details. (C:\Users\etien\AppData\Roaming\UnrealEngine\AutomationTool\Logs\C+Program+Files+Epic+Games+UE_4.21\UBT-TestLocationIfCrash1-Android-Development.txt)
LogPlayLevel: (see C:\Users\etien\AppData\Roaming\Unreal Engine\AutomationTool\Logs\C+Program+Files+Epic+Games+UE_4.21\Log.txt for full exception trace)
@stevenjt - Just wanted to feedback that your plugin is really helping in our development at the moment Thanks!
Can you tell me simply which is the blueprint node that provides me with the local coordinates of, for example, my iPhone or iPad as I move it in my city?
|
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500719.31/warc/CC-MAIN-20230208060523-20230208090523-00744.warc.gz
|
CC-MAIN-2023-06
| 2,988
| 23
|
https://1library.net/document/9ynnl6ky-a-foundation-for-machine-learning-in-design.html
|
code
|
Additional services for
Email alerts: Click here
Subscriptions: Click here
Commercial reprints: Click here
A foundation for machine learning in design
SIANG KOK SIM and ALEX H.B. DUFFY
AI EDAM / Volume 12 / Issue 02 / April 1998, pp 193 209 DOI: null, Published online: 08 September 2000
Link to this article: http://journals.cambridge.org/abstract_S0890060498122096
How to cite this article:
SIANG KOK SIM and ALEX H.B. DUFFY (1998). A foundation for machine learning in design. AI EDAM, 12, pp 193209
Request Permissions : Click here
A foundation for machine learning in design
SIANG KOK SIM1andALEX H.B. DUFFY2
1School of Mechanical and Production Engineering, Nanyang Technological University, Nanyang Avenue,
Singapore 639798, Republic of Singapore
2CAD Centre, University of Strathclyde, 75 Montrose Street, Glasgow G1 1XJ, Scotland, UK
(ReceivedJune 27, 1997;RevisedOctober 17, 1997;AcceptedNovember 10, 1997)
This paper presents a formalism for considering the issues of learning in design. A foundation for machine learning in design (MLinD) is defined so as to provide answers to basic questions on learning in design, such as, “What types of knowledge can be learnt?”, “How does learning occur?”, and “When does learning occur?”. Five main elements of MLinD are presented as the input knowledge, knowledge transformers, output knowledge, goals/reasons for learning, and learning triggers. Using this foundation, published systems in MLinD were reviewed. The systematic review presents a basis for validating the presented foundation. The paper concludes that there is considerable work to be carried out in order to fully formalize the foundation of MLinD.
Keywords: Design Knowledge; Design Process Knowledge; Design Reuse; Knowledge Transformation/Change;
Learning in Design; Machine Learning Techniques
Design experience provides a wealth of knowledge that de-signers can (re)-use to design better products within a shorter time-to-market period and at the same time be economi-cally competitive. The research work in the area of machine learning has contributed many methods that have been ap-plied to the acquisition of knowledge in design. This has been evident from the body of work reported in the field of machine learning in design (MLinD) (Duffy, 1997). What is evident from this work is the application of particular ma-chine learning methods to the acquisition of some specific design knowledge. In the area of MLinD, key questions raised by Persidis and Duffy (1991) are now being ad-dressed (Duffy, 1997):
•What type of knowledge is learned?
•How is learning taking place?
•When is learning taking place?
Given that there is sufficient evidence from MLinD re-search in which generalized past design knowledge can be
acquired using certain machine learning methods, there is now a need for a systematic approach to formalizing learn-ing in design. Leith (1990) argues strongly for a formalism to arrive at the state of what “ought to be” in artificial in-telligence (AI) and computer science rather than what “is” (i.e., current state of these disciplines) so as to overcome the ad hoc basis of software writing and the inefficient de-velopment of software systems. Using the same argument of Leith, given that machine learning in design is a special-ized application of AI and computer science, there is there-fore a need to put the study of MLinD research and the development of the MLinD systems to support designers on a formal basis.
To answer the questions raised in a structured basis, Sec-tion 2 presents basic elements of learning and a foundaSec-tion for learning using these elements. This foundation is pre-sented here as a basis for the “dimensions of machine learn-ing” raised by Grecu and Brown (1996) and similarly by Persidis and Duffy (1991). Since the research work in MLinD has resulted in the development of many published systems in support of knowledge acquisition in design, this founda-tion has been used to analyze and evaluate these MLinD systems. The five elements presented here are the input knowledge, the output knowledge, the knowledge trans-former, the learning triggers, and the learning goal. Sec-tion 3 presents the types of design knowledge learned (both
Reprint requests to: Dr. A.H.B. Duffy, CAD Centre, University of Strath-clyde, 75 Montrose Street, Glasgow G1 1XJ, Scotland, UK. Tel: (1 44)141-548-3134; Fax: (144)141-552-3148; E-mail: email@example.com Copyright © 1998 Cambridge University Press 0890-0604/98 $12.50
product knowledge and design process knowledge) given the input knowledge and the goal of learning as described in the MLinD systems. Section 4 provides evidence of var-ious knowledge transformers used to learn design knowl-edge but implemented in various machine learning methods and techniques. Evidence of the types of learning triggers and when learning occurs are described in Section 5. Sec-tion 6 compares the foundaSec-tion for machine learning with “dimensions of machine learning” raised by Grecu and Brown (1996) and concludes that the paper presents a struc-tured basis upon which to research and develop the field of machine learning in design.
2. DEFINING A FOUNDATION OF DESIGN LEARNING
It is proposed that a systematic approach to the study of learning design knowledge can be based on analyzing the knowledge change of design activities. This perspective is based upon the hypothesis posited by Persidis and Duffy that learning is inextricably linked to design. In this paper, learning in design is viewed as a knowledge-gaining activ-ity associated with the activactiv-ity of design.
A foundation for learning in design must be able to ad-dress some of the key questions in learning that were raised in Section 1. Since learning can be viewed as an activity, any formalism must invariably consider what knowledge is input into that activity, what is the output knowledge, and the knowledge change that transforms the input knowledge into output knowledge. Knowledge change involves the transformation of the existing knowledge into some new knowledge. Since there are many possible ways in which knowledge can be transformed, it is necessary to define what kind of knowledge can be learned. For example, given a past design, one can learn about the composition of the prod-uct in terms of “part-of” hierarchy, or about the relation-ships between attributes of components. Further, the learning activity often will have a specific goal, for example, gain new knowledge of a product, explore and generalize a
par-ticular domain. It is also necessary to define what triggers learning and when that trigger can occur.
Given the above, the basic elements of a learning activity may consist of:
•existing knowledge as input knowledge, Ik;
•knowledge transformers, Kt;
•output knowledge, Ok;
•learning goal or reason, Gl;
•learning triggers, Tl(Tlwand Tlt!.
These basic elements of learning may be related as shown in Figure 1.
In this figure, it is suggested that the input knowledge is transformed into new output knowledge, which then can feed back into the learning activity as input knowledge for yet more new knowledge. This output knowledge in itself may also trigger or act as a reason or goal for a learning activity.
2.1. Elements of a foundation for machine learning in design
Using the above elements as a basis, we now can map Persidis and Duffy’s, and Grecu and Brown’s issues and “dimensions” as shown in Table 1. It can be noted from Table 1 that Grecu and Brown go into far greater detail than Persidis and Duffy, possibly reflecting the evolution of our
Fig. 1. Elements of learning.
Table 1. Elements of machine learning in design
Basic elements of learning Persidis and Duffy Grecu and Brown
Ik(input knowledge) • Not explicitly addressed • What are the elements supporting learning?
• Availability of knowledge for learning Kt(knowledge transformer) • How is learning carried out? • Methods of learning
• Local vs. global learning Ok(output knowledge) • What knowledge is learned? • What might be learned?
Tl(trigger) • What can trigger learning? • What can trigger learning?
• When is learning triggered?
own understanding of the field. Having said this, the ap-proaches seem complementary and would seem to reflect the basic elements identified above. Some issues associated with the above are discussed further.
2.2. Types of design knowledge learned
The types of design knowledge to be learned are dependent on the activity of the design process, the types of input knowl-edge, the goal of the learning process, and when learning takes place. Persidis and Duffy (1991) state five main types of design-related knowledge: design requirements, design description, domain knowledge, case histories, and design management. Failures in design provide an opportunity to learn about the causes of the failure in order to avoid com-mitting the same failure again. Knowledge learned from de-sign failure could be in the form of types of failures and conflicts (Grecu & Brown, 1996; Vancza, 1991), heuristics for failure recovery, and conflict resolution (Grecu & Brown, 1996; Vancza, 1991). On the other hand, successful designs and design processes (Grecu & Brown, 1996; Vancza, 1991) also provide opportunities for learning about the character-istics of the successful past designs and goals/plans of de-sign processes. The review of MLinD systems in Section 3 shows that indeed there are numerous types of design knowl-edge that can be learned from past designs. It is therefore reasonable to conclude that there are many types of design knowledge that can be learned.
2.3. The learning goal
Because of the variety of knowledge that can be learned, it is sometimes necessary to specify a learning goal. The learn-ing goal influences what parts of the existlearn-ing knowledge are relevant, what knowledge is to be acquired, in what form, and how the learned knowledge is to be evaluated. Such goals, as in design, can be hierarchical in nature in that a specific goal could be a subset of a more general goal. For example, learning about a particular technology may help improve the quality of the product or speed up the design process. Another example of a specific learning goal such as learning about constraint violation in design may result in better conflict anticipation and resolution, which in turn may lead to a better design plan and hence a shorter design cycle (Leith, 1990). Thus, specific learning goals may be the stimulus for many of the unforeseen consequences of learning.
2.4. Types of knowledge change
The design process is the vehicle that designers use to change the state of the design through the application of scientific and design knowledge. As the state of the design changes, there is a corresponding change of knowledge state of the design. Persidis and Duffy presented three main
classifica-tions under which knowledge change can be considered, namely, acquisition, modification, and generation:
•Acquisition: the process in which new knowledge, ei-ther through direct input from the user or derived from knowledge sources, is received.
•Modification: the process of altering the existing struc-ture of knowledge through either addition or deletion of the knowledge component from the structure. The modified knowledge structure remains the same type.
•Generation: the process of creating completely new
knowledge from existing knowledge.
While knowledge change through modification is explic-itly described as the addition/deletion of knowledge com-ponent from the knowledge structure, the several ways in which knowledge change through generation is achieved are described below. The different processes by which knowl-edge change takes place are called knowlknowl-edge transform-ers. In the study of machine learning in design systems, seven knowledge transformers for knowledge generation and one knowledge transformer for knowledge modification are iden-tified. Some of these transformers are similar to those de-scribed by Michalski’s set of 11 knowledge transmutations (Michalski, 1993). Of these 11, the knowledge transmuta-tion of replicatransmuta-tion/destructransmuta-tion, sorting/unsorting, and se-lection/generation are not considered here as knowledge transformers because they are basically processes for reor-ganizing knowledge bases. There is no learning in the form of knowledge change per se.
2.4.1. Knowledge transformers
Because of the nature of design knowledge, which may exist in symbolic or numerical form, knowledge change operators or knowledge transformers consist of two types: the symbolic type and/or the non/subsymbolic type. A knowledge transformer is an operator that derives a piece of new knowledge from a given input or an existing piece of knowledge.
Group rationalization (or clustering)/decomposition (un-group). Group rationalization involves the grouping of past designs according to their similarities when considering par-ticular perspective(s) or criteria (Duffy & Kerr, 1993). The grouping may be based on single attributes and/or nested attributes. Decomposition removes the groupings.
Association/disassociation. Association determines a de-pendency between given entities or descriptions based on some logical, causal, or statistical relationships. For exam-ple, two entities may be related together by a taxonomic relationship (i.e., “kind-of”) or compositional relationship (i.e., “part-of”). The opposite is disassociation, which as-serts a lack of dependency (e.g., “not part-of”).
Derivations (reformulation)/randomization. Derivations are transformations that derive one piece of knowledge from another piece of knowledge (based on some dependency be-tween them). In contrast, randomization transforms one knowledge segment into another by making random changes. Because the dependency between knowledge components can range from logical equivalence to a random relationship, derivations can be classified on the strength of dependency.
Generalization/specialization. Generalization generates a description that characterizes all of the designs based on a conjunction of all of the specializations of that concept. Typically, the underlying inference is inductive. But gener-alization also can be deductive, when the more general de-scription is deductively derived from the more specific one using domain knowledge or existing knowledge (e.g., as in explanation-based generalization). Specialization increases the specificity of the description.
Abstraction/detailing. Abstraction generates a new ver-sion of the knowledge with less detail than the original through the use of representation of abstract concepts or op-erators. As distinct from generalization, the underlying in-ference is deductive in nature. Detailing is the opposite, in which the new knowledge is generated with more details.
Explanation/discovery. Explanation derives additional knowledge based on domain knowledge. Discovery derives new knowledge without an underlying domain knowledge. The knowledge transformers described above are not syn-onymous with the machine learning methods that are used in machine learning systems. The knowledge transformers describe the cognitive learning process whereas the ma-chine learning techniques describe how that process can be implemented as algorithms. For example, explanation-based generalization uses generalization and explanation of domain knowledge to derive new knowledge.
2.5. Learning triggers
Persidis and Duffy (1991) posit the need to identify events that act as “triggers” for the self-activation of the learning process. In this paper, it is suggested that these triggers can be classified into three main categories: in situ, provisional, and retrospective.
•In situ triggers are activated when there is a need to acquire new knowledge while the design is under focus of attention.
•Provisional triggers are activated when there is a fore-seen event that is envisaged to require additional knowl-edge. An example of a provisional trigger arises when there is a need to learn heuristics that may lead to a reduced search for a design solution. This learning is to be done in anticipation of or provisionally for a fore-seen event.
•Retrospective triggers are activated after an event. That is, learning is triggered by the need to learn from suc-cessful design(s)/failed design(s) and/or processes in hindsight.
Grecu and Brown (1996) identify several situations that could trigger learning. The reasons for learning are numer-ous but may occur due to the following:
•Novelty driven: when there is a new design problem, a new technology, a new process, or a new design requirement.
•Excellence driven: can the design be improved to achieve better quality and reliability? Can the design process be more streamlined or time-efficient or re-source-efficient? Can the design be market-competitive?
•Failure avoidance driven: can failure(s) in the design
be avoided? Can design constraints be overcome?
The above examples serve to illustrate reasons that give rise to learning and are by no means exhaustive in cover-age. For specific examples (gleaned from the study of MLinD systems) of what triggers learning and when learning is trig-gered, more details are given in Section 5 and Table 3.
2.6. Current work of machine learning in design
There is now considerable effort in supporting computer-based learning in design. Such work can act as a basis upon which to discuss the above elements. The examination of such systems can be broadly classified into two groups:
•systems that acquire knowledge of a product (i.e., de-sign concepts);
•systems that acquire knowledge of the design process (i.e., procedural knowledge and control knowledge).
It is difficult to ascertain the true nature of the goal or rea-son for an associated MLinD system’s learning activity. Con-sequently, the examination has concentrated on attempting to identify:
•the types of design knowledge acquired (Element Ok! and the associated learning goal (Element Gl!;
•the knowledge change as a result of the learning
activ-ity (Elements Ok/Ik!;
•what triggers that learning activity and when (Element Tl!.
3. TYPES OF DESIGN KNOWLEDGE LEARNED
Many MLinD systems have been developed to automate the learning of design knowledge. The purpose of this section is to show the variety of design knowledge that has been acquired through these MLinD systems. For the conve-nience of description, the types of design knowledge learned can be categorized into the following:
•product design knowledge;
•design process knowledge.
This categorization is important because it was revealed through the examination of the MLinD systems that, in most cases, the learning of meaningful product design knowl-edge occurs retrospectively. Similarly, certain types of de-sign knowledge are best learned while the dede-sign is in progress.
3.1. Product design knowledge
Systems such as BRIDGER (Reich, 1993), NODES (Persi-dis & Duffy, 1991; Duffy et al., 1995), CONCEPTOR (Li, 1994), PERSPECT (Duffy & Duffy, 1996), and NETSYN (Ivezic & Garrett, 1994) have addressed the issue of learn-ing product knowledge from past design concepts. The knowledge learned can be discussed under the headings of:
•composition of the components/subsystems that con-stitute the product;
•constraints imposed on the attributes due to the physics of the problem, geometrical and spatial requirements, life-cycle issues, design requirements, and so on;
•decomposition of the components that constitute the
•performance evaluation knowledge;
•dynamic learning of implicit knowledge.
3.1.1. Learning knowledge of design composition
Knowledge of the composition of past designs serves as as a useful starting point to initiate the synthesis process of design problems that have a similar design specification. For example, BRIDGER was developed primarily for the synthesis of different concepts of cable-stayed bridges. NODES also was developed primarily to support concept modeling operations during the early stages of design (Duffy et al., 1995).
While BRIDGER expresses the design composition knowledge as a taxonomic hierarchy (i.e., “kind-of” rela-tionship), NODES expresses both the taxonomic relation-ships as well as the compositional (“part-of”) relationrelation-ships of a design object. BRIDGER uses the concept formation capability of ECOBWEB to learn design composition
knowl-edge, and NODES uses maximum conjunctive generaliza-tion (MCG) to build the generalizageneraliza-tion of a new concept into existing concept(s).
The knowledge input into NODES or BRIDGER con-sists of past design instances or abstractions in terms of tax-onomic hierarchy (i.e., “kind-of” relationship and/or “part-of” relationship). Through the process of knowledge change (i.e., clustering or generalization) the output knowledge will be a new taxonomic structure updated to reflect the changes caused by the inclusion of the new design instance.
3.1.2. Learning empirical knowledge of quantitative and qualitative relationships
The numerical relationships of attributes of design con-cepts are important in defining the preliminary definition of the form or structure of the concept. The estimation of these values has been achieved by scaling or interpolation (BRIDGER; Reich, 1993), empirical network of associa-tion (CONCEPTOR; Li, 1994), statistical approximaassocia-tion (NODES using Designer; Duffy et al., 1995), and neural networks (NETSYN, Ivezic & Garrett, 1994). Each of these types of estimation of attribute values based on records of past design cases is described below.
Although there is in BRIDGER a taxonomic structure gen-erated from past design concepts in terms of attribute value pairs, the numerical relationships of these attributes are not explicitly represented. Therefore, for a given design speci-fication, BRIDGER derives appropriate values for the attri-butes through the interpolation of suitable design cases.
In the case of CONCEPTOR, the empirical relationships of design attributes of a given design domain are repre-sented as an empirical network. The empirical network cap-tures both quantitative variables and qualitative variables as nodes, which can be linked by empirical formulas (for quantitative variables) or design patterns (for qualitative vari-ables). The empirical formula represents proportional rela-tionships among the quantitative variables. Design patterns capture important associations among qualitative variables on which the designer should focus his or her attention. Each node represents a design feature, its default value, and range value. Collectively, the empirical network of nodes linked either by empirical formula or design patterns represents the description attributes of a learned concept.
In NODES, the numerical relationship of attributes of the design object is described by characteristics and formulas. Characteristics represent numerical attributes of concepts and always are associated with a single object. They also are associated with formulas and together form a constraint network where the values of characteristics are constrained by the values of other characteristics appearing in the same formula.
3.1.3. Learning knowledge of design constraints and design expectation
Both design constraints and design expectation play a cru-cial role in controlling the design process. A design expec-tation is a designer’s prediction about a design attribute’s value. Expectations may consist of a default value, a range of possible values, or a simple relation between design attribute values. Relationships between key design attributes and dependent design attributes are expressed as expecta-tion rules by Chabot and Brown (1994). An expectaexpecta-tion rule may be triggered when a value for a key attribute has been decided. These expectations and expectation rule(s) are an important source of design knowledge that can be learned from past design experiences and past design cases by in-duction (Chabot & Brown, 1994).
Design constraint knowledge provides the means for de-tecting design failures (Chabot & Brown, 1994). Although there are a large number of constraint types, categorizing them into hard or soft constraints aids in detecting design failures. Hard constraints are either satisfied or violated. If a hard constraint is violated, then the design decision be-comes unacceptable. Values for soft constraints are bounded by the degree of error tolerance that is acceptable or allow-able. Therefore, while expectation rules provide values to progress a design, constraint(s) test the values and detect any design failure.
Since different constraints and different expectation rules are called into play during a design process, Chabot and Brown use the mechanism of constraint inheritance as the machine learning method of knowledge compilation (Brown, 1991) whereby new knowledge of constraints is incremen-tally formed or updated.
3.1.4. Learning design decomposition knowledge
Liu and Brown (1992, 1994) posit that decomposition knowledge can be generated from design object knowledge, functional knowledge, design cases, design heuristics, gen-eral problem-solving knowledge, and domain knowledge. To extract knowledge that is appropriate for a given design context, Liu and Brown introduce the concept of “decom-position factors.” A decom“decom-position factor is a suggestion about how to partition an entity that may be a design object, a component, or a set of attributes.
The learning mechanism used is that of knowledge com-pilation (Brown, 1991). Knowledge comcom-pilation is a type of learning in which existing knowledge is converted into new forms with the intent to improve problem-solving efficiency.
The final decomposition knowledge is represented in a tree structure. Each node of the decomposition tree con-tains a problem description and a list of possible alternatives, competing decomposition hypotheses, as well as a set of links to subproblems, knowledge sources, interaction infor-mation among subproblems, and other relevant inforinfor-mation. From design object knowledge, Liu and Brown identi-fied three types of interactions among the components of an
object. The first type of interaction deals with explicit rela-tionships among the components, while the second type of interaction arises from the relationships between the object’s attributes and the components’ attributes. The third type deals with interactions arising from the intercomponent relations giving rise to interactions among the subcomponents.
Representing the components and interactions as nodes and links, respectively, in a graph, several pivotal nodes with many links converging to them can be identified. Each piv-otal node plus its connected nodes form a decomposition factor.
Design cases in which there are specific design prob-lems, with their associated design solutions, can be another source for generating decomposition hypotheses. Given a design problem, a set of relevant design cases can be re-trieved by using the object’s case index. A suitable case is selected if every constraint of the current design problem matches the constraints of a design case to a predetermined degree; then the design case becomes a candidate for de-composition factor seed. All of the selected cases are con-verted into decomposition factors. If a selected case has a matching quality above some limit, it can be used for gen-erating decomposition hypotheses.
3.1.5. Learning knowledge for performance evaluation Performance evaluation of a design in the various stages of the design process often determines if the design should be progressed further. But performance evaluation is not an exact science. Multiple criteria that are conflicting in nature are used to evaluate the performance of a design or designs. Current methods of evaluation often are either too simple or too complex for configuration design decisions (Mur-doch & Ball, 1996). Several researchers (McLaughlin & Gero, 1987; Murdoch & Ball, 1996) have posited that known configuration solutions in terms of their topological or geo-metric layouts, components, and materials can be analyzed and evaluated in terms of their performance so that good aspects of these designs may be reused and poor ones changed or discarded. That is, knowledge of performance evaluation can be learned from past design configurations. Several machine learning approaches (McLaughlin & Gero, 1987; Murdoch & Ball, 1996) have been developed to elicit the acquisition of design rules that map from the design so-lution space (i.e., structure or form) to the design behavior space.
Pareto optimization is used to isolate the set of best solu-tions in terms of the criteria defining the performance space. The knowledge acquired either through the Pareto/ID3 (McLaughlin & Gero, 1987) or GA/Kohonen Feature Map (Murdoch & Ball, 1996) is:
•Knowledge of the mapping between performance space (described by evaluation criteria) and the design de-cision space (described by design variables and their values).
3.1.6. Learning function knowledge and the causal models
Bhatta and Goel (1996) posit that generic teleological mechanisms (GTMs) are one type of design abstraction that can be learned from past designs through cross-domain anal-ogies. Cross-domain analogy involves the recognition of sim-ilarity between two problems from two different domains and determining what knowledge to transfer and how to transfer between them. GTMs take as input the functions of a desired design and a known design and suggest patterned modifications to the structure of the known design that would result in the desired design.
3.1.7. Dynamic learning of implicit design knowledge So far, the systems that use machine learning methods to acquire the design object knowledge for a given domain have been based on a fixed viewpoint that is predetermined by the knowledge engineer. Since it is not possible to com-pletely predict designers’ knowledge requirements, because each designer has different knowledge needs at different times and for different reasons, Kerr (1993) presents a new approach to utilize experiential knowledge called custom-ized viewpoint.
The key concept behind this flexible approach to edge utilization is that it generalizes experiential knowl-edge directly from specific experiences, according to designers’ knowledge needs, and subsequently utilizes this knowledge in design. To illustrate the concept, Kerr devel-oped, tested, and evaluated the utility of the customized view-point approach within the realm of numerical design. PERSPECT is supported by four subsystems, namely, DE-SIGNER, ECOBWEB, S-PLUS, and GRAPHER (Duffy & Duffy, 1996; Kerr, 1993).
PERSPECT uses ECOBWEB, a concept formation sys-tem to generate and organize the generalization of a set of past designs (described by attributes and values) into a con-ceptual hierarchy. It classifies each example (past design) and incorporates it permanently into a hierarchy by incre-mentally changing the hierarchy’s structure. The knowl-edge acquired by the approach is:
•Multiple forms of experiential knowledge of a domain (described by name, meaning, and units of attributes) in terms of empirical equations, numerical generaliza-tion, and heuristics. For example, empirical equations
quantify the relationship between these attributes to-gether with a measure of the unreliability of these equations.
•Multiple forms of implicit experiential knowledge through generalizations of past design information and identifying the most suitable generalization of past de-sign that supports the customized viewpoints for a new design.
3.2. Design process knowledge
Learning about the design process provides just as an im-portant wealth of knowledge as learning from past designs. A design process usually consists of design decisions made that results in design actions taken to progress the design. Design actions clearly interact with the evolving design. Learning about the design process invariably involves learn-ing about the decisions made, the rationale for the decisions made, and the effect of that decision on the evolving de-sign. Certain phases of the design process are exploratory in nature, and the use of control knowledge to manage the exploration is important.
Systems that capture design rationale (Gruber et al., 1991) and design history thus provide vital sources of knowledge for learning about the design process. While capturing de-sign rationale usually involves recording successful dede-sign decisions, the design history records both successful and failed decision lines. Both successful decisions and failed decision lines are objects of learning about the process.
Knowledge of the design process can be captured in the form of design plans or as a hierarchy of activations (con-dition and action pairs) on the blackboard of a blackboard system (Erman et al., 1980; Hayes-Roth, 1985). In artificial intelligence (AI) parlance, a sequence of dependent actions is called a plan. As such, systems such as ADAM (Knapp & Parker, 1991), BOGART (Mostow, 1989), ARGO (Huhns & Acosta,1992), and DONTE (Tong, 1992) have used the AI approach of planning to describe and capture the design process, whereas only DDIS (Wang & Howard, 1994) has used the blackboard concept. Design plans usually do not capture failed lines of design actions, whereas the black-board approach usually captures both the successful as well as the failed design actions, resulting in capturing the de-sign history.
The knowledge change of design plans can be accom-plished by the acquisition of new design plans, modifica-tions of existing plans through design re-use, and generation of new plans or generalized plans from existing plans at the end of one or more similar design processes. The review of systems such as BOGART, ARGO, DONTE, CDA, and DDIS provides evidence of such knowledge change.
3.2.1. Acquiring design plans
A knowledge change in a design plan can be learned through the process of acquisition. This simply involves be-ing told by the designer or recordbe-ing the series of actions taken by the design, by inferring the design plan from a pre-vious design case or solution (Duffy, 1997).
Both BOGART and ARGO record user inputs as a means of acquiring new plans. BOGART uses VEXED to record successive design steps in a tree-like design plan that con-sists of nodes representing design modules. Each module can be decomposed or refined into submodules by a catalog of “if-then” refinement rules provided by VEXED. If VEXED lacks any of these rules, it will learn the manual decomposition step by generalizing that step into a new rule. This learning facility within VEXED is provided by LEAP (Learning Apprentice).
ARGO acquires a plan as it solves a problem and is rep-resented using an acyclic graph of dependencies among plan steps (instantiated rules). If one plan step adds an assertion that satisfies the condition part of a rule, the second step becomes dependent on the first. The dependency graph may contain independently solvable subproblems or dependent subproblems with justifications maintained by the truth main-tenance system.
In ARGO, the design plans are represented as schemas of corresponding preconditions and postconditions that are rep-resented as a database of assertions stored as slots of frames in a truth maintenance system. A module is represented as a collection of assertions describing its specification, com-ponents, interconnections, and so forth, each with a belief status of IN or OUT supported by a set of justifications. The OUT status is caused by actions that fail, and the cor-responding rule instances are not included in the depen-dency graph for representing the design plan. Therefore, ARGO does not learn plans that incorporate failed lines of reasoning.
3.2.2. Acquiring case-dependent design plans
DDIS integrates both domain-based reasoning and case-based reasoning in its strategy for solving design problems. It uses case-dependent knowledge that it acquires from cur-rent design session(s) to supplement its domain-independent knowledge for future design(s) or redesign(s). DDIS there-fore records all design actions as knowledge source activa-tion records (KSARs) and the design history (a sequence of executed KSs and their bindings) on its design blackboard. By analyzing these records, DDIS abstracts case-dependent
plans and goals. These plans and goals can be posted di-rectly on the control blackboard by case-dependent control actions during subsequent design sessions. These case-dependent actions compete with case-incase-dependent knowl-edge sources at every design cycle to allow case-based reasoning to influence domain-based reasoning, so that past design actions leading to dead-ends or failures are avoided. DDIS therefore generates case-dependent design edge after the completion of the design session. The knowl-edge generated consists of the following:
•control knowledge of a particular session is also ab-stracted to a global design plan and several redesign plans so that they can be used separately in a flexible manner according to new situations encountered;
•knowledge of constraint violations that can be applied
to new cases in order to focus early on critical con-straints that are most likely to cause problems.
3.2.3. Generalized design plans
BOGART, ARGO, and DDIS also generate generalized plans from several plans or by abstracting new plans. The details of how the generalized plans are generated through various knowledge transformers are given in Section 4.
3.3. Summary of the types of knowledge learned
Table 2 gives a summary of the various types of design knowledge that can be learned from the MLinD systems. The table shows the knowledge input Ik, the product/process knowledge learned Ok, and the reason for learning that knowledge, Gl.1
4. KNOWLEDGE TRANSFORMERS IN MLinD SYSTEMS
In this section, the purpose is to show the evidence of the knowledge transformers considered in Section 2 that MLinD systems use to transform input knowledge into new or mod-ified design knowledge. Although these knowledge trans-formations are implemented in terms of various machine learning methods (either symbolic or subsymbolic), it is the nature of the characteristics of each type of knowledge transformation that makes their identification within the MLinD systems possible. Each machine learning method used may be either symbolic or subsymbolic in nature, de-pending on the nature of the representation of the input/ output knowledge. In describing the various knowledge transformers, no distinction is made as to whether the knowl-edge transformed is product design knowlknowl-edge or design pro-cess knowledge.
1This is usually inferred from the context of the design problem
4.1. Group rationalization (or clustering)/ decomposition (ungroup)
BRIDGER and CONCEPTOR are examples of machine learning systems that learn design concepts using the knowl-edge transformer of group rationalization or clustering. While CONCEPTOR uses the concept clustering system called COBWEB, BRIDGER uses ECOBWEB (Reich, 1993), which is an extension of COBWEB. Both machine learning systems for concept formation generate design concepts from the characteristics of similar past designs. Using past bridge designs as training examples, each system generates hierar-chical classification structures that can be used to assist in the synthesis of similar bridge designs.
CONCEPTOR not only learns design concepts from past designs through the knowledge transformers of clustering (or concept aggregation—a term used in CONCEPTOR), it also derives numerical relationships (or concept char-acterization in CONCEPTOR’s terminology) among the quantitative design attributes of a concept. Derivation as a knowledge transformer is described in Section 4.4.
COBWEB/ECOBWEB. To build the hierarchical struc-ture, COBWEB/ECOBWEB integrates the processes of clas-sifying examples and incorporating them into a hierarchy.
It employs five operators to determine how best to incorpo-rate an example into the hierarchy. Each resulting partition (classification) is evaluated using a utility function to deter-mine the category utility (Gluck & Corter, 1985), which is a measure that quantifies the similarity between members of a partition. It selects the classification that results in the highest category utility value, incorporates the example per-manently into the hierarchy, and generates the appropriate con-ceptual description that suits the new incorporated example. Each partition is described as a conjunction of attribute– value pairs, and each partition has a probability to indicate its frequency of occurrence in the training examples. As a re-sult, a hierarchy represents only one concept, while the nodes in the hierarchy represent subsets of the concept.
Hence, in learning taxonomic-type design concept(s) from past designs, the knowledge change involved is the cluster-ing of past designs into a hierarchical structure of concepts, each node in the hierarchy representing subsets of a main concept. Knowledge of the design artefact such as function-ality, structure, or behavior is not explicitly represented. The knowledge of the artefact’s decomposition structure or com-posite structure in terms of “part-of” links is not reflected in BRIDGER’s/CONCEPTOR’s classification hierarchy structure.
Table 2. Relationships between input knowledge, output knowledge, and learning goal
Input knowledge Ik Output knowledge Ok Learning goal Gl
Instance(s) of past design(s) together with existing taxonomic or compositional knowledge
• Taxonomic knowledge of design concepts (e.g., BRIDGER)
• Compositional knowledge of design concepts (e.g., NODES)
• Expedite synthesis of preliminary design concepts
Records of past designs in terms of attributes and attribute values
• Empirical knowledge of quantitative information (e.g., CONCEPTOR, NODES) • Design patterns of qualitative relationships
(e.g., CONCEPTOR, NODES)
• Expedite preliminary definition of form and structure of design concept
Knowledge of current design constraints
• Knowledge of new/updated design constraints (e.g., Chabot & Brown, 1994)
• Streamline design process by detecting and avoiding design failure
Records of failed constraints • Knowledge of anticipated crucial constraints (e.g., DDIS)
• By checking crucial constraints early in the design, leading to shorter design cycle Past design cases of design problems
and corresponding solutions
• Design decomposition knowledge in terms of decomposition factors (e.g., Liu & Brown, 1992)
• Streamline design process by focusing on interrelated systems and/or components Past design configurations and
performance evaluation criteria
• Knowledge of mapping between performance evaluation space and design decision space (e.g., McLauglin & Gero, 1987;
Murdoch & Ball, 1996)
• Excellence driven to achieve better design
Past design concepts described by attributes and values
• Multiple forms of explicit design knowledge (e.g., PERSPECT)
• Multiple forms of implicit design knowledge (e.g., PERSPECT)
• Excellence driven by utilizing knowledge from multiple sources
Records of design actions described by preconditions and postconditions
• Abstracted design plan (e.g., ARGO) • Case-dependent design plan (e.g., DDIS)
Another example where the knowledge transformer of clustering is used is in the context of learning design per-formance evaluation knowledge. The design solution space has to be categorized into different classes of solutions for the purpose of mapping them with the performance evalu-ation space. To elicit knowledge for performance evalua-tion, meaningful mappings between the decision space and the performance space modeled must be established. While McLaughlin and Gero (1987) concentrate on the solutions near the Pareto boundary as sources of knowledge in the evaluation space, Murdoch and Ball (1996) suggest that the entire solution set, from both the design space and the eval-uation space, represents valuable design information that must be analyzed for effective reuse. The entire evaluation space and the design spaces must be analyzed to identify different classes of solutions and trends in design practice. Because of different areas of the design solution space and the performance space considered for mapping, McLaugh-lin and Gero (1987) chose ID3 to distinguish solutions that are Pareto-optimal (i.e., near the Pareto boundary) and those that are not, while Murdoch and Ball (1996) chose the clus-tering capability of the self-organizing neural network called the Kohonen Feature Map to categorize the entire design solution space.
ID3. The induction algorithm ID3 is used as a means of inferring general statements about the nature of solutions that exhibit Pareto optimal performance in terms of a set of performance criteria. The positive example set consists only of decision and performance data of solutions that are Pareto-optimal in terms of the chosen criteria. The negative exam-ple set is generated by combinations of design decisions that are inferior in performance. The heuristic rules that best rep-resent the concept to be learned are those with the most pos-itive examples and the least negative examples.
Kohonen Feature Map. The Kohonen Feature Map is a neural network that can learn clustering patterns unsuper-vised. That is, the mapping between criteria in the perfor-mance evaluation space and the configuration parameters (i.e., component and materials) in the design space can be clustered without precategorization of the design space.
Each solution consists of a set of parameterized compo-nents that can be applied as a training example to the net-work. The network uses an unsupervised learning algorithm to generate a mapping between the high-dimensional de-sign space of component parameters and the neurons in the network. The map generated provides a topological (i.e., nearest neighbor) relationship among the component parameters. The network topology then can be inspected to identify clusters or archetypes that span the original set of design solutions. By decoding the nodes within each cluster back to the performance evaluation space, a map-ping between the two spaces is achieved. The archetype solutions then are analyzed to identify the characteristics
of configurations that contribute to high or low technical merit.2
4.2. Similarity/dissimilarity comparison
Knowledge change of a design plan through similarity com-parison is made possible if there exists an original target design plan. The machine learning method used is learning by analogy, which involves a transfer of information/knowl-edge from a base domain/plan to a target domain/modified plan.
Having acquired the history of design decisions made in a previous design, BOGART ( Mostow, 1989; Mostow et al., 1992) uses the derivational analogy method by Car-bonell (1983, 1986) to change the design plan by reasoning from the previous plan. The derivational analogy method represents a problem-solving plan as a hierarchical goal struc-ture, showing how and why each goal was decomposed into subgoals. It solves a new problem by replaying this plan top-down. When the subplan for a subgoal fails, the plan is modified by solving that subgoal from the user input of a new solution. By this process of similarity/dissimilarity com-parison, a new design plan is constructed.
NODES learns/builds a model of the conceptual design linked by a compositional network of concepts through “part-of” and “kind-“part-of” associations and a numerical network of characteristics through association between objects and for-mulas. In the compositional network the nodes denote ob-jects or assemblies, and the arcs denote the directed relation (or association) “part-of” between two nodes. In the numer-ical network, nodes represent the characteristics of objects or formulas, and arcs represent the link or association be-tween two nodes, one of which is a characteristic and the other of which is a formula in which the characteristic ap-pears (Duffy & Duffy, 1996).
The concept characterization phase of CONCEPTOR is an example of deriving new knowledge based on some depen-dency between them. After concept aggregation, CONCEP-TOR derives two types of relationships within a concept: empirical formulas among quantitative design attributes and design patterns among qualitative attribute–value pairs that frequently appeared in past examples.
2Technical merit combines in one generic measure of design merit three
For some design problems there is a need to apply the probability estimation function to acquire the probability es-timate of each value of each unassigned design property (e.g., estimates of loading, the material behavioral properties are statistical in nature).
Ivezic and Garrett (1994) developed a system called NETSYN to learn the Bayesian a posteriori probabili-ties of design properprobabili-ties. NETSYN uses the feed-forward backpropagation neural network as the machine learning technique to acquire and represent the probability estima-tion funcestima-tion. The probability estimaestima-tion funcestima-tion is ac-quired through inductive learning using past designs to train the neural network to estimate the desired probabili-ties. The trained network estimates Bayesian a posteriori probabilities.
To use conventional classifiers, one has to estimate the conditional probabilities P~D6Hi!for each design property
and the a priori class probabilities. The main difficulty lies in the elicitation of conditional probabilities P~D6Hi!that
reflect the actual design knowledge. This is estimated by assuming some idealized probability distribution (e.g., Gaussian distribution). The neural network approach esti-mates the Bayesian probabilities in a direct way, offering an approach where prior assumptions on probability distribu-tions need not be made.
The computational model estimates the probability for each value of each property being used in a given design context. Each design context involves several design prop-erties for which values have to be assigned. Therefore, the construction of NETSYN architecture is modular, that is, for each design property a neural network structure is as-signed to act as a probability estimation function for that property.
Knowledge derived through generalization has a greater problem-solving scope. This is because generalized rules or knowledge generally can be applied to a wider range of prob-lems for a given domain or complex problem. Different types of knowledge related to the design product/process can be derived through the process of generalization. For example, NODES enriches its design knowledge base, called the con-cept library, by progressively accumulating solutions of problems defined within a particular domain. It uses gener-alization as a knowledge transformer so that new concepts are reflected in all of the concepts that are generalizations of that concept. By integrating with the DESIGNER sys-tem, numerical aspects of the concept (i.e., characteristics and associated formulas) can be analyzed. Both BOGART (Mostow, 1989) and ARGO (Huhns & Acosta, 1992) ac-quire generalized design plans from several plan instances or by abstracting new plans. Design actions interact with the evolving design. The interaction between design action and the design could be generalized into useful design rules.
Generalized design concepts. NODES generalizes knowl-edge from the most comprehensive concepts within a con-cept library to the less specific. Numerical parameter ranges and compositional knowledge are generalized to all of the associated superclasses to ensure that there is no contradic-tion between a particular concept and its specializacontradic-tions. The generalization mechanism that is responsible for the updat-ing of knowledge is invoked automatically whenever a new concept is saved in the concept library.
When a design has been completed, the evolved model in NODES is used to increase its knowledge by acquiring the relevant knowledge of the new design. The mechanism in-volved is the decomposition of the design into its constitu-ent concepts (or specializations), along with appropriate constituent and connective relations, and merging each con-cept with its corresponding library.
NODES uses a machine learning technique called maxi-mal conjunctive generalization ( MCG) ( Dietterich & Michalski, 1983). MCG ensures that no item of knowledge is associated with a concept unless it is associated with all of the concepts that are a specialization of that concept. In terms of set theory, this means that the set of items of knowl-edge associated with a concept is the intersection of the sets associated with the specialization of that concept.
Generalized design plans. BOGART uses VEXED’s abil-ity (Steinberg, 1992) to interactively record decisions in terms of general rules that can be easily replayed in a new con-text, rather than specific operations that cannot be general-ized. In BOGART, a design plan contains a node for each module. When the module is refined, the node is annotated with the name of the decomposition rule and the values of its parameters, and connected to a new child node for each submodule.
Specialized (compiled) design constraint knowledge. De-sign constraint knowledge is the primary method of detect-ing design failure (Chabot & Brown, 1994). Past design knowledge can be expressed as design expectation rules that relate the key design attribute with the dependent design attributes. When the key design attribute’s value is decided, the design expectation rule is triggered. This results in the numerical or symbolic computation of the expected values of dependent design attributes. The expected values of ei-ther type (expected value range or expected symbolic value) are compared with the corresponding design attribute val-ues. An expectation violation (Chabot & Brown, 1994) oc-curs when an inconsistency is noticed between the two sets of values, resulting in the creation of a DSPL3expectation violation structure. The information in the expectation vio-lation structure is used by the Generic Object Knowledge Base (GOKB) Reasoner to transform the relevant constrain-ing knowledge in the GOKB into a DSPL constraint. The knowledge compilation process consists of four sequential subprocesses of Reasoner, Transformer, Inheritor, and Ex-ecutor. The Reasoner analyzes the role descriptors of the dependent (target) attribute into either potential numeric val-ues or a member of a list of symbolic valval-ues and the struc-tural descriptors of the design attributes. Constraints are inherited from the GOKB when a relevant explanation has been found by the Reasoner. The Transformer component supervises the transformation of the relevant GOKB knowl-edge into a DSPL constraint structure. The newly inherited constraint is tested by the Executor component. A “success-ful” test ensures that the new design attribute is valid and the newly inherited constraint knowledge is learned as a DSPL entity for future use.
Chabot and Brown therefore view constraint inheritance as a form of failure-driven learning that transforms a less efficient generalized deep object knowledge into surface knowledge that is highly specialized, tuned, and effective for the given design problem.
Abstraction in empirical equations. Abstraction in em-pirical equations may become necessary in the event that no useful empirical equations exist or because not enough attribute values are known. PERSPECT can be used to es-timate the values of the unknown attributes. Using their own or PERSPECT’s knowledge of design attribute depen-dency, designers using PERSPECT can define a perspec-tive consisting of unknown attributes and related attributes, generate a viewpoint of experiential knowledge that can be used to find a past design or group of designs similar to the current design, and use associated similar attributes as val-ues for the uninstantiated attributes in the current design.
If the domain model is too complex (i.e., described by too many empirical equations), designers can delete un-wanted variables from empirical equations to generate sim-pler equations, which then can be used to estimate values of the design model. PERSPECT achieves this capability by the process of abstraction. Abstractions of empirical equa-tions mean that dependent attributes can be assigned with fewer required attributes. DESIGNER can be used to deter-mine the least influential input variable of the equation and suggest the variable as most suitable for deleting from the empirical equation.
Design plan abstraction. Both ARGO and DDIS use the knowledge transformation of abstraction to generate design plans. In ARGO, this task is accomplished by computing macrorules for increasingly abstract versions of the plan and inserting these rules into a partial order according to some abstraction relation. Macrorules, consisting of relevant pre-conditions and postpre-conditions, are computed for each plan and stored in a partial order according to an abstraction scheme. These macrorules are built by compiling through the instances rules of the plan using a variant of explanation-based generalization (EBG) (Mitchell et al., 1986; DeJong & Mooney, 1986). The abstraction is accomplished by in-crementally merging each set of edge macrorules into a set of cumulative macrorules for previously merged rules. The plan abstraction scheme consists of deleting all of its leaf rules that have no outgoing dependency edges, since these leaf rules are those that deal with design details.
At the end of each design session, DDIS abstracts the con-trol knowledge recorded on the concon-trol blackboard to one glo-bal design plan and several redesign plans. The processes the DDIS uses to abstract design plans are as follows:
•All knowledge source activation records (KSARs) that modified the solution blackboard are identified and un-necessary design steps that led to unsuccessful alter-natives or that did not contribute directly to the design process are filtered or removed.
•For each identified major action (KSAR), a case-dependent goal is created to prefer the same knowl-edge source or same type of action in the future.
•The major design actions are classified into design and
redesign actions. Design actions are those that lead di-rectly to the eventual solution. Redesign actions are those that are executed when a constraint violation is present on the blackboard. The goals corresponding to design actions are grouped into the design plan, while the redesign goals are grouped into redesign plans cor-responding to each backtracking episode that resulted from constraint violations.
•The intentions of a global plan and case-dependent re-design plan are stated. The intention of a global plan is to generate design values for all of the design attributes and to satisfy all of the applicable constraints of the design. The intention of a case-dependent redesign plan 3Design Specialists and Plans Language ( DSPL) is a
is to satisfy all of the unsatisfied constraints that trig-gered the redesign process.
•The critical constraints that were violated and caused backtracking are recorded so that they can be consid-ered early in future designs.
Detailing to reconstruct design history. Unlike BOG-ART and ARGO, which rely on records of past design de-cisions in the form of a design plan, CDA first reconstructs from a similar solution a design plan using predefined rules (Britt & Glagowski, 1996). So while BOGART uses a der-ivational analogy to solve a new design problem, CDA uses a reconstructive derivational analogy (RDA) algorithm to automatically reconstruct design plan(s) from a large col-lection of past working design(s). Using the knowledge base of circuit design domain rules and information about the new circuit problem, CDA’s reconstruction expert finds ap-plicable rules, and selects and applies the preferred rule to the current circuit, adding design components each time un-til the final circuit meets the design requirement. In this pro-cess of detailing CDA reconstructs the design plan and acquires the rules for the composition of the design.
DONTE illustrates the learning of control knowledge to ex-plore the design space through the process of discovery about the design space. Through the process of discovery, IDEAL (Bhatta & Goel, 1994) demonstrates learning physical prin-ciples of a “concept” description from examples without knowing the target concept a priori. The process of discov-ery is generally considered to consist of two distinct phases: hypothesis formation and hypothesis testing.
Learning design control knowledge in DONTE. The dis-covery learning task is initiated by hypothesis formation in which a current hypothesis on the design space is represented as a set of subproblems that are presumed to be independent. Through design decisions made on these independent subproblems guided by control heuristics, interactions of sub-problems are discovered and these interactions are aggre-gated into what is referred to as a macrodecision. The objective of a design decision is to minimize a cost evalu-ation function that favors certain design solutions (e.g., NAND gates are preferred over other gates). The current hypothesis is updated by the formation of the macrodeci-sion, resulting in a new hypothesis about the design search space. Through this process of hypothesis formation of the design search space, information assimilation through each design decision made, and updating the hypothesis, DONTE learns control knowledge to optimally search the design space.
Learning models of physical principles in IDEAL. Using hypothesis formation on past designs’ structure-behavior-function (SBF) models of physical devices, Bhatta and Goel (1994) show how behavior-function (BF) models of
phys-ical principles can be acquired for future use in design. Dis-covering physical principles from abstract design models of physical devices is implemented as a learning component of IDEAL (Integrated DEsign by Analogy and Learning).
The models of specific devices (SBF models) provide both the content and constraints for learning the models of phys-ical principles (BF models) by incremental generalization over design experiences. In particular, Bhatta and Goel show that the function of a device determines what to generalize from its SBF model, the SBF model suggests how far to generalize, and the topology of functions indicates what method to use for generalization. By using content and con-straints of the model, IDEAL is able to discover physical principles using fewer examples.
Table 3 gives a classification of MLinD systems in terms of the knowledge transformers used to generate new design knowledge and implementation of that knowledge change through the machine learning methods supplemented by other methods (e.g., Pareto optimization), which results in cer-tain types of knowledge structures.
5. TYPES OF TRIGGERS FOR MACHINE LEARNING
The purpose of this section is to show evidence of what can trigger learning and when that trigger is likely to occur. Knowing what these triggers are and when these triggers initiate the learning process are important questions that must be answered if machine learning capability is to be incor-porated into design support systems. To discuss these trig-gers by themselves and not relate them to the context of the knowledge learned and the knowledge transformer in-volved would not show the relationship between these ele-ments of learning. Sections 5.1 to 5.3 give some examples of learning design knowledge under different types of gers: namely, the retrospective, in situ, and provisional trig-gers that are implemented/implied in the MLinD systems reviewed. Since these examples do not represent exhaus-tively the range of the types of triggers and their related triggering events, Table 4 gives a summary of what can trig-ger learning and when these trigtrig-gers occur in relation to the knowledge learned and the knowledge transformer in-volved in the MLinD systems reviewed.
5.1. Retrospective triggers
5.2. In situ triggers
In situ triggers of learning occur during the design process, when design decisions are made. Design decisions are made in relation to the design object and/or design process. These design decisions may lead to a successful design action or to a failure. Learning can occur under such design decisions and actions. Some examples of in situ triggers implemented in MLinD systems are discussed below.
Failure in achieving behavioral specifications. Design ad-aptation is a common practice in conceptual functional de-sign. Design adaptation usually occurs in several phases: in adapting a design retrieved from past design cases to satisfy the new behavioral specification, and in diagnosing and re-designing a failed design to achieve the desired behavior. Thus, Goel and Stroulia consider the design adaptation task as learning (Ashok & Stroulia, 1996). This process of
learn-ing takes place in situ as the design adaptation processes. Goel and Stroulia identify three types of diagnosis in de-sign adaptation that could trigger the learning process dur-ing design:
•The design does not achieve the desired function of the device. The device fails to achieve the desired func-tion because of incorrect specificafunc-tions of one/more of the components.
•The design results in undesirable behavior. The un-desirable behavior is due to the under/overspecifica-tion of the attribute of the component that influences its behavior.
•The specified structural component in the design has poor behavior. The component fails because of the over-specification of another component whose behavior has an adverse effect on the specified component.
Table 3. Knowledge transformers used in various MLinD systems and the related machine learning or other methods
Knowledge transformer MLinD systems involved Machine learning/Other methods Design knowledge represented Group rationalization/
• BRIDGER • ECOBWEB/EPROTOS • Hierarchical structure of concept/ subconcept
• CONCEPTOR • COBWEB • Decision tree of rules for Pareto optimum design
• McLaughlin & Gero (1987)
• ID3/Pareto • Clusters or archetypes of design solution mapped to performance evaluation space
• Murdoch & Ball (1996) • Kohonen neural network/GA Similarity/dissimilarity
• BOGART • Derivational analogy • Design plan as a hierarchical goal structure
• DDIS • Case-based reasoning • Design plan/history
Association/disassociation • NODES • Semantic links in network • Compositional network of concepts • Numerical network of characteristics Derivation/randomization • CONCEPTOR • Concept aggregation • Empirical formula among quantitative
attributes. Design patterns among qualitative attributes
• NETSYN • Modular backpropagation neural network • Bayesian a posterior probabilities of design properties represented as network of weights in neural network structure
Generalization/specialization • NODES • Maximal conjunctive generalization (MCG) • Generalized rules of design concepts • BOGART/LEAP • Generalization using EBL generalization • Generalized design rules
• DSPL • Knowledge compilation through constraint inheritance
• Generalized design plan. Constraint rules
Abstraction/detailing • PERSPECT • ECOBWEB/DESIGNER • Abstracted empirical equations • ARGO • Merging edge macrorules into cumulative
macro by removing leaf rules
• Abstract plan of macrorules
• DDIS • Identify, classify all activated KSARs into two types of design plans
• A global design plan and several redesign plans
• CDA • Reconstructive derivational analogy • Detailed design plan built bottom-up Explanation/discovery • DONTE • Hypothesis formation/hypothesis testing • Discovery of macrodecision rule to
They attributed the design failure in the three types of diagnosis to structural causes. Therefore, while knowledge of function to structure (FrS) mapping is useful for new conceptual designs that may be learned retrospectively, knowledge of structure-behavior-function (SBF) models can be learned in situ during design adaptation. Each of the three different diagnosis tasks requires different schemes for accessing the internal behaviors that result in the device tions. In particular, task (1) requires the use of design func-tions as indices into the internal behaviors that result in the device functions; task (2) requires indices from the primary behaviors of the device; and task (3) requires the use of struc-tural components of the device as indices into the internal behaviors in which they play a functional role. KRITIK2’s SBF models (Ashok & Stroulia, 1996) support all three kinds of indexing schemes. This enables the system to flexibly access the internal behaviors relevant to the current diagno-sis task and to thereby localize the diagnostic search.
Violation of design expectations. The DSPL system with constraint inheritance implemented by Chabot and Brown (1994) is an example of in situ triggers that occur when de-sign expectations are violated. Whenever dede-sign expecta-tion violaexpecta-tions occur, constraint inheritance as a form of failure-driven learning is activated. When an inconsistency is detected in the evaluation of design expectation rules, the relevant design object knowledge is identified for knowl-edge compilation into new constraint surface knowlknowl-edge through the constraint inheritance learning mechanism. The new constraint knowledge then is used to test the value of the design attribute for which there is an expectation. The addition of the new constraint knowledge to the existing de-sign constraint knowledge base leads to expectation-failure– driven learning (Chabot & Brown, 1994).
Customized viewpoints. The concept of customized view-points is an example of learning design knowledge in situ.
Table 4. Learning triggers in relation to knowledge learned and knowledge transformer involved
Knowledge transformer, Kt Knowledge learned, Ok What triggers learning, Tlw
When is learning triggered, Tlt
Group rationalization/decomposition • Taxonomic knowledge of design concepts • New concept • Retrospective • Clusters of design configuration map to
performance evaluation space
• Performance trends in new design • Retrospective
Similarity/dissimilarity comparison • Knowledge of design plan • New but similar design • In situ • Knowledge of case-based design plan • New design case • In situ Association/disassociation • Compositional knowledge of design
• New design configuration • In situ
Derivation/randomization • Empirical formula among quantitative design attributes
• New/updating empirical relationship(s)
• Design patterns among qualitative attribute–value pairs
• New/updating design patterns • Retrospective
• Posterior probabilities of design properties • New knowledge of a posterior probabilities
Generalization/specialization • Generalized design concepts • New concept saved • In situ
• Generalized design plans • Module(s) in plan refined • In situ/retrospective • Generalized design rules • No existing design rules • In situ
• Specialized design constraint knowledge • Constraint violation • In situ Abstraction/detailing • Abstracted empirical equations • Nonexistence of useful empirical
equation or insufficient knowledge of attribute values
• In situ/provisional
• Abstracted design plan by removing leaf nodes from plan
• New abstracted design process • Retrospective
• Abstracted from session control • Past design cases to improve design • Retrospective knowledge of the following: process of similar design(s)
lglobal design plan
lrelated redesign plans • Crucial constraints that triggered • Provisional
lconstraint violations redesign process
• Detailed design plan reconstructed • No similar design plan existed • Provisional bottom-up
|
s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704803308.89/warc/CC-MAIN-20210126170854-20210126200854-00689.warc.gz
|
CC-MAIN-2021-04
| 72,331
| 294
|
http://web3d.org/pipermail/x3d-public_web3d.org/2019-September/011260.html
|
code
|
[x3d-public] X3Dv4 Annotation Component revision, request review
Brutzman, Donald (Don) (CIV)
brutzman at nps.edu
Sat Sep 14 16:43:03 PDT 2019
Attached please find a review of X3Dv4 Annotation Component.
The current version is in github version control. The SIGGRAPH 2019 version can be found at
X3Dv4 Draft Specification available to public now!
X3Dv4 Public Working Draft Specification
X3D Annotation Component
Here are some key questions:
a. Can we move url field to 42.3.1 X3DAnnotationNode and then omit 42.4.6 URLAnnotation?
b. Can we better rename 42.4.3 GroupAnnotation node as InlineAnnotation or ModelAnnotation? (The name GroupAnnotation is ambiguous and seems to imply that it is grouping annotations themselves.
c. Can we better rename 42.4.4 IconAnnotation as ImageAnnotation? (The term "icon" implies characteristics such as small size or mostly square, but the node had no such restriction.)
d. Examples needed.
All feedback welcome. Please provide initial comments prior to next X3D teleconference Friday 14 SEP. Once we have names sorted out OK, am ready to pursue DTD Schema Tooltips, prototype stubs, github spec revision, and initial implementation in X3DJSAIL Java and X3DPSAIL Python.
all the best, Don
Don Brutzman Naval Postgraduate School, Code USW/Br brutzman at nps.edu
Watkins 270, MOVES Institute, Monterey CA 93943-5000 USA +1.831.656.2149
X3D graphics, virtual worlds, navy robotics http://faculty.nps.edu/brutzman
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 20190820_42 Annotation component_001.pdf
Size: 523458 bytes
Desc: 20190820_42 Annotation component_001.pdf
More information about the x3d-public
|
s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400191780.21/warc/CC-MAIN-20200919110805-20200919140805-00229.warc.gz
|
CC-MAIN-2020-40
| 1,675
| 25
|
http://www.coderanch.com/t/355409/Servlets/java/strange-session-management-websphere
|
code
|
This week's book giveaway is in the OCAJP 8 forum. We're giving away four copies of OCA Java SE 8 Programmer I Study Guide and have Edward Finegan & Robert Liguori on-line! See this thread for details.
Hi, We are facing a strange problem regarding session implementation. We have set a session timeout of 25 min but some of the users (ONLY SOME)of our site are getting timed out in less than 25 mins . We are using WebSphere 3.5.3 and Netscape Enterprise Server .We have overridden the default timeout provided by WebSphere by using setMaxInactiveInterval() method of HttpSession class. We are using cookies for session management and on extensive debugging of our problem , we found that at some of the client's browser , our session cookie(created by WebSphere) is not being created altho' the user has cookies enabled on his browser's settings. I know one of the solution to this problem is to go for URL rewriting for session implementation but we don't want to go for this implementation because it has some inherent drawbacks,moreover it would require a huge effort on our part because our site is pretty big. Any thoughts on this problem or anybody else has also faced this kind of the problem.??? Thanks in advance Sudhakar
Hi, I've got this trouble using Tomcat 4.0 . The sessions references sometimes simply desapeared or get invalidated nearly than it should be. The solution was implement a session manager that is executed by a Thread, and takes care everything about sessions in each 2 minutes. I know it's hardous, but it was the only thing that worked. Claudio Gualberto
|
s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440645315227.83/warc/CC-MAIN-20150827031515-00031-ip-10-171-96-226.ec2.internal.warc.gz
|
CC-MAIN-2015-35
| 1,586
| 3
|