Unnamed: 0
int64
0
3k
title
stringlengths
4
200
text
stringlengths
21
100k
url
stringlengths
45
535
authors
stringlengths
2
56
timestamp
stringlengths
19
32
tags
stringlengths
14
131
1,900
The implications of M1 Apple Silicon for Web Developers
The web in context The implications of M1 Apple Silicon for Web Developers If you’ve been on a social media blackout over the past few weeks then you may have missed the announcement, Apple’s first generation of ARM SOCs are out. The M1 is now available in three Macs and first impressions are positive. The change will have a positive impact on consumers but what about web developers? What does this shift in architecture mean for developers working on the web with JavaScript, HTML and CSS? Here are a few implications that M1 will have on developers. Availability of tools Apple has started off strong by ensuring there is a good compatibility story when it comes to the transition from x86 to ARM based architecture. Most consumer facing applications will work through Rosetta 2, a translation layer, just fine and will likely see performance improvements once developers make Apple Silicon optimised builds. This covers many use cases but as a developer, there are some gaps. Developer Tooling is incompatible on M1 Not everything is working on Apple Silicon and it will likely be that way for some time as developers get their hands on M1 equipped computers to test and rebuild tools. At the time of writing, here’s how developer tools are fairing. The Good Let’s start with the positive, here are some relevant tools you might use as a web developer. Electron : if you’re making cross platform apps with Electron, v11.0 has full native support. : if you’re making cross platform apps with Electron, v11.0 has full native support. Python : I’m including this because it’s a common language used for web servers, Python v3.8, v3.9 and even v2.7 are compatible. : I’m including this because it’s a common language used for web servers, Python v3.8, v3.9 and even v2.7 are compatible. Sublime Text: This editor isn’t native but it works via Rosetta 2 The Bad These are some common tools you might be using as part of daily development that aren’t working. Docker: This is a big one if you rely on spinning up your infrastructure for local development. It’s currently in available in the Preview Program. The Ugly Finally, the mixed bag. These two are very popular and have support coming but they just aren’t ready yet for everyone. Node : v15 is supported but previous versions will require patches which aren’t ready yet. If you’re a node developer on an LTS release, you’re out of luck right now. : v15 is supported but previous versions will require patches which aren’t ready yet. If you’re a node developer on an LTS release, you’re out of luck right now. VS Code : Insider builds are working on M1 but so far, no final release. It’s only a matter of time with Electron having M1 support. : Insider builds are working on M1 but so far, no final release. It’s only a matter of time with Electron having M1 support. Atom: Still quite a popular editor, it’s available via Rosetta 2 but it’s unclear when full support will be added. Take a look at “Does it ARM” for compatibility of specific developer tools to see if they’ll work on a shiny new M1 equipped Mac. What does this mean for Linux? There are implications across the industry because Apple is a large company with a lot of sway. A positive flow-on effect of putting powerful ARM processors in the hands of developers comes with an easier path for ARM tooling on other platforms. ARM isn’t new and there are lots of popular devices that are built with these chips like the Raspberry Pi. The difference is that until now, the majority have been low powered devices. My hope is that with ARM processors in the hands of more developers, the barrier to making tools compatible with other devices will be lower than it is today. This might take the form of more optimised compilers, programming languages and tooling that is more efficient than what we see with x86 version today. This also introduces an issue, however, compatibility. Incompatibility across systems I’ve already discussed the issues that now exist when trying to run software on new Apple Silicon devices but what about the other way around? This migration away from what has been a common architecture for decades, will create incompatibility and limit interoperability between computers. When Apple moved to Intel based Macs, one compatibility advantage comes to mind immediately, video games. PC is still dominant in this field but due to the similarity between Macs and Windows PCs, developers have been more likely to release their games to both platforms. The barrier for entry is relatively low. The new architecture shift is creating another platform for developer’s to support and it’s unlikely that this is a a market share large enough to pay attention to. It’s also unclear if Apple Silicon will be powerful enough to deal with games that are more graphically intensive and visually demanding. Game developers will need to make a decision on what path they take for cross-platform titles and it might not be worth it right now. New possibilities M1 promotional image, credit: Apple There are some possibilities for Apple Silicon devices that I think are going to shake things up in the industry. I’m not sure how many of them will reach web developers but they’re tantalising all the same. Media enhancements for safari WebKit, the engine that powers Safari, now has support for a new media query. It’s to detect if the display supports HDR (High Dynamic Range). <style> @media only screen (dynamic-range: high) { /* HDR-only CSS rules */ } </style> <script> if (window.matchMedia("dynamic-range: high")) { // HDR-specific JavaScript } </script> It’s pretty mundane at the moment but there could be some great potential to expand in the future. Web developers have a knack for taking simple queries like these and crafting compelling experiences, one example that comes to mind is light and dark mode with CSS Media Queries, based on prefers-color-scheme . Low power/high efficiency The biggest advantage that Apple is toting with Apple Silicon is power efficiency but how does that affect web developers? We often have our machines plugged in to electricity the majority of the time. This is a win for users who can browse more websites with less battery drain while still having performant systems when needed. I’m really interested in if there will be more support for leveraging the different types of cores, either directly with an API or indirectly via the browser. There are two types of cores in M1 chips, high efficiency and high performance. High efficiency cores are for battery, they are lower power drain cores. What if we could offload background tasks to these cores? As web developers we can make websites that have smaller bundle size, more optimised images, more efficient code but we can’t directly save someones battery life. If browsers start to utilise this efficiency or even better, allow developers to use these cores too, we can create a whole new metric for how performance on websites is measured. Media encoding/decoding The general purpose CPUs and GPUs of computer architecture we think of today have some downsides when compared with more dedicated chips. The M1 has dedicated support for ending and decoding media like video, images and audio with lower power consumption. It’s unclear how this will impact web developers right now but I hope that this will lead to better video codecs and file formats that are more smaller even if they do require more processing on the client. With a wider range of options, as web developers, we can provide experiences that cater to low bandwidth environments more easily. Conclusion I’m looking forward to seeing how this shift plays out and that impacts it has on users and developers alike. There are some opportunities around performance and efficiency that I hope we can take advantage of soon but there are going to be some teething problems until the tooling we use everyday, is compatible with Apple Silicon. Let’s see if developers take to the platform and if user’s expectations of efficiency change too. References Updates Since this article has been written, the following changes have been made:
https://medium.com/front-end-field-guide/the-implications-of-m1-apple-silicon-for-web-developers-5b272e4b21dd
['Seth Corker']
2021-01-17 11:59:43.047000+00:00
['Apple', 'Software Development', 'Apple Silicon', 'Web', 'Technology']
1,901
5 Considerations for Building a 5-Star FireTV App
The Android operating system (OS) is being used across multiple devices and platforms and is currently the most popular mobile operating system. At the moment, Android powers more than 2 billion devices and many of those devices operate on variations of the Android software development kit (SDK), such as Amazon’s FireTV OS, Nokia’s X platform, and Alibaba’s Aliyun OS, to name a few. As a result, with multiple architecturally various applications that can be built, they share the same set of application programming interfaces (APIs) from the Android SDK. After years of experience developing exclusively for Android mobile devices, I’ve come up with certain development patterns that can be replicated to help bring a quality mobile app to market. On the other hand, FireTV development requires some additional review and slight adjustments to those patterns based on the specifics of this platform. In this article, we will look into some aspects of FireTV development and some of the lessons learned from our experience developing a 5-star app for one of the biggest media providers. Performance At the time of writing this article, from the hardware perspective, FireTVs are less performant devices compared to the most up-to-date mobile Android phones. This means a developer will need to take a more diligent approach with regards to memory allocation, data processing, and algorithms while developing a FireTV app. These technical limitations often cause image stuttering and slowness. To avoid these issues, the best approach is to make as much data processing work as possible on the server-side, and send only the necessary data through the RESTful API. This will avoid unnecessary sorting and filtering on the client side that is both memory- and processing power-expensive. Power Supply Unlike mobile devices, FireTVs have uninterrupted access to a power supply and at first sight, may not require battery saving related optimizations. However, most of the battery-expensive work is related to the central processing unit’s (CPU) usage and network connectivity, which are the same factors that impact performance. This means that even though FireTVs do not have batteries and have permanent access to a power supply, implementing the necessary optimizations will considerably help with performance and will help avoid any lags in the user interface (UI) rendering. Network Connection Another out of the box advantage that comes with FireTV devices, unlike mobile phones, is the reliable, fast, relatively inexpensive and large bandwidth network connection that gives developers a bit more freedom in architecting the app. Depending on the case, engineers can reduce caching size and can rely more often on data updates from the network. They will not need to worry about network costs, bandwidth or reliability. However, consider making non-urgent network updates while the app is closed with Android’s WorkManager, this will help refresh, process and prepare the data before the user opens the app, and will avoid additional resource allocation when the user re-opens the app. Overall App Architecture FireTV applications have similarities and differences relative to “traditional” mobile application architectures. Networking and caching layer architectures can be lifted and shifted. These components can be used as is and do not require any adjustments or modifications. The major adjustments and differences revolve around the user experience (UX) and user interface (UI). FireTV does not have touch screen functionality and works exclusively with the remote. This fact requires engineers to follow guidelines for the cursor movement and UI for the selected states. The fastest way to include the UI elements into the app is by using the Leanback library, which has built-in navigation, however, it may be a bit limited in terms of customizations. User Interface FireTVs have a so-called “10-foot” user interface because the screen is roughly 10-feet from the user’s eyes, versus the 2-foot distance of a computer screen. This means some additional considerations must be taken into account in order to accommodate the distance and provide the right user experience. Developers should consider using appropriate (i.e.: larger) sizes for UI elements and fonts so they can easily be seen from a relatively longer distance. Also, make sure that every remote input can easily be reflected on the screen and is visible from a 10-foot distance. However, do not use larger assets than what is needed as this may negatively impact the performance and slow down UI rendering.
https://medium.com/tribalscale/5-considerations-for-building-a-5-star-firetv-app-8d2456a81513
['Tribalscale Inc.']
2019-06-20 14:37:50.729000+00:00
['Mobile App Development', 'Technology', 'OTT', 'Fire Tv', 'Android']
1,902
Exploring How Smart Infrastructure Can Help Ford Build A Great Self-Driving Service for Miamians
How this could benefit self-driving vehicle development: We want to understand how smart infrastructure could provide our self-driving vehicles with as much information as possible to navigate complex intersections. This could ultimately help customers by allowing us to expand service areas for our business. The intersection at Lincoln Road and Lenox Avenue in Miami Beach is a narrow two-way street with buildings close to the crosswalk. Whether it’s due to road geometry or buildings close to the curb, complex intersections like this could minimize the sensor view of a self-driving vehicle — which is really difficult for human drivers to navigate. In cases like these, a smart node can provide a comprehensive view of the environment to the vehicle before it arrives at the intersection. The vehicle can then use this additional information in its decision-making process. For example, if the vehicle sensors’ direct line of site is blocked due to a building close to the curb. A smart node could provide information about what is behind that building so that it’s aware of what’s in the environment well in advance and prepare accordingly. But, what information is the node gathering? Our sensor node is not capturing any data that is personally identifiable. All data from the smart node is anonymized to ensure privacy is maintained. This work is part of an advanced research project and we are focused on the learnings for how this could potentially help create a great self-driving service. The Ford approach — working collaboratively: The most important piece of our approach to building a self-driving service is to make sure we are working collaboratively with the cities and states in which we operate, to ensure we can co-create a vision for the future and share lessons learned. Miami-Dade County, the City of Miami Beach and the Florida Department of Transportation have been key collaborators in developing this research project and we will continue to work closely with them on this. What Miami-Dade County is saying: Miami-Dade County’s Assistant Director of the Department of Transportation and Public Works, Carlos Cruz-Casas has been working with us since day one. “Many have said Miami is the ‘next’ tech hub and we are no stranger to innovating and thinking about new ways for our city to move,” said Carlos Cruz-Casas. “Together with Ford over the last three years, we have been highly collaborative partners to ensure we safely launch self-driving services for Miamians. “One part of this, is making sure we support their research and testing so that when commercial services come online, we provide an efficient and safe way for people to hail a ride or have orders delivered,” he added. “We have worked with Ford to ensure they have satisfied all approvals and permits necessary, and we are thrilled Ford chose Miami-Dade for this first-of-its-kind research.” What’s Next: Ford is committed to launching a trusted commercial self-driving service in Miami-Dade in 2022. Our goal is to continue to test emerging technology, like the smart node, to ensure we provide a great experience for Miamians in the future.
https://medium.com/self-driven/exploring-how-smart-infrastructure-can-help-ford-build-a-great-self-driving-service-for-miamians-d1f6fccd0c14
['Ford Motor Company']
2021-03-31 15:32:41.307000+00:00
['Automation', 'Smart Cities', 'Technology', 'Self Driving Cars', 'Business']
1,903
Time to use IT for achieving equality in education
Time to use IT for achieving equality in education Since the old days, education has provided momentum for development on two levels; the societal level and the individual level. While nurturing outstanding individuals so that they can contribute to society, education has also played a main part in rearing a social ladder that enables those from underprivileged backgrounds to climb up. However, due to a growing economic gap, the faith in education’s role of providing class mobility is increasingly being considered a mere “mythology” these days. Globally, it is being proven by multiple statistics that the economic disparity between countries or regions ultimately leads to the gap in educational achievement. For instance, as of 2014, over 70 per cent of the global out-of-school population in primary and secondary education were from sub-Saharan Africa and Southern Asia.[ii] What about the domestic situation within states? The only difference with the educational inequality on a global scale is that the income gap of families instead of countries brings on difference in education. In fact, the notion that the privileged are prone to get better and higher level of education is being solidified as de facto idee recue in every country. In all countries with data, children from the richest 20 per cent of households achieved greater proficiency in reading at the end of their primary and lower secondary education than children from the poorest 20 per cent of households, and in most countries with data, urban children scored higher in reading than rural children.[iii] Where did it go wrong? It differs by the wealth of each country. For developing countries, the shortage of skilled teachers and school facilities are to blame. The situation in sub-Saharan Africa directly shows this; sub-Saharan Africa, which, as above mentioned, is suffering from extremely high rate of out-of- school youths, has relatively low percentages of trained teachers in pre-primary, primary and secondary education (44 per cent, 74 per cent and 55 per cent respectively), while the majority of schools in the region do not have access to electricity or potable water.[iv] For developed countries, disadvantaged family backgrounds should be held liable. Even with skilled teachers and excellent facilities, children still need an opportunity to exercise what they have learned at school so that they can fully assimilate. In this aspect, parental backgrounds can have a huge impact on children’s accomplishment (see Pierre Bourdieu & Jean-Claude Passeron, Les héritiers: les étudiants et la culture (1964)). What is fortunate is that there is a common solution that can solve both problems at the same time; internet. Thus, I would like to take this opportunity to introduce the solution which South Korea has adopted. In South Korea, there is a special television and radio network only for the educational purpose called “EBS” (acronym for Korea Educational Broadcasting System). EBS strives to supplement school education and promote lifelong education for anyone who yearns to learn. Fig. 1 EBS Logo But what makes EBS distinguished from other countries’ educational television networks is its effective utilisation of internet; EBS provides internet lectures on all levels of education, from primary to high school. EBS provides not only one-way lectures, but also a platform for interactive learning. If students leave any questions on a Q&A bulletin board, lecturers answer them as soon as possible. Also, each and every affiliate service of EBS is specialised for its target group. For example, in case of EBS Primary, not only students but their parents can also consult with teachers. Moreover, as EBS provides differentiated education, every student from low to high achievement can be the beneficiaries. But above all, EBSi, the service for high school students, is the most popular. Figure 2 Ebooks provided by EBS EBSi covers every single subject that students can learn in high school curriculum, including but not limited to Korean language, Korean history, mathematics, English, nine social sciences subjects, four sciences subjects, and nine second foreign language courses. Students have a wide range of selection as EBSi hires many skilled teachers from high schools or for-profit private academies. Indeed, a lot of students are taking advantage of EBSi. Especially for those who live in isolated regions, or for those who cannot afford private lectures, EBSi can be their saviours. Figure 3 One of the English lectures provided by EBSi played on iPad Some might say that the EBSi-model lacks universal applicability because it requires high speed internet. True, EBSi has been successful thanks to the wide-spread internet network of South Korea with the fastest internet speed in the world as of 2015[v] and the highest rate of internet use per inhabitant as of 2015.[vi] However, what should be taken into account is that EBSi launched its service no less than 15 years ago in 2004, when less than 70 per cent of Korean households possessed internet accessible electronic devices.[vii] To put it bluntly, the current situation in the poorest parts of the globe is not much different from the internet situation of South Korea in 2004. According to the data from 65 developing countries, the average percentage of schools with access to computers and internet for teaching purposes is above 60 per cent in both primary and secondary education.[viii] Of course, there is a huge difference between having a computer at home and at school, but still, I assume that it will be much easier and more efficient to install internet networks than training teachers and building new facilities. Still, even if every household around the world has a computer, what is the point if lectures are provided at a high cost? In case of South Korea, the budget for EBS is comprised of money from various sources, but most of them are from the government.[ix] In fact, the easiest way to finance the necessary budget is to draw taxes; for example, there is a unique tax named “education tax” in South Korea. It was first introduced in 1958 in order to raise funds required for education, and the reason why it has not roused taxpayers’ antipathy until now is because it takes the form of indirect tax.[x] It is hard for people to realise that they are paying “education tax.” I think this can set a good precedent. However, some countries can still find it hard to meet the demands of students as they have to hire teachers for various subjects. This is the part where higher education plays a role. In South Korea, all the students affiliated to College of Education shall have four credits or more of practice teaching in order to graduate.[xi] Four credits are equivalent to approximately four weeks of real teaching experience. For four weeks, these students, as preservice teachers, have opportunities to actually teach real students on the front lines. Then what if we let these students of College of Education film their lectures for mandated time and recognise their credits? If conventional Open Course Ware (OCW) classes are provided by professors, this can be referred to as students- made OCWs. Of course, I cannot say that my solution is perfect. In fact, South Korea also suffers from educational inequality despite all these efforts. But I venture to say that without the help of information technology, the situation could have been much worse. Thus, it is my belief that with the help of internet, mankind can take a step forward in its long-running strenuous effort to achieve educational equality, both on international and interfamilial level. “Knowledge is power. Information is liberating.” Kofi Annan, the seventh Secretary-General of the United Nations, said, “Knowledge is power. Information is liberating.” Education is the premise of progress, in every society, in every family.[xii] In the midst of the wave of neoliberalism, deteriorating educational equality might seem inevitable, but the 21st century is also the age of information technology. It is time to make the most out of modern civilisation to achieve equality in education.
https://medium.com/asefedu/time-to-use-it-for-achieving-equality-in-education-a468afc0d192
['Asefedu', 'Editor']
2019-07-22 07:16:08.290000+00:00
['Sustainable Development', 'Technology', 'Sdgs', 'Arc']
1,904
friday lost and found: touch my lion edition
[caption id=”attachment_947" align=”alignnone” width=”600" caption=”Touch my Lion! Love it! Liebe mein affe-Lion!”] [/caption] Apple released Lion this week and MG Siegler immediately made sweet, sweet love to it. More useful than @parislemon’s 3,000 word opus: Greg Kumparak’s, “Nine Things you Should Do After Installing Lion” and Cult of Mac’s ludicrously comprehensive, six-page review/manual. Via TechChrunch, Cult of Mac *** Speaking at Fortune BrainstormTech, Twitter CEO Dick Costolo said the company will eventually offer self-serve ads for brands and is looking to get into the commerce game. Twitter, he said, wants to “remove friction” from the process of buying stuff on the platform. If I may offer a humble suggestion from a lowly non-profit marketer: Can you please also help “remove friction” from the process of donating to charities through Twitter? Via Fortune *** Mmm…. That’s some good Gojee. I’ve seen a lot of people using Gojee recently and am trying it out myself — I’ve been in a bit of a cooking rut these days. The service lets users input ingredients they have in their pantry and spits out recipes (an gorgeous photos) based on those ingredients. I wonder what it will suggest for pita chips, Kraft mac & cheese and craisins? Seriously though, I just typed in “black beans” and favorited 10 recipes. This thing is seriously addictive. *** Dude… your Prius is pretty badass. Toyota responds to the Honda CR-Z Sport hybrid with the Prius Performance Package. For an extra three grand, you get front and rear spoilers, 17-inch alloy wheels, custom tires, a tuned rear sway bar and…. wait for it… limited edition floor mats. w00t! Via PSFK *** Finally, start your weekend off right with this awesome video of the Sesame Street Muppets rockin’ the Sure Shot, by the Beastie Boys: Sesame Street breaks it down from Wonderful Creative on Vimeo. Via LaughingSquid
https://medium.com/david-connell/friday-lost-and-found-touch-my-lion-edition-a202f72d24c7
['David Connell']
2016-08-20 13:06:15.101000+00:00
['Nonprofit', 'Technology', 'Green Tech', 'Environment', 'Links']
1,905
Bad Blood by John Carreyrou— Notes
Outline Elizabeth Holmes dropped out of her Stanford undergraduate programme and started Theranos at the age of 19. The company claimed to revolutionise blood testing by taking very small amounts of blood from patients using a fingerprick and perform a large variety of tests. She claimed the results would be instantaneous instead of the traditional approach where it usually takes days. This technology is very useful for pharmaceutical companies as it will help them to speed up drug trial tests helping them to instantly change the dosage for the patient. It also has huge implications for doctors as more than 70% of their decisions are based on blood tests. In 2015, the company had a more than $9 Billion valuation and Elizabeth was the youngest self-made woman billionaire. She was an inspiration to a lot of young women in the field of medicine and was all over the media for creating a huge technological shift in the field of medicine. But very soon the reality came out and Theranos became a defunct healthcare company and soon Holmes net worth became almost zero. To know-how let’s go deeper and understand the story from different angles. Family and Loyalists Elizabeth was a straight-A student at a young age. Her family was very well known and her dad ensured that she got the best private school education. From her childhood, she wanted to do something different and was very much inspired by the stories from Silicon Valley. Her parents wanted her to do a PhD in basic science but Elizabeth always knew that she wanted to start a company and make money. One of the main characters in the story was the Fuisz family. Elizabeth’s mother and Lorraine Fuisz were very good friends. Lorraine’s husband Richard Fuisz, a doctor by profession was an extremely proud and wealthy investor known for filing patents which he then licenses it to make profits out of it. He came to know about Theranos after his wife told a conversation he had with Mrs Holmes and was really annoyed on the fact that even though they knew each other very well they didn’t consult him before starting the company. He patented a process which later became very crucial for Theranos. They later filed a lawsuit against him to revoke his patent which he had to do finally. Theranos’ legal team well very known and feared and was one of the primary reasons they were able to go undercover without any public scrutiny. Richard Fuisz later grouped together with ex-employees of the company to go against Theranos and their malpractices which I will go in detail in the next section. The right hand of Elizabeth was Sunny Balwani, an Indian who became extremely wealthy during the dot-com bubble. It was also said that Elizabeth and Sunny were romantically involved with each other, something which they kept it as a secret with their employees and board members. He was extremely bossy and tried to create an atmosphere of fear so that everyone knew that it was their privilege they worked in such a great startup. He tried to monitor everyone’s dedication by seeing their computer activity and the number of hours they dedicated every day. He was one of the primary reasons why people left the company. Even though he did not understand the field properly he always boasted regarding the same and was notorious for not giving even a proper farewell to most of his employees when they resigned. Theranos Employees Elizabeth Holmes energy and charisma attracted a lot of good employees who were quite driven by her mission to solve the healthcare problem. Since she always had the backing of professors from her alma mater they were able to recommend some of the best students from Stanford to Theranos. She started the company with Shaunak Roy whom she met at Stanford. Very soon she recruited the first engineering team who were responsible for building ‘Edison’ company’s first commercial product. She was greatly inspired by Steve Jobs and tried to emulate him as much as possible. She wore black turtlenecks something which Steve Jobs was famous for wearing during product announcements. Her vision was just like Apple that everyone should have an Edison at their home, an extremely easy to use, portable and versatile device. She was able to attract a lot of employees from Apple like Ana Arriola who became the Product Design head for Edison. The company never allowed inter compartment communication and she pushed to have information as siloed as possible. This often led to miscommunication and did not allow anyone to have a full picture of what’s happening in the company. This led to many employees leaving the company as there was an atmosphere of constant fear and intimidation. Once the board even decided to remove her as CEO because of the complaints from the Sales and Legal team as she was trying to sell products which were not even tested and gave false projections to clients which were impossible to achieve. Throughout the board meeting, she convinced the members that it was a mistake and will try to compensate for it. They gave her a second chance and she used this opportunity to fire anyone who was up against her. Employees regularly came and left the company within a very short period of time but this was all done secretly to raise no suspicions within the company or outside. In fact, her employees were not allowed to write the company’s name in their Linkedin profile. Edison was only able to conduct immunoassays which was just a small part of all the tests which Theranos pitched. Soon, they started building MiniLab where the idea was that all the test could be conducted by this machine. They code-named it ‘4S’ a tribute to Apple’s iPhone. Soon Sunny started hiring more Indians to the companies as they all had H1-B visas and were more likely to stay within the company. He knew all of them wanted the company experience to stay in the U.S and showed more loyalty to him as compared to the American counterparts. Things took a severe turn when Ian Gibbons joined the microbiology team from the Dean of Stanford’s recommendation but soon came to know that things were not being done ethically in the company. He had constant fights with the engineering team and slowly felt that no one was listening to him in the company. Because of his old age, he felt more desperate about the job and soon committed suicide. The company officials still were more worried about whether there were any company trade secrets left with him. Investors and Clients Elizabeth got a lot of support in the beginning from the Dean of Stanford, Channing Robertson, Larry Elison founder of Oracle and Tim Draper founder of a prominent VC firm. They were one of the first few investors in the company. When Elizabeth introduced the device to the military, General Mattis (Secretary of Defense under the Trump administration) was very impressed by it as it had immense potential on the battlefield for all the soldiers as this would help soldiers to recover quickly. He soon joined as the board member of the company and was followed by Henry Kissinger, George Schultz who was former Secretary of State of the country. All these investors added a lot of credibility to the company. She mainly targeted government officials as it gave more public credibility and helped her to avoid scrutinies from public agencies. But no one noticed all the board members were from the non-medical field and didn’t really understand the product or the field. All of them thought that Elizabeth was the next Jobs and really wanted to be part of it. Since she always had the majority shares she took the majority of the decisions in the company and considered board members nothing equivalent to placeholders. Some of the biggest deals Theranos had with was Walgreens and Safeway. Walgreens was seeing stagnation in their revenues and wanted to pivot out from their primary source which was from drugstore sales. They upgraded their stores to have Theranos clinics which took blood tests of visiting customers. She convinced both the companies that their device Edison was able to perform to a wide variety of tests using a fingerstick. The company received combined more than $200 million in investments and spent even more in upgrading their stores. Theranos didn’t conduct the majority of the tests on their devices rather they took the blood samples from the stores and tested on modified machines they bought from Siemens and other third-party devices. Most of the result showed false values as they regularly diluted the blood samples and tests were conducted by certified lab professionals. Whistleblowers, Author and the Lawyers Deceased Ian Gibbon’s wife, Richard Fuisz and Alan Beam who was heading the laboratory department in Theranos initially reached out to the author who is a Wall Street Journal journalist regarding the company’s false claims. The author then tried to find out more employees to corroborate their story. He reached out to Erika and Tyler who left the company because they knew that what was happening in the labs were not ethically right. Cherry-picking test results, hiding actual test devices when authorities came for checking, using false approval claims of FDA and using competitor’s devices for testing the blood while being given a completely different picture to the investors. The author tried to set up meetings with the doctors in Walgreens and Safeway who used Theranos products and it was true that even their patients got false results when comparing with other traditional means. Most of the wellness centre used both intra-venous and finger stick method which was not being told to the board members and there was enough evidence that a story could be made. But very soon the lawyers came to know about the whistleblowers, David Boies one of the most feared lawyers was taking up Theranos’ case. His firm was known for its aggressive tactics and using private investigators to hunt and intimidate people. Many wondered how the company were able to pay the legal fees associated with it, they didn’t as Boies was given some % of shares of the company. All the employees were told to sign a non-disclosure agreement while leaving which prevented them from sharing any info of the company to the public so Tyler, Erika and Alan were threatened with a lawsuit of sharing this info to a journalist. The story is out… The story was published on the Journal’s front page on Thursday, October 15th, 2015. The headline, “ A Prized Startup’s Struggles”, was understated but the article itself was devastating. Prior to releasing the story the legal team of David Boies regularly contacted the author to return back all the content which was against the company rules as it was going against the non-disclosure agreements of the employees. They used a lot of threatening approaches even after the first article was released but the damage was already done. FDA and other medical agencies made secret visits to the labs and found a lot of unauthorized devices and banned them from using it again. On March 14, 2018, the Securities and Exchange Commission charged Theranos, Holmes and Balwani with conducting “an elaborate, years-long fraud.” To resolve the agency’s civil charges, Holmes was forced to relinquish her voting control over the company, give back a big chunk of her stock and pay a $500,000 penalty. She also agreed to be barred from being an officer or director in a public company for ten years. Favourite quotes and terminologies Folie a deux: delusion or mental illness shared by two people in close association. A classic example which Sunny and Elizabeth share among each other. La mattanza: It was an ancient Sicilian ritual in which fishermen waded into the Mediterranean Sea up to their waist with clubs and spears and then stood still for hours on end until the fish no longer noticed their presence. Eventually, when enough fish had gathered around them, someone gave an imperceptible signal and in a split second the scene went from preternatural quiet to a gory bloodbath as the fishermen struck viciously at their unsuspecting quarry. Mike, editor of WSJ gave this example to the author as an analogy where they patiently waited until they were ready to publish. Vaporware: It was used to describe a new computer software or hardware that was announced with a lot of fanfare only to take years to materialize. It was a reflection of the computer industry’s tendency to play it fast and loose when it came to marketing. A similar strategy was used by Theranos but they were in the medical industry which was highly regulated and a lot of patients life were at risk.
https://medium.com/@martinzachthomas/bad-blood-by-john-carreyrou-notes-b3feab3f0d24
['Martin Z Thomas']
2020-12-11 17:53:19.679000+00:00
['Investors', 'Technology', 'Silicon Valley', 'Theranos', 'Healthcare']
1,906
Why Your iPhone is Slow
#3. Check the Memory Usage The phone may be performing slow if the amount of free memory is low. If you’ve had the iPhone for over a year, take a lot of photos, or have a lot of apps installed, checking the memory usage periodically is good practice. Unused apps and photos that take up a lot of memory should be deleted. The iPhone storage can be managed by going to: Settings > General > iPhone Storage Image provided by author Delete unused apps As shown above, the iPhone storage screen provides a breakdown of the memory usage per app. If you choose to delete an app, select it from the list and press the Delete App option. Image provided by author 2. Delete Messages, Photos and Music Messages, photos and music all contribute to the overall iPhone storage usage. Selecting Messages from the iPhone Storage list gives the top conversations, photos, and GIFs/Stickers that can be deleted. Selecting Music displays all of the songs and memory size, with the option to delete. Photos will need to be deleted through the Photos app. 3. Clear “Other”
https://medium.com/macoclock/why-your-iphone-is-slow-a7edd7f23ebc
['Julie Elise']
2020-12-24 05:36:17.749000+00:00
['Technology', 'iOS', 'Productivity', 'iPhone', 'Phone']
1,907
Education Ecosystem Proof of Work: Run Project Quality Check and Earn LEDU
Education Ecosystem is a decentralized learning ecosystem that teaches people how to build complete products in future technology fields. Building a project-based learning platform to tackle the $360B professional development market requires not only a great team but also the support of a great community. Another area for our community to participate and earn LEDU is with our next Proof of Work initiative, running project quality checks. Site moderators on Education Ecosystem are independent bounty hunters who conduct quality assurance checks. Quality assurance encompasses reporting bugs, technical streaming issues, video quality issues and content moderation. Education Ecosystem moderators will be allocated tokens for each quality assurance activity. Quality assurance is important as it improves user experience for all subscribers and project creators. Site moderators represent an integral part of the Education Ecosystem community and without them the site cannot run. For this reason they are rewarded with LEDU tokens for carrying out site moderation weekly. How to Become a Site Moderator Prospective site moderators will first need to apply by submitting this Google Form. To qualify as a site moderator an applicant should first share the passion and vision for Education Ecosystem so that they can explain why they would make a great moderator. The applicant should have experience in a category on our platform. Having experience in quality assurance is a bonus. After an application is received qualified applicants will interviewed and then notified if they are selected. How to Report Quality Issues Moderators run random quality checks on the Education Ecosystem learning platform. A Quality Assurance (QA) Checksheet is provided to moderators which they can fill in and report any issues found. If other issues are found they can also be reported by the moderator. These issues are then double checked by our quality assurance team. How to Earn LEDU Site moderators will be rewarded 50 LEDU tokens for each confirmed quality issue. Rewards are accumulated and moderators are paid at the end of each month. Site moderator performance is evaluated monthly and a moderator must maintain active status to continue receiving LEDU rewards. Once 80% of quality issues reported each month are confirmed as real issues by our quality assurance team then the moderator is considered active. Do you hate finding bugs on a website? Do you wish you could help get rid of these bugs and earn LEDU for doing so? Apply to become a site moderator. Get LEDU Coin Get LEDU coins now on Exrates, Livecoin, Mercatox and IDEX or join the LEDU OTC Trading program for large purchases. Read more about LEDU coins on our project page and ask any questions you might have in our Telegram group chat.
https://medium.com/ledu-tokens/education-ecosystem-proof-of-work-run-project-quality-check-and-earn-ledu-faddf40e5dc6
['Dr. Michael J. Garbade']
2019-01-11 19:47:01.566000+00:00
['Quality Assurance', 'Education', 'Blockchain Technology', 'Software Development']
1,908
Sony’s shift towards premium TVs meant lower Q2 FY19 sales
And sticking to basic math, Sony is fast approaching a point that even if each device they sell holds a 100% profit margin (which is of course impossible), the cost of running the division (R&D, marketing, inventory, components, etc.) would be higher than the return. Could TVs be headed towards a similar fate? It’s still early and there’s yet another transition ahead of the industry — 8K, though I doubt despite hype from companies that it’s coming any time soon, nor do I think consumers will care, seeing how so much content is still not offered even in 4K— but beyond that, unless there’s a massive shift in Sony’s strategy, it’s hard not to look 5 years ahead and see what happened to Mobile happen to their television business as well. Producing a TV and shipping it costs a hell of a lot more than it does for a smartphone and with sales likely topping off at around 10 million for FY19, when will the division not be generating enough profits to justify their entire lineup? At that point, the number of models Sony offers would be slashed. But then, with fewer models in the market, would they be conceding and exiting certain markets all together, which would result in lower revenues and thus, trapping them in a downward spiral? If Mobile is any indication, that answer is yes. For the past decade, Sony’s mantra towards solving its fledgeling electronics business has been — “lets cut costs” — and while I’m all about a company running a lean operation, you seldom see struggling companies recover by ‘cutting’ their way out. Instead, what each and every division within Sony needs is bold thinking — roadmaps that extend years ahead, taking them to where the puck will be, not where the puck is, marketing that excites consumers (none of this Kondo and make you feel bullshit), and ultimately quality products that embrace the connected, software driven world that we live in. Oh, and it wouldn’t hurt if Sony’s entire home theater lineup actually worked and complemented each other. With PS4, Sony understood this, having learned from their mistakes with PS3. But Sony Electronics continues to lack the vision that made Trinitron and Bravia household names and XBR, THE TV home theater enthusiasts lusted over. Sony still has a chance to make sure that in the future, when people are visiting the electronics museum, that their products don’t stop appearing after 2010, but that clock is ticking and I’m not sure current executives at the company can hear it.
https://sonyreconsidered.com/sonys-shift-towards-premium-tvs-meant-lower-q2-fy19-sales-abfc9667b301
['Sohrab Osati']
2019-11-01 19:04:21.447000+00:00
['Technology', 'Advertisement', 'Tech', 'Business', 'Home Theater']
1,909
My phone is not a fridge
One of the critical problems in technology today is a lack of empathy. People are stuck in their echo chambers in these hip tech companies staffed with young people of remarkably little cognitive and moral diversity. When their company’s workforce bears no resemblance to most of their customer base, it is no wonder they don’t get the problems. No matter how much user research they do, their customers will always be foreign to them, leaving plenty of room for projection where empathy should reign supreme. Creator responsibility Along the lines of a big push for sustainability, topics like extended producer responsibility are hotly discussed and debated. Essentially, the goal is to expand the responsibility a producer of goods has from mere warranty into areas of dealing with disposal and more. This is supposed to ensure that they already care about repairability and recyclability during the design and manufacturing process. We need a similar concept for creators of software systems and expand their responsibility for their creation. If a product team releases a feature that shows apparent harmful effects, they must be liable to fix it or otherwise face penalties. This is needed to address the diffusion of responsibility that happens in large companies, where in the end no one feels responsible for anything, and ethics is left unconsidered. That would also mean that there needs to be a monitoring of impact beyond just the company’s bottom-line and business metrics. Social and environmental responsibility means monitoring second-and third-order effects, which might sometimes be fuzzier than conversion numbers, to understand what impact a product has. Sounds complicated? No one said this would be easy. Or that it should be easy. Creation comes with enormous responsibility and should not be taken lightly. By bringing something into existence, you alter the nature of our reality. This should be complicated and well-considered. My phone is not a fridge. Photo by Alexandru Acea on Unsplash The producer of my fridge has a clear responsibility. Today it is the warranty they serve, tomorrow hopefully, the full recycling and disposal as well. Responsibility evolves as we learn that impact products have. What they are indeed not responsible for is what I put into the fridge — because it neither acts as a gatekeeper (like the AppStore), nor does it make recommendations or provides features for third parties to interact with me on a regular and intrusive basis. The producer of my smartphone today only takes responsibility for the hardware in terms of the warranty. Tomorrow hopefully also for the repairability of it. Soon the disposal as well. Yet, they also need to take responsibility for what they allow onto my phone because it passes their gates. And for what they enable those third-parties to do because they provide the infrastructure. We can, and should, hold companies like Facebook and Google accountable for their part in all of this. But that does not mean that someone like Apple should be off the hook. If a mobile app uses provided platform features to make it more addictive, someone like Apple is a part of that — and needs to own up to it. They have, admittedly, made relevant updates and moves in the last year. Quite late, and still too little. Creation should not be taken lightly. And responsibility does not expire.
https://uxdesign.cc/my-phone-is-not-a-fridge-9aa85a215f11
['Sebastian Mueller']
2021-02-15 13:12:44.781000+00:00
['Ethics', 'Design', 'Responsibility', 'Technology', 'Innovation']
1,910
The Differences Between Provider Pattern and Bloc Pattern
Building a login screen using Provider Pattern What are we trying to achieve here? We are trying to write the code in which as soon as user starts typing in the text field, the text field keeps showing the error message to the user as the email or password he is typing is a valid email or password. Based on that, if email and password match, the regex will allow the UI button to be enabled, making it clickable. login_screen.dart class LoginScreenWithProvider extends StatelessWidget { @override Widget build(BuildContext context) { return ChangeNotifierProvider<LoginProvider>( create: (context) => LoginProvider(), child: Scaffold( appBar: AppBar(title: Text(‘Login’)), body: Padding( padding: const EdgeInsets.all(8.0), child: ListView( children: [ // Email Field Consumer<LoginProvider>(builder: (context, provider, child) { return TextField( decoration: InputDecoration( labelText: “Email”, errorText: provider.email.error, ), onChanged: (String value) { print(‘provider.email.error ${provider.email.error}’); provider.changeEmail(value); }, ); }), ], ), )), ); } } Note: To do this, we will have to keep rebuilding the widget (in our case it is a text field) but not the whole widget tree to comply with programming best practices. You need to wrap the Scaffold widget inside the ChangeNotifierProvider widget, so that we have the instance of Provider object to listen the changes. Most importantly, the widget you want to rebuild should be the child of the Consumer widget, because it listens for changes in data should any event happen, so that it can rebuild and update the widget on change in data. The best use of the Consumer widget is to use it as deep as possible within the widget tree. The best practise here would be to rebuild the desired widget in the widget tree which needs the update. The best way to do that is to use the Consumer as deep as possible. Now we check for change notifier or provider class: class LoginProvider with ChangeNotifier { ValidationModel _email = ValidationModel(null,null); //Getters ValidationModel get email => _email; bool get isValid { if (_email.value != null){ return true; } else { return false; } } //Setters void changeEmail(String value){ if (ValidatorType.email.hasMatch(value)){ _email=ValidationModel(value,null); } else if (value.isEmpty){ _email=ValidationModel(null,null); } else { _email=ValidationModel(null, “Enter a valid email”); } notifyListeners(); } @override void dispose() { super.dispose(); } } As of now, for the sake of simplicity, we have picked Email text field only to check how it works. On every keyboard hit we call the changeEmail method to check if the email typed matches the regex or not. NotifyListeners() will keep notifying the Consumer widget to rebuild the text field widget and errorText will be updated accordingly. The password field also works in the same way: ListView( children: [ // Email Field Consumer<LoginProvider>(builder: (context, provider, child) { return TextField( decoration: InputDecoration( labelText: “Email”, errorText: provider.email.error, ), onChanged: (String value) { provider.changeEmail(value); }, ); }), // Password Field Consumer<LoginProvider>(builder: (context, provider, child) { return TextField( decoration: InputDecoration( labelText: “Password”, errorText: provider.password.error, ), onChanged: (String value) { provider.changePassword(value); }, ); }), //Update Indicator Consumer<LoginProvider>(builder: (context, provider, child) { return Container( height: 150, child: Center( child: provider.state == ViewState.Idle ? Container() : CircularProgressIndicator(), ), ); }), //Submit button will be enabled on correct format of email and password Consumer<LoginProvider>(builder: (context, provider, child) { return RaisedButton( color: Colors.blue, disabledColor: Colors.grey, child: Text( ‘Submit’, style: TextStyle(color: Colors.white), ), onPressed: (!provider.isValid) ? null : () { provider .submitLogin() .then((LoginResponse response) { print(‘response: ${response.token} ‘); Navigator.push( context, MaterialPageRoute( builder: (context) => DashboardScreen())); }); }, ); }) ], ) The login provider class will be: login_provider.dart class LoginProvider with ChangeNotifier implements LoaderState { LoginProvider(){ setState(ViewState.Idle); } ValidationModel _email = ValidationModel(null,null); ValidationModel _password = ValidationModel(null,null); //Getters ValidationModel get email => _email; ValidationModel get password => _password; bool get isValid { if (_password.value != null && _email.value != null){ return true; } else { return false; } } //Setters void changeEmail(String value){ if (ValidatorType.email.hasMatch(value)){ _email=ValidationModel(value,null); } else if (value.isEmpty){ _email=ValidationModel(null,null); } else { _email=ValidationModel(null, “Enter a valid email”); } notifyListeners(); } void changePassword(String value){ if (ValidatorType.password.hasMatch(value)){ _password=ValidationModel(value,null); } else if (value.isEmpty){ _password=ValidationModel(null,null); } else { _password=ValidationModel(null, “Must have at least 8 characters”); } notifyListeners(); } ViewState _state; @override void setState(ViewState viewState) { _state = viewState; notifyListeners(); } @override ViewState get state => _state; Future<LoginResponse> submitLogin() async { setState(ViewState.Busy); final Mapable response = await apiClient.serverDataProvider.login(_email.value, _password.value,); setState(ViewState.Idle); if (response is LoginResponse) { print(‘response.token ${response.token}’); return response; } } @override void dispose() { super.dispose(); } } Upon typing of email and password, isValid is getting updated. If isValid is received as true by Consumer with RaisedButton child, then the button gets enabled, else it will remain disabled.
https://medium.com/swlh/the-differences-between-provider-pattern-and-bloc-pattern-fc93dc523672
['Dlt Labs']
2020-06-30 03:08:32.541000+00:00
['Technology', 'Ios Development', 'Dltlabs', 'Programming', 'Android Development']
1,911
What Makes A Successful Data Scientist? 5 Traits to Success
What Makes A Successful Data Scientist? 5 Traits to Success Photo by Fab Lentz on Unsplash An important question that we all ask ourselves at some point through our data science learning journey is, “what do I need to do or learn to become successful in the field?”. I know I ask myself that equation in every learning journey I went through in life because when we stay on a learning journey, we often want that journey to end up with success. But, the tricky part about that equation is, the answer varies based on the person's perspective of success. For some, success is getting a high-pay job; for others, it’s learning something new or have a job where you learn a new thing every day. For me, it is to do something I love and get paid to do it. Not necessarily a big amount of money, but as long as I enjoy what I do, I am successful. If you ask any data scientist who considers themselves successful or who is viewed as such, they will probably tell you how did they achieve their success from the personal perspective that they gained throughout their journey. But, the interesting part is, although success is personal and subjective, if you ask 100 or even 1000 data scientists, we will end up with some common answers and some unique ones. Although the unique ones make each learning journey special, the common ones are the ones you probably need to focus on. Because throughout your journey, you will develop your own unique traits for success. So, in this article, I will focus on the top 5 traits that all data scientists would agree you need to obtain to succeed as a data scientist. №1: Have a curious mind hungry for knowledge I will order the traits of a successful data scientist based on my personal experience and what I believe is important yet not always addressed. The first trait any data scientist needs to succeed is curiosity and hunger to learn. As I always say, data science is an ongoing field; a lot is happing within it all the time. So, to succeed in this field, you need to be curious, curious about the next trend in the field, curious about the possible problems that you can solve, and curious about the different ways you can solve them. If you have a curious mind, you will always be learning and always be improving, both essential to succeed in data science. №2: Can Easily adapt to new data and situations Data science is an interdisciplinary field; it connects maths, statistics, business, programming, design, and science communication. Because of that, data science applications are endless; there are problems data science can solve in every field. So, adaptability is an important trait for any data scientist to obtain. Being able to adapt to new tools, to new algorithms, new data, new problems, new teammates, and new situations. This is getting more important if you are a freelance data scientist or changing jobs. The ability to apply your knowledge to any application and data can make you either build a great project or fail to come above the crowd. №3: Critical thinker and problem solver One of the main tasks expected from data scientists is to find trends, patterns and obtain insights from given data. The key to finding these insights and trends and making use of them is critical thinking and problem-solving techniques. Therefore, you will need to explore data, analyze it and think critically to decide on a model to use in further steps of the project. Problem-solving and critical thinking also help you decide on which tools to use and which decisions to make. Thus, utilizing these skills can help you save time and effort while working on your project. An important thing to mention here is, your skills to think about data critically will improve as you advance in your career. №4: Have good communication skills Having good communication skills is, in my opinion, one of the most essential yet overlooked soft skill every data scientist need to master. Imagine spending hours working on some data, analyzing it, cleaning it, visualizing it, yet you can’t explain your findings to others. Then, all your work would be for nothing because if they can’t understand your findings they won’t be able to act upon it. A data scientist needs to be good in science communications, and they need to be good at simplifying concepts and only focusing on what really matters. In addition, they should be able to give actionable advice and be good at utilizing tools to create effective visualizations that deliver a clear message. №5: Have solid technical skills I intentionally left technical skills to be the last on my list. That’s not to say that technical skills are not important. In fact, you probably can’t become a data scientist without solid technical skills. But, if you’re trying to get into data science or already have, and as the rest of us are working on your skills daily, then you know the technical skills are the core focus of many online materials. So, you will learn the technical skills anyway because that the skills that are often tested during the job hunting process. I put technical skills last because the other 4 traits I mentioned are often overlooked or not focused on, especially new data scientists. Yet, there are essential to building a career in data science. Takeaways We all want to be successful, work hard on ourselves, learn new skills, develop the ones we already have, and work on trying to reach a better version of ourselves all the time; that’s just a part of life. It is also an important part of succeeding in tech as a whole and data science in particular. Data science is an ongoing field; new algorithms and models are designed every day to offer better efficiency and performance and lower error rate. Also, because of the interdisciplinary nature of data science, you will need to learn various topics and concepts that may not seem related at first. Still, their connections appear once you start building projects and getting involved in the field. It’s important to remember that each has a unique learning journey, a unique experience, and a unique definition of success. Still, within this uniqueness, we all share some common traits that allow us to show our individuality, prove our skills, and become the best data scientist we could be, so, in short, these common traits allow us to succeed.
https://towardsdatascience.com/what-makes-a-successful-data-scientist-5-traits-to-success-b106e7bc7cf1
['Sara A. Metwalli']
2021-09-04 14:17:58.120000+00:00
['Data Science', 'Careers', 'Women In Tech', 'Technology']
1,912
Bitcoin: What is Bitcoin? Complete Guide 2021
Bitcoin : Definition Bitcoin: How does it work? Bitcoin: Decentralization Bitcoin: Who accepts Bitcoin? Bitcoin: Who determines the price of Bitcoin? We will try to understand and explain what Bitcoin is today; its characteristics, functionalities, usefulness, potential use and how it works. Bitcoin : Definition Etymologically, the word Bitcoin comes from the two words bit and coin; - Bit is a unit of binary information and - Coin designates a token. Bitcoin is an innovative payment network and a new form of unit of value. It is the very first virtual or cryptocurrency. Being completely virtual, Bitcoins do not exist in a physical form. These currencies are kept on a public book to which everyone has transparent access. In other words, Bitcoin is a crypto-active encryption that designates virtual assets stored on an electronic medium that allows a community of users to use it as a means of payment without having to resort to so-called “traditional” currencies such as the euro or the dollar. These crypto-actives or also called cryptocurrency are neither issued nor guaranteed by banks or governments. However, they can be used to make transactions such as purchasing products and services. Bitcoin: Why was Bitcoin created? We remember the financial crisis of 2008 that devastated the world economy; it was called a “krash”, marked by a liquidity crisis and sometimes solvency crises at both the bank and state levels, leading to a scarcity of credit to businesses and the insolvency of many borrowing households. The crash began with difficulties encountered by low-income American households who found themselves unable to repay the loans they had been granted for the purchase of their homes. Following this crisis, we can point out that: In the United States: 9 million homes foreclosed between 2009 and 2012, more than 8 million citizens lost their jobs and about 2.5 million businesses were closed. Faced with food insecurity and the resulting income inequality, people’s confidence in banks has been severely shaken. In Europe: Ireland is the first country on the European continent to enter recession. Its GDP fell by 3.5% and its public deficit reached 7.2% of GDP. In France, there were 800,000 more unemployed. The number of unemployed in the International Labor Office (ILO) rose from less than 2 million to nearly 2.8 million, an increase of 40% between mid-2008 and mid-2014. The unemployment rate rose from 7.3% to 10.2% of the labor force. In Asia: The crisis has hit most of Southeast Asia and Japan, which has seen the collapse of currencies, a stock market crash and a sharp increase in private debt. Taiwan saw its currency depreciate more than 45 percent against the dollar in the space of three weeks. This led to the so-called Great Recession, which caused property prices to fall and unemployment to rise sharply. In other words, the financial crisis has exposed the inherent shortcomings of banks and other financial institutions. After the financial crisis, people began to desire a new monetary system that would not have the same flaws as those with traditional currencies. Bitcoin, this cryptographic currency, came into being about two months after the crisis. But who is behind this new form of money? Bitcoin : Who invented BTC? Whoever created Bitcoin wanted to promise decentralized online payment with lower transaction fees than government-issued currency. The identity of the person(s) who created this technology remains a mystery. However, the creator chose a pseudonym: SATOSHI NAKAMOTO Nakamoto wanted to build a digital “cash” system, working peer-to-peer, like cash while remaining discreet and totally untraceable. In Bitcoin’s white paper, Nakamoto explains: “What is needed is an electronic payment system based on cryptographic evidence rather than trust, allowing two consenting parties to deal directly with each other without the need for a trusted third party”. Satoshi could be a group of people and appears for the first time on October 31, 2008 on the P2P Foundation bitcoin.org forum. A few messages recorded in his name made it possible to continue the search from his personal account. On his profile Nakamoto claimed to be a Japanese man born on April 5, 1975; Nakamoto also mentioned his first research on crypto-money and the design of Bitcoin from 2007. Bitcoin: How does it work? Bitcoin is not administered by a single central authority such as a bank, it works in a decentralized way through a set of nodes. These “nodes” form the network through which all transactions take place. But how does this work in practice? Bitcoin: Understanding the Bitcoin Blockchain Bitcoin is powered by an open-source code known as blockchain. The blockchain is a digital, shared, public ledger that supports the entire Bitcoin network. All confirmed transactions are included in the blockchain. Each transaction is listed in a “block” that is “chained” to the code, creating a permanent record of each transaction. Blockchain technology is at the heart of more than 2,200 crypto-currencies that have followed the development of Bitcoin. The function of the Bitcoin blockchain is to keep track of the transactions carried out on the network, while being impossible to tamper with. The Bitcoin blockchain eliminates the need for a central control body to validate transactions (the network nodes act as validators). The digital ledger is replicated on all the nodes of the Bitcoin network and takes the form of blocks that are linked together using a cryptographic process. Bitcoin: understanding bitcoin mining To obtain Bitcoins, people install and use powerful computers to perform calculations. This is called mining. Simply put, bitcoin mining is the process of creating a new bitcoin by solving a computer puzzle. It is necessary in order to keep the record of transactions on which the bitcoin is based. Bitcoin mining is performed by specialized computers. Those who do the mining are called miners. It is these miners who provide security and confirm Bitcoin transactions. Without Bitcoin miners, the network would be dysfunctional. In other words, Bitcoin mining is a distributed consensus system that is used to confirm current transactions by including them in the blockchain. Mining imposes a chronological order in the blockchain, protects network neutrality, and allows different computers to agree on the status of the system. To be confirmed, transactions must be written in a block that meets very strict cryptographic rules that will be verified by the network. These rules prevent previous blocks from being modified, as this would invalidate all subsequent blocks. Bitcoin: Decentralization This is probably the most important characteristic of Bitcoin and it seems necessary to explain what decentralization is. “Decentralization is the process of distribution and dispersion of power away from a central authority.” The advantages of the blockchain include: - A theoretically limited money supply, - Lower transaction costs. One of the ambitions of Blockchain technology when it first appeared in 2008 was to decentralize the international monetary system. The Blockchain technology records the actions of each individual in a distributed manner over a network and thus enables large-scale coordination without a central “authority” or supervisory body. In other words, the system operates without a central authority or a single administrator. It is managed in a decentralized manner thanks to the consensus of all the nodes in the network. However, most of the financial and government systems that currently exist are centralized, which means that only one higher authority is responsible for their management, such as a central bank or state apparatus. The Blockchain therefore allows decentralization, as it offers each user the opportunity to become one of the many audit processors in the network. Bitcoin: Where do I put my Bitcoins? In fact, we put our cryptocurrency like Bitcoin in special dedicated wallets; we call them digital wallets or “wallets”. It’s in these wallets that we store information about your crypto reserves translated into code. This is called a crypto wallet. A ‘wallet’ or wallet is therefore a secure digital storage process of cryptos. This ‘wallet’ is the equivalent of a bank account. It allows you to receive bitcoins and other tokens, store them, and then send them to other people. Private and Public Keys: A Bitcoin wallet contains a public key and a private key, which work together to allow its owner to initiate and digitally sign transactions, providing proof of authorization. To better understand how wallets work, we invite you to read our Top 10 Best Cryptocurrency Wallets for 2020. Bitcoin: Who accepts Bitcoin? Bitcoins as a store of value are very useful, but often, before you get them, you have to ask yourself the following question: Who accepts Bitcoin as a method of payment today? It can be seen that the number of sites and shops that accept Bitcoin and cryptocurrencies is constantly increasing. According to Coinmap.org, there are nearly 15,000 places worldwide where you can purchase goods and services using Bitcoin. An HSB survey also found that 36% of small and medium sized businesses in the U.S. accept Bitcoin. Through these locations, you can buy cars, apartments, etc. At Magna Numeris, we have reviewed and listed a few sites where you can spend your Bitcoin. Among the biggest players in e-commerce we can mention Amazon, we have selected several examples for you: ShopIfy: Shopify started accepting Bitcoin payments in November 2014, choosing at that time to offer BitPay as a payment option to its merchants. Shopify is one of the largest e-commerce platforms on the Internet. Microsoft: the company allows the use of Bitcoin to top up your Microsoft account. Microsoft accepts the use of Bitcoin in its online Xbox Store since 2014. AT&T: it is the first major US cell phone operator to offer its customers a payment option using cryptography such as Bitcoin. NewEgg: Newegg is a well-known company in the crypto-community for the purchase of crypto mining equipment. They sell all kinds of hardware at all ends of the spectrum. Bitcoin: Where to buy Bitcoin? To be able to spend it, we first need to own Bitcoin; then we can ask ourselves how to buy Bitcoin quickly and easily? There are services that will allow you to exchange your dollars or euros for Bitcoins or other crypto-currencies. Several platforms offer their services to enable the purchase and sale of cryptocurrencies Coinbase: Founded in 2012, Coinbase is one of the largest and most popular platforms for buying Bitcoins in the United States. It also operates in several European countries, including France. Its easy-to-use interface is perfectly suited for the most novice users and allows them to quickly make their first purchases. Binance: Founded in 2017, this platform allows you to buy Bitcoin. There are also nearly 200 other tokens, a very large choice of assets compared to other platforms. Localbitcoin: Founded in June 2012 by Jeremias Kangas, LocalBitcoin is a peer-to-peer bitcoin market based in Helsinki, Finland. Its service facilitates peer-to-peer trading in local currency against Bitcoins. eToro: Founded in 2007, this platform offers the purchase of dozens of cryptos including Bitcoin. It is one of the most widely used platforms in Europe with nearly 10 million subscribers. It is also the case of Cartam.world which wants to offer new innovative solutions to the crypto community by offering a free P2P (peer-to-peer) marketplace dedicated to crypto users, whether beginners or veterans. On cartam.world, you can use your crypto-currency with your own wallets to buy or sell second hand goods against cryptos and view other users’ offers through the map facilitating local meetings thanks to geolocation and also allowing more advanced searches thanks to advanced functions. Bitcoin: Who determines the price of Bitcoin? As we now know, Bitcoin is not issued by a central bank or supported by a government. Consequently, monetary policy, inflation rates and economic growth measures that generally influence the value of currencies such as the dollar or euro do not apply to Bitcoin. Therefore, there is no standard or global price for Bitcoin; it depends on supply and demand in each market, i.e., the number of people who want to buy and sell it, the price and the quantity. In general, the price of Bitcoin is influenced by: 1. The supply of bitcoin and the demand of the market 2. The cost of producing a bitcoin through the mining process 3. Rewards for Bitcoin Miners for verifying transactions in the blockchain Bitcoin ATH: All Time High Bitcoin is highly volatile in nature; Bitcoin has experienced significant price fluctuations, over a three-month period from October 2017 to January 2018, for example, Bitcoin’s price volatility has been close to 8%. Initially, the price of Bitcoin was low; it wasn’t until 2013 that Bitcoin began to take off. In October 2013, the price of Bitcoin was set at $123.50. It began to climb rapidly, reaching over $140 in April, and surpassed $1,000 in December of the same year. Bitcoin: Bitcoin Price History When the currency was first launched in 2009, it was at a value of $0.00. In 2010, the popularity of Bitcoin surged, and during the year Bitcoin’s value increased from $0.00 to $0.39 later in the year. By February 2011, Bitcoin had reached a price of $1.00, four months later it was worth approximately $31.00. What goes up must come down, however, and by the end of 2011, Bitcoin collapsed to a value of about $2. By the end of 2012, the price had gone from $4 at the start to about $13. - April 2013: $100 - November 8, 2013: $1,000 - May 20, 2017: $2,000 - June 11, 2017: $3,000 - August 13, 2017: $4,000 - September 1, 2017: $5,000 - November 28, 2017: $10,000 Bitcoin: The price at 23,000 Dollars in December 2020 ATH stands for “all-time high”, i.e. exactly the highest value of all times. This term refers to the highest price reached by an asset on a given market or stock exchange, and is very often used in the context of investors in cryptosystems. On December 17, the price of Bitcoin (BTC) reached $23,000 for the first time in history. This historic milestone comes just over two weeks after reaching its previous record. On Thursday, December 17, at approximately 09:14 UTC, the BTC reached an astonishing $23,441.60 USD, making it a record value. The price of bitcoin jumped more than 400% in 2020, after reaching a low of $3,600 in March. Bitcoin is a digital currency, also known as cryptocurrency, founded in 2008 by Satoshi Nakamoto, which uses blockchain technology to facilitate instant payments. Being stored in wallets as coinbase; Bitcoins allow its holder to buy goods and services on the Internet or in real life as to reserve to buy games on Microsoft’s Xbox or to buy computer equipment on Newegg. Has it helped you to understand how bitcoin works? Written by Laetisia Harson, Project Manager at Magna Numeris https://twitter.com/CartamOfficial https://www.linkedin.com/company/28594185 https://cartam.world/fr/ Cartam: Free Marketplace for Cryptocurrency Users Magnanumeris.com
https://medium.com/@magnanumeris/bitcoin-what-is-bitcoin-complete-guide-2021-c21f1ac30f3f
['Magna Numeris']
2020-12-23 09:28:09.906000+00:00
['Blockchain Technology', 'Bitcoin', 'Bitcoin Mining', 'Cryptocurrency']
1,913
4 Lessons We Can Learn From The Digital Revolution
When Steve Jobs was trying to lure John Sculley from Pepsi to Apple in 1982, he asked him, “Do you want to sell sugar water for the rest of your life, or do you want to come with me and change the world?” The ploy worked and Sculley became the first major CEO of a conventional company to join a hot Silicon Valley startup. It seems so quaint today, in the midst of a global pandemic, that a young entrepreneur selling what was essentially a glorified word processor thought he was changing the world. The truth is that the digital revolution, despite all the hype, has been something of a disappointment. Certainly it failed to usher in the “new economy” that many expected. Yet what is also becoming clear is that the shortcomings have less to do with the technology itself — in fact the Covid-19 crisis has shown just how amazingly useful digital technology can be — than with ourselves. We expected technology and markets to do all the work for us. Today, as we embark on a new era of innovation, we need to reflect on what we have learned. 1. We Live In A World Of Atoms, Not Bits In 1996, as the dotcom boom was heating up, the economist W. Brian Arthur published an article in Harvard Business Review that signaled a massive shift in how we view the economy. While traditionally markets are made up of firms that faced diminishing returns, Arthur explained that information-based businesses can enjoy increasing returns. More specifically, Arthur spelled out that if a business had high up-front costs, network effects and the ability to lock in customers it could enjoy increasing returns. That, in turn, would mean that information-based businesses would compete in winner-take-all markets, management would need to become less hierarchical and that investing heavily to win market share early could become a winning strategy. Arthur’s article was, in many ways, prescient and before long investors were committing enormous amounts of money to companies without real businesses in the hopes that just a few of these bets would hit it big. In 2011, Marc Andreesen predicted that software would eat the world. He was wrong. As the recent debacle at WeWork, as well as massive devaluations at firms like Uber, Lyft, and Peloton, shows that there is a limit to increasing returns for the simple reason that we live in a world of atoms, not bits. Even today, information and communication technologies make up only 6% of GDP in OECD countries. Obviously, most of our fate rests with the other 94%. The Covid-19 crisis bears this out. Sure, being able to binge watch on Netflix and attend meetings on Zoom is enormously helpful, but to solve the crisis we need a vaccine. To do that, digital technology isn’t enough. We need to combine it with synthetic biology to make a real world impact. 2. Businesses Do Not Self Regulate The case Steve Jobs made to John Sculley was predicated on the assumption that digital technology was fundamentally different from the sugar-water sellers of the world. The Silicon Valley ethos (or conceit as the case may be), was that while traditional businesses were motivated purely by greed, technology businesses answered to a higher calling. This was no accident. As Arthur pointed out in his 1996 article, while atom-based businesses thrived on predictability and control, knowledge-based businesses facing winner-take-all markets are constantly in search of the “next big thing.” So teams that could operate like mission-oriented “commando units” on a holy quest would have a competitive advantage. Companies like Google who vowed to not “be evil,” could attract exactly the type of technology “commandos” that Arthur described. They would, as Mark Zuckerberg has put it, “move fast and break things,” but would also be more likely to hit on that unpredictable piece of code that would lead to massively increasing returns. Unfortunately, as we have seen, businesses do not self-regulate. Knowledge-based businesses like Google and Facebook have proven to be every bit as greedy as their atom-based brethren. Privacy legislation, such as GDPR, is a good first step, but we will need far more than that, especially as we move into post-digital technologies that are far more powerful. Still, we’re not powerless. Consider the work of Stop Hate For Profit, a broad coalition that includes the Anti-Defamation League and the NAACP, which has led to an advertiser boycott of Facebook. We can demand that corporations behave how we want them to, not just what the market will bear. 3. As Our Technology Becomes More Powerful, Ethics Matter More Than Ever Over the past several years some of the sense of wonder and possibility surrounding digital technology gave way to no small amount of fear and loathing. Scandals like the one involving Facebook and Cambridge Analytica not only alerted us to how our privacy is being violated, but also to how our democracy has been put at risk. Yet privacy breaches are just the beginning of our problems. Consider artificial intelligence, which exposes us to a number of ethical challenges, ranging from inherent bias to life and death ethical dilemmas such as the trolley problem. It is imperative that we learn to create algorithms that are auditable, explainable and transparent. Or consider CRISPR, the gene editing technology, available for just a few hundred dollars, that vastly accelerates our ability to alter DNA. It has the potential to cure terrible diseases such as cancer and Multiple Sclerosis, but also raises troubling issues such as biohacking and designer babies. Worried about some hacker cooking up a harmful computer virus, what about a terrorist cooking up a real virus? That’s just the start. As quantum and neuromorphic computing become commercially available, most likely within a decade or so, our technology will become exponentially more powerful and the risks will increase accordingly. Clearly, we can no longer just “move fast and break things,” or we’re bound to break something important. 4. We Need A New Way To Evaluate Success By some measures, we’ve been doing fairly well over the past ten years. GDP has hovered around the historical growth rate of 2.3%. Job growth has been consistent and solid. The stock market has been strong, reflecting robust corporate profits. It has, in fact, been the longest US economic expansion on record. Yet those figures were masking some very troubling signs, even before the pandemic. Life expectancy in the US has been declining, largely due to drug overdoses, alcohol abuse and suicides. Consumer debt hit record highs in 2019 and bankruptcy rates were already rising. Food insecurity has been an epidemic on college campuses for years. So while top-line economic figures painted a rosy picture there was rising evidence that something troubling is afoot. The Business Roundtable partly acknowledged this fact with its statement discarding the notion that creating shareholder value is the sole purpose of a business. There are also a number of initiatives designed to replace GDP with broader measures. The truth is that our well-being can’t be reduced to and reduced to a few tidy metrics and we need more meaning in our lives than more likes on social media. Probably the most important thing that the digital revolution has to teach us is that technology should serve people and not the other way around. If we really want to change the world for the better, that’s what we need to keep in mind. Greg Satell is an international keynote speaker, adviser and bestselling author of Cascades: How to Create a Movement that Drives Transformational Change. His previous effort, Mapping Innovation, was selected as one of the best business books of 2017. You can learn more about Greg on his website, GregSatell.com and follow him on Twitter @DigitalTonto
https://greg-satell.medium.com/4-lessons-we-can-learn-from-the-digital-revolution-d483cb496368
['Greg Satell']
2020-10-24 12:38:42.980000+00:00
['Digital Transformation', 'Innovation', 'Technology']
1,914
World leaders congratulate Joe Biden and Kamala Harris
World leaders congratulate Joe Biden and Kamala Harris Presidents and Prime Ministers around the world went to Twitter to celebrate the next president of the US. The moment all major news outlets projected Joe Biden and Kamala Harris as the next President and Vice President of the United States, world leaders around the world flooded Twitter with messages of congratulations. Most G7 leaders sent their congratulations: French President Emmanuel Macron, British Prime Minister Boris Johnson, CanadianPM Justin Trudeau, Italy’s Prime Minister Giuseppe Conte and President Sergio Mattarella, German Chancellor Angela Merkel, as well as European Council President Charles Michel and European Commission President Ursula von der Leyen. Among other world leaders, India’s Narendra Modi tweeted not only to Joe Biden, but also a separate tweet to Kamala Harris, whose mother was originally from India. “Your success is pathbreaking, and a matter of immense pride not just for your chittis, but also for all Indian-Americans,” he tweeted. “I am confident that the vibrant India-US ties will get even stronger with your support and leadership.” Andrew Holness, Prime Minister of Jamaica, also sent special congratulations mentioning Harris’ Jamaican heritage — her father is originally from Jamaica. Among women world leaders, Jacinda Ardern of New Zealand, Erna Solberg of Norway, First Minister of Scotland Nicola Sturgeon, Kersti Kaljulaid of Estonia, Simonetta Sommaruga of Switzerland, Zuzana Čaputová of Slovakia, Mia Amor Mottley of Barbados, Sanna Marin of Finland, Katerina Sakellaropoulou of Greece, Denmark Danish PM Mette Frederiksen, and Christine Lagarde of the European Central Bank all went to Twitter to congratulate Joe Biden and, in particular Kamala Harris, the first woman to become Vice President of the United States. Twitter has created a list of world leaders congratulated the next US President and Vice President: https://twitter.com/i/events/1325143166773653505 Young global leader like Malala also congratulated Kamala Harris on Twitter; while Greta Thunberg retweeted CNN about Biden’s win.
https://medium.com/digital-diplomacy/world-leaders-congratulate-joe-biden-and-kamala-harris-f9f94d84a577
['Andreas Sandre']
2020-11-07 23:14:30.031000+00:00
['Tech', 'Social Media', 'Elections', 'Technology', 'Government']
1,915
My New Apple Watch is a Privacy Nightmare
Image courtesy of Techcrunch.com The Apple Watch comes equipped with some labor-saving apps that are designed to make life easier when using it with your paired iPhone. I can use my wrist to answer phone calls like a secret agent, I can take notes, talk to Siri, and even ping my phone when I’ve forgotten where I put it. For those of us who remember the 1990 movie with Warren Beatty and Madonna, or even the classic comic strip, there’s even a walkie-talkie app so we can live out our Dick Tracy dreams. It also comes equipped with two apps that are going to cause a lot of people headaches; a remote shutter release, and a way to activate the voice recorder remotely. I work in an industry where we are exposed to reams of confidential information, much of which we can’t even share with other parts of my organization. This combination of the Apple Watch and iPhone make surveillance and espionage not only unobtrusive, but also as easy as checking the time. These are the result of me “checking my watch” in a public parking lot Taken from my iPhone in my belt holster using the Apple Watch app. (Author) From the same position, this time using the zoom function. (Author) As far as anyone could tell, I was checking my watch…maybe answering a message, or reading a notification. There was no indication that I was taking a photograph, not even a shutter release noise. That school in the background? It’s actually almost an entire city block away, across a football field-sized grassy area. (Author) When it’s combined with the iPhone 11 Max Pro’s impressive zoom features, you can get clear pictures without ever moving your phone off your hip, or out of your pocket. When it’s combined with the ability to start a sound recording from my wrist, I could conceivably record an entire walk-through of a facility to be reviewed at a later point. Of course, with the ability to slip the bands off, it becomes really easy to hide the watch face in a pocket or even in a hand and take photos that way. What makes it a voyeur’s dream? The phone (and camera) could be dozens of feet away, and the remote shutter release still works, makes sense right? Except that if the phone is set on silent, it doesn’t make any noise. Still not getting it? Voyeur + Apple Watch + iPhone = This is our (empty) bathroom at home…taken while I was on the other side of the house. (Author) Consider the thought of a silent phone taking photos after being placed in the ceiling of a dressing room or bathroom. It gets worse I forgot to mention, the Apple Watch App provides a live view of what the camera is seeing, even if you don’t take a picture. Anyone with a watch connected to a phone can monitor the room where the phone is while they’re dozens of feet away. So there’s no evidence to prove what happened if the phone itself isn’t found in an incriminating place.
https://medium.com/swlh/my-new-apple-watch-is-a-privacy-nightmare-fcf6c84662c5
['Matthew Woodall']
2019-12-05 15:39:51.054000+00:00
['Privacy', 'Technology', 'iPhone', 'Apple', 'Apple Watch']
1,916
XinFin and CSI-SPIT presents Hands on Blockchain Workshop.
XinFin and CSI-SPIT presents Hands on Blockchain Workshop. Team XinFin with Team CSI-SPIT college presented workshop on Blockchain Technology at SPIT college Mumbai, India on Dec 07, 2018. XinFin is building a strong developer ecosystem to involve developers globally by organising workshops locally and globally. XinFin is engaged in integration of blockchain with other technologies and encouraging blockchain enthusiasts from various parts of the world by conducting workshops and hackathons. XinFin also welcomes new ideas from experts who are unable to attend the workshop and hackathon, by organising Open Hackathon. Open Hackathon will innovate and replace existing non-efficient closed systems with more transparent, efficient system using blockchain, for details Click Here.. Team XinFin with Team (Computer Society of India) CSI — SPIT student branch. At your left, Dr. Sudhir Dhage, Professor, Department of Computer Engineering, Faculty Coordinator CSI-SPIT Students Branch, Regional Student Coordinator (CSI Region VI: Maharashtra and Goa). Rudresh Solanki, XinFin community member in middle and Nishant Kumar, Senior Blockchain Engineer at right. XinFin and CSI-SPIT [ Sardar Patel Institute of Technology] college jointly organised a free blockchain workshop held at Sardar Patel Institute of Technology, Bhavan‘s Campus, Munshi Nagar, Andheri, Mumbai, India on Dec 7, 2018. Workshop started at 10 am [IST], it started with Dr. Sudhir Dhage, Professor, Department of Computer Engineering, Faculty Coordinator CSI-SPIT Students Branch, Regional Student Coordinator (CSI Region VI: Maharashtra and Goa) felicitating the instructors from XinFin, Rudresh Solanki & Nishant Kumar. Students and Start ups from various colleges came to attend the workshop from different parts of the city, Mumbai. Due to large number of attendees, the workshop had to be split into two sessions held concurrently in different computer labs. Blockchain enthusiast from various colleges The workshop portion was split into two parts before & after lunch. First part of the workshop was an introduction to blockchain followed by students installing and deploying their very blockchain network based on XDC nodes. Major terms like “mining”, “gas”, “transactions”, etc were explained with example by students actually performing them on their laptops. Rudresh Solanki, XinFin community member explaining at the workshop Nishant Kumar, Senior Blockchain Engineer explaining at the workshop The second part of the workshop was focused on application building on blockchain i.e. decentralized applications. Second session started with learning of basic syntax of coding language Solidity. Students coded out & deployed two smart contracts one based on a real world auction & other on an ERC20 token. For deployment Remix was used. Other deployment methods like via Truffle were discussed & shown. Nishant Kumar, Senior Blockchain Engineer explaining concepts at the workshop Rudresh Solanki, XinFin community member helping students at the workshop Next the web3.js library was discussed with a node based application to connect to XDC blockchain network setup in the previous part of the workshop. All the attendees were given, Participation certificates by the instructors already signed by Ankit Patel ( Training Academy XinFin ). We do not limit to this workshop. XinFin is now hosting an Open Hackathon, where-in we invite all the developers to find solutions to the problem statement mentioned on our GitHub, and can write code to receive prizes in XDC tokens worth $10000. Here(https://github.com/XinFinOrg/Hackathon) is the link to repository of Open Hackathon. We invite all of you to join the XinFin Community at: Open Hackathon: https://github.com/XinFinOrg/Hackathon GitHub: https://github.com/XinFinorg/ Telegram: https://t.me/xinfin Twitter: https://twitter.com/XinFin_Official Subscribe to the XinFin Community Newsletter by clicking here. You will be receiving our newsletter direct to your inbox every month. XinFin Hybrid Blockchain [XDCE] www.xinfin.org
https://medium.com/xinfin/xinfin-and-spit-presents-hands-on-blockchain-workshop-a25695e61640
['Xinfin Xdc Hybrid Blockchain Network']
2019-01-21 10:30:51.127000+00:00
['Latest News', 'Technology', 'Hackathons', 'Blockchain', 'Github']
1,917
Hosting A Modded Minecraft 1.16.4 Server on a Raspberry Pi
Hosting A Modded Minecraft 1.16.4 Server on a Raspberry Pi Recently, some friends and I wanted to check out the recent major updates made to Minecraft. There’s something especially inviting about diving into a new world, exploring the infinite landscape and building huge castles with friends that has captured Minecraft players for years. Here, I’m going to describe the process of setting up the latest version of Minecraft Forge and installing your favorite mods, because who doesn’t want to add Tesla coils and rail guns to their game?! But Minecraft is better with friends! In Part Two, we’ll be hosting our own modded server on a Raspberry Pi. Make sure to follow the steps here first. First thing’s first — you’ll need to be an owner of Minecraft Java Edition. As of this article’s publication, the most recent version is 1.16.4. You can buy a copy of Minecraft here. Even if you’re on Windows 10, you want the Java Edition! Since we’ll be playing modded Minecraft, you’ll also need a copy of Minecraft Forge 1.16.4. Make sure to download the one that says ‘installer’ under the ‘Download Recommended’ banner. If you’re going to be following along in Part Two, we’ll use the same file to install the server on our Raspberry Pi. Select Some Mods There are a ton of community made mods for Forge. You can browse and select which ones you want here. My favorite is Immersive Engineering, which adds a robust electricity and mechanics system. Just make sure the mod is compatible with your version of Minecraft, then click the download button to get the zip file. For now, download the mods you want and keep the zip files someplace you can find them later. The game version should match your version of Minecraft Java Edition, which should also match your version of Forge. A note about mods — Some mods, like the very popular Optifine, are client side only. That means we won’t need to put them on the server later and players are free to use them as they please. All other mods must be installed on the server and the client, so all players must have a matching mod list! If a mod doesn’t specifically say it’s client side only, assume it’s for both. If you’re not going to be building the server, don’t worry too much about this. If you are, make sure you add the same exact mods later. Launching Forge from the Minecraft Launcher When you launch Minecraft after first installing it, you should be treated to something like this. Notice the version just to the left of the play button. We’re going to install Forge so we can select it from our launcher instead. Installing Java Stay with me here, this is likely the worst part. In order to use Forge, your computer needs to have Java installed. Java used to be installed alongside Minecraft, but lately the game has used a bundled version to avoid needing that (this is a good thing). Forge, however, still needs Java running on your computer. For MacOS — Open your terminal with the launchpad or by going to Applications > Utilities and clicking on Terminal. Check to see if you have Java installed already by entering this into your terminal. java --version If you see something like this and as long as the number after Java is 8 or above, you’re good to go! java 13-ea 2019-09-17 Java(TM) SE Runtime Environment (build 13-ea+33) Java HotSpot(TM) 64-Bit Server VM (build 13-ea+33, mixed mode, sharing) Otherwise, we need to run a couple of commands here to install Java. We’ll use Homebrew, an excellent package manager for MacOS. Run this first to install, (or update if you have it already), Homebrew. Then run this to install Java. brew cask install java FYI, there are two ways to install and use Java — the officially supported version from Oracle that requires a license and the open source version called OpenJDK. Since we’re not Java programmers who need production level support, we’re happy with OpenJDK. That’s what Homebrew will install. For Windows 10 — This process is a bit manual. You’ll need to download a version of OpenJDK from here. The latest version should be fine. Extract the zip file in a folder at C:\Program Files\Java\ (you’ll need to create the Java folder). Open the control panel, then select system > advanced > environment variables. You’ll want to append the location of the ‘bin’ folder from what you just extracted in C:\Program Files\Java\ to the PATH variable. After doing so, it should look something like this, but with whatever version you have — C:\WINDOWS\system32;C:\WINDOWS;"C:\Program Files\Java\jdk-11\bin" Finally, under system variables, click new and add a variable called JAVA_HOME. Set this to be the same Java location, but this time without the bin folder. Click ok, then apply changes. C:\Program Files\Java\jdk-11 Installing Forge If you didn’t download Forge before, go ahead and do so here. Click on the installer option under the recommended banner. This should download a .jar file. Open the file to be walked through installing Forge for the client. Once it’s finished, relaunch the Minecraft launcher. You should now have an option launch Minecraft with Forge. Installing Mods On MacOS — Go back to the terminal and paste this — open ~/Library/Application\ Support/minecraft If there isn’t a folder called mods, create one. Simply drop the .jar files of the mods you downloaded into this folder. On Windows 10 — Click start and type ‘run’ into the search bar. Type %appdata% into the text field of run and hit the enter key. It should bring up a location in the file explorer. You should have a folder called .minecraft in this location. If there isn’t a folder called mods in the .minecraft folder, create one. Drop the .jar files of the mods you want here. At this point, you should be able to play single player Minecraft with mods! Make sure to check out Part Two so your friends can join!
https://medium.com/@curtmorgan3/hosting-a-modded-minecraft-1-16-4-server-on-a-raspberry-pi-9dc5952b29c0
['Curt Morgan']
2020-12-01 18:47:43.521000+00:00
['Raspberry Pi', 'Technology', 'Minecraft', 'Programming', 'Raspberry Pi 4']
1,918
TechNY Daily
TechNY Daily Keeping the NYC Tech Industry Informed and Connected 1. NYC’s Bizzabo, an online platform for event organizers, has raised $138 million in a Series E funding. Insight Partners led the round and was joined by investors including Viola Growth, Next47 and OurCrowd. Bizzabo’s platform can be used to plan and run both virtual and in-person conferences. (www.bizzabo.com) (TechCrunch) 2. Tom Brady and John Mayer have invested in NYC’s Hodinkee, a content and e-commerce site for high end watches. Other investors in the $40 million Series B round included TCG, GV (Google Ventures), LVMH Luxury Ventures, True Ventures and Future Shape. (www.hodinkee.com) (Forbes) 3. NYC co-living startup Ollie has been acquired by a competitor, San Francisco’s Starcity. The purchase price for the acquisition of Ollie’s assets and roommate-matching technology was not disclosed. Ollie, which had raised a total of $15 million from investors including the Moinian Group, and the Texas Employees Retirement System, operates co-living spaces in NYC, Los Angeles and Boston. (www.ollie.co) (The Real Deal) 4. Brooklyn’s Actasys, a startup that applies aerodynamic principles to solve problems in the automotive industries, has raised $5 million in a seed funding.The round was led by Volvo Cars Tech Fund and NextGear Ventures. The company’s ActaJet product uses air to clean vehicle sensors. (www.actasysinc.com) (BusinessWire) _________________________________________________________________ Small Planet partners with the world’s most innovative companies to create inspired apps and websites. Development, UX design, monetization strategies, user acquisition, and more. Contact us. (Sponsored Content) _______________________________________________________________ 5. NYC’s Current, a digital “challenger” bank, has raised $131 million in a Series C funding. NYC’s Tiger Global Management led the round and was joined by Sapphire Ventures, Avenir, Foundation Capital, Wellington Management Company and QED Investors. Current, which started as a teen debit card controlled by parents, now is offering digital banking services to more than two million members. (www.current.com) (TechCrunch) 6 NYC’s Materialize, an SQL database platform for streaming data, has raised $32 million in a Series B funding. Venerable Silicon Valley VC firm Kleiner Perkins led the round with participation by Lightspeed Ventures. The company’s platform implements an SQL interface for streaming data. (www.materialize.com) (PR Newswire) 7. NYC’s Funnel (formerly, Nestio), a platform that helps landlords manage rental portfolios, has raised $14 million in a Series AA funding. RET Ventures led the round and was joined by Trinity Ventures and Camber Creek. Funnel’splatform is an inventory management tool for landlords which includes an AI-powered bot that responds to potential renters in real time. (www.funnelleasing.com) (The Real Deal) 8. NYC’s Hellosaurus, an e-learning video platform for kids, has raised $3.5 million in a seed funding. General Catalyst led the round with participation by GSV Ventures, Shrug Capital, Next 10 Ventures, BDMI, Runway Fund, GFC, as well as Dave Gilboa (Founder, Warby Parker), Neil Blumenthal (Founder, Warby Parker), Jeff Raider (Founder, Warby Parker, Harry’s), and Joey Zwillinger (Founder, Allbirds) through their Good Friends fund. The company’s app is a video platform for kids filled with episodes they can interactive with instead of just watch. (www.hellosaurus.com) (GlobeNewswire) We have special sale pricing on TechNY Daily sponsorship and advertising opportunities. For information, contact: Lauren@techny.nyc NYC Tech Industry Virtual Event Calendar This Afternoon (3pm) Getting Ready for 2021 Legal and Finance Panel includes TechNY’s Paul Goodman Hosted by Daily Stack Free with promo code: TECHNY (** calendar continued below **) ____________________________________________ TechNY Recruit Jobs Job Postings are on sale. Contact us at jobpostings@techny.nyc Circle (new) Circle was founded on the belief that blockchains and digital currency will rewire the global economic system, creating a fundamentally more open, inclusive, efficient and integrated world economy. Solutions Engineer Director, Account Management Enterprise Sales Director (Banking and Financial Services) Senior Software Engineer, Frontend Manager, Software Engineering Agilis Chemicals Transforming chemical industry with modern commerce technology Full-stack Engineer Business Development Manager — Enterprise SaaS Marketing Director — Enterprise SaaS LiveAuctioneers Awarded for four consecutive years as one of Crain’s 100 Best Places to Work in NYC, LiveAuctioneers is the largest online marketplace for one-of-a-kind items, rare collectibles, and coveted goods. Product Marketing Manager Senior Marketing Manager Lukka We are a SaaS solution that makes crypto accounting easy. We are a trusted, blockchain-native technology team that is passionate about digital asset technology. Our team is continuously collaborating and designing new products and initiatives to expand our market presence. Technology and customers are at the center of our universe. We dedicate our energy to learning, building, adapting, and achieving impactful results. Customer Success Specialist Third Party Risk Manager or Director Sales Proposal Manager Summer Summer’s mission is to help the 45 million Americans burdened by student debt save time and money through smart, algorithm-based recommendations. Summer combines policy expertise and innovative technology to serve student loan borrowers across the country. Back-End Engineer Logikcull.com Our mission: To democratize Discovery. Enterprise Account Executive The Dipp A personalized subscription site for pop culture’s biggest fans. Director of Engineering Ridgeline Founded by Dave Duffield (founder and former CEO of Workday and PeopleSoft) in 2017, Ridgeline’s goal is to revolutionize technology for the investment management industry. We are building an end-to-end cloud platform on a modern infrastructure using AWS, serverless technologies, which up to this point hasn’t been done in the enterprise space. Software Engineering Manager, UI Simon Data Simon Data is the only enterprise customer data platform with a fully-integrated marketing cloud. Our platform empowers businesses to leverage enterprise-scale big data and machine learning to power customer communications in any channel. Senior Front-End Engineer Product Designer Director, Enterprise Sales Full Stack Engineer Vestwell Retirement made easy. Senior Fullstack Engineer Package Free Package Free is on a mission to make the world less trashy though offering products that help you reduce waste daily. We source our products from individuals and brands with missions to create a positive environmental impact and since launching, have diverted over 100 million pieces of trash from landfills. Head of Operations Hyperscience Hyperscience is the automation company that enables data to flow within and between the world’s leading firms in financial services, insurance, healthcare and government markets. Founded in 2014 and headquartered in New York City with offices in Sofia, Bulgaria and London, UK, we’ve raised more than $50 million raised to date and are growing quickly. We welcome anyone who believes in big ideas and demonstrates a willingness to learn, and we’re looking for exceptional talent to join our team and make a difference in our organization and for our customers. Machine Learning Engineer Senior Security Engineer Braavo Braavo provides on demand funding for mobile apps and games. We offer a flexible and affordable funding alternative to the traditional sources of capital like working with a VC or bank. We’re changing the way mobile entrepreneurs finance and grow their app businesses. Our predictive technology delivers on demand, performance-based funding, without dilution or personal guarantees. By providing non-dilutive, yet scalable alternatives to equity, we’re helping founders retain control of their companies. Business Development Manager VP of Marketing Head of Sales Yogi At Yogi, we help companies decipher customer feedback, from ratings and reviews to surveys and support requests. Companies are inundated with feedback, but when it comes to turning this data into actionable business decisions, most companies fall short. That’s where Yogi fits in. Full Stack Software Engineer Ordergroove We’re passionate marketers, engineers and innovators building the technology to power the future of commerce. We’re a B2B2 SaaS platform helping the world’s most interesting retailers and direct-to-consumer brands remove friction from the customer experience to deliver recurring revenue through subscriptions programs — shifting their consumer interactions from one-and-done transactions to long-lived, highly profitable relationships. Data Scientist Upper90 Upper90 is an alternative credit manager based in New York City that has deployed over $500m within 18 months of inception. Investor Relations Analyst Upscored UpScored is the only career site that uses data science to connect you with jobs suited specifically to you while automatically learning your career interests. Its AI-powered platform decreases job search time by 90%, showing you the jobs you’re most likely to get (and want) in less than 2 minutes. Data Engineer Senior Frontend Developer Senior Backend Developer Frame.io Frame.io is a video review and collaboration platform designed to unify media assets and creative conversations in a user-friendly environment. Headquartered in New York City, Frame.io was developed by filmmakers, VFX artists and post production executives. Today, we support nearly 1 million media professionals at enterprises including Netflix, Buzzfeed, Turner, NASA & Vice Media. Frontend Engineering Manager Sr. Swift Engineer Lead Product Designer Attentive Attentive is a personalized text messaging platform built for innovative e-commerce and retail brands. We raised a $230M Series D in September 2020 and are backed by Sequoia, Bain Capital Ventures, Coatue, and other top investors. Attentive was named #8 on LinkedIn’s 2020 Top Startups list, and has been selected by Forbes as one of America’s Best Startup Employers. Enterprise Account Executive Sales Development Representative Senior Client Strategy Manager Director of Client Strategy KeyMe NYC startup revolutionizing the locksmith industry with innovative robotics and mobile technology. Customer Experience Representative Inbound Phone Sales Representative Systems Software Engineer Button Button’s mission is to build a better way to do business in mobile. Enterprise Sales Director — New York Postlight Postlight is building an extraordinary team that loves to make great digital products — come join us! Full Stack Engineer Deliver your job listings directly to 48,000 members of the NYC tech community at an amazingly low cost. Find out how: jobpostings@techny.nyc ____________ NYC Tech Industry Virtual Event Calendar This Afternoon (3pm) Getting Ready for 2021 Legal and Finance Panel includes TechNY’s Paul Goodman Hosted by Daily Stack Free with promo code: TECHNY December 16 Fundraising 201: How to Raise a Seed Round Efficiently Hosted by Startup Grind December 16 Rise Refresh: Remote Selling with Winning by Design Hosted by Rise New York Contact Us for Free Listing of Your Web-based Events Send us your events to list (it’s Free!) to: editor@techny.nyc Did You Miss Anything Important? Read Our TechNY Daily Past Editions TechNY Daily is distributed three times a week to 48,000 members of NYC’s tech and digital media industry. Connecting the New York Tech Industry Social Media • Mobile • Digital Media • Big Data • AdTech • App Development • e-Commerce • Games • Analytics • FinTech • Web • Software • UX • Video • Digital Advertising • Content • SaaS • Open Source • Cloud Computing • AI • Web Design • Business Intelligence • Enterprise Software • EduTech • FashionTech • Incubators • Accelerators • Co-Working • TravelTech • Real Estate Tech Forward the TechNY Daily to a friend Not a Subscriber to TechNY Daily, Click Here Copyright © 2020 TechNY, All rights reserved.
https://medium.com/@smallplanetapps/techny-daily-3cf9054c13a0
['Small Planet']
2020-12-09 19:21:57.308000+00:00
['Technology News', 'Venture Capital', 'Startup', 'Funding', 'Technews']
1,919
No more exploitation — A revolution in money transfer transactions for international migrants
The remittance industry is gaining its position in all nations’ economy due to the current trend of globalization. This vast ecosystem consists of various stakeholders, among whom are approximately 250 million international migrants worldwide. It is projected that remittances sent back home would then add up $444 billion in 2017, according to World Bank. This cross-border transfer is economically significant for not only the receiving families to maintain basic living conditions but also GDP of many developing or underdeveloped countries as one of the largest capital inflows. Monopoly in the non-commercial market Despite the potential of the money transfer industry, there are currently only a few players that dominate the market and control the infrastructure that enables money transfers across borders (i.e. Western Union/MoneyGram). Considering the biggest player alone, Western Union, Bloomberg reported the company had a vast 13% market share with $8.95 billion market capital in 2017. Alongside with MoneyGram, they dominate the current market and this monopoly allows them to dictate terms in the market. Unnecessarily high transaction costs result in greater complication An obvious example is the punitively high transactions costs varying substantially among different corridors. For instance, sending $200 from the US to El Salvador costs $8, but it’s $42 to send $500 to Burkina Faso through the Western Union website. This reflects not only the excessive cost but also financial exclusion of the system. As explained, these costs are due to consumer protection costs, local remittance taxes, market distribution, regulatory structure, volume, currency volatility and other market efficiencies. However, logistical and regulatory costs alone cannot account for such extortionate fees and there are great technological alternatives to lower their costs, which these companies are reluctant to apply. Having no better choice to make, migrants have to compromise with these costs, resulting in at least $32 billion lost in transaction fees. The only other way to lower this cost is through unofficial money operators with informal contracts and no money arrival guarantee, which will put you at high risk of losing your money. The world’s official volume of remittance will also lack up to $100 billion only because of this issue. Restrictive business environment does not respond to market need Another term they create is the restrictive business practices the incumbents have set in place. For instance, banks and agents have exclusive agreements which stipulate that once they work with each other, they are not allowed to work with other providers. This is an impassable barrier that restricts newcomers from entering the market as they traditionally need banks to enable cross-border payments. This practice enforces migrants to have a bank account but signing up for one takes more time and more hectic procedures than the indigenous, thus making them reluctant to open an account and become the “unbanked”. With the current heavy reliance on the centralized banking system, it would cause lots of inconveniences to transfer money abroad. This issue gives our solution more validity as we have designed a money transfer system that completely removes the reliance on banks by using cryptocurrencies to increase financial inclusion worldwide. It’s time to think about another convenient way to transfer money abroad! A technological solution is on the way! ZeroBank employs the currently flourishing technologies — blockchain and smart contract — along with the proven sharing economy model to tackle the aforementioned problems in the market. Via this platform, money can be sent, received, and exchanged quickly, directly, at any time, anywhere in the world through a network of community agents. End-users including immigrants can bear minimal costs since ZeroBank platform completely removes the interference of the centralized intermediaries like banks, or MTOs. Moreover, the user registration is consumer-oriented, which not only helps customers but also our system to verify their identities. Once verified through our KYC, with a smartphone connected to the Internet, they can make a transaction without any knowledge about cryptocurrencies. ZeroBank is currently putting every effort in completing MVP, which is expected to be released on 15th August for trial as well as testing and finalizing product to be launched in three markets including France, United Kingdom and Vietnam next year. The World Bank recently said that lower transaction fees could spur an increase in remittances and lift another 30 million people out of poverty. With this project, we hope to contribute to achieving this goal in the near future. Stay updated on our channels:
https://medium.com/zerobank-cash/no-more-exploitation-a-revolution-in-money-transfer-transactions-for-international-migrants-be8d4c5fb8a2
['Zerobank - Your Local Currency']
2018-08-08 02:37:04.579000+00:00
['Remittances', 'Technology Solution', 'Migrants']
1,920
Why is Blockchain NOT the challenge but the SOLUTION to sustainable energy
1. Decentralized energy supply system and household participation One of the most special opportunities that Blockchain brings to the energy sector is that it encourages the growth of electricity ‘prosumers’. Prosumers refers to households that are able to generate electricity or become the supplier of electricity, for example, by having a solar panel installed in their buildings. By putting consumers at the heart of the system, both renewable energy consumption and production can be enhanced, creating a positive feedback loop that accelerates the transformation of society from non-renewable to renewable energy in a self-sustaining way. In order to bring such a system to reality, there need to be an incentive mechanism, for instance, prosumers should expect to be rewarded with positive return from investing in installing a solar panel. However, in the current electricity supply system, the challenge is not only that there isn’t a trusted platform to facilitate household electricity transactions, but also the fact that multiple intermediaries and complexity in regulatory compliance cuts the potential income for households from engaging in electricity trades. Many governors and scholars in this field have recognized the need for changes, in 2017 the European Union’s Clean Energy Package was implemented, with a set of rules on household energy producing, storage and trading, aiming to encourage consumers to actively participate in electricity trading. Academic literature has extensively focused on improving the market design by establishing local storage systems into national operations of power systems, nevertheless, progress is slow and we haven’t seen significant changes in the past few years. This is where a decentralized Blockchain system could come into play. Firstly and most importantly, Blockchain Technology could create a trustless network to encourage Peer to Peer (P2P) electricity transactions. In any P2P transaction, trust between parties are extremely important. If we think about how Uber, Airbnb have built up their P2P exchange system, parties involved in the transaction have trust in the paring and verification mechanisms of the platform, for example, the registration process to become a Uber driver. Blockchain technology provides an even stronger ‘trustless’ system, where you don’t have to trust the counterparty, but only the technology. By offering cryptographic ways of tracking transactions, the transparency in terms of transaction records and anonymity in terms of personal information are both enhanced. Without a centralized control, identical databases that records each and every transaction ( in an encrypted form) are stored in the device of every player involved, and in order for a transaction to be recorded, all parties need to reach consensus and validate the transaction. image source: unsplash “Historically, although there are some exceptions, people tended not to share with strangers or those outside their social networks. Sharing was confined to trusted individuals such as family, friends and neighbours. Today’s sharing platforms facilitate sharing among people who do not know each other, and who lack friends or connections in common.” (Frenken and Schor, 2017). Furthermore, through the use of a distributed ledger system, many intermediaries are removed from the transaction, a pure peer to peer transaction is possible to achieve. This reduces cost and time of transaction, increasing the “profit margin” and incentive for households to engage in energy trading activities. We have already seen practical implementations in many countries around the globe, both startups and traditional large energy suppliers are actively pushing for such a revolution. For instance, in 2018, Centrica- the owner of British Gas worked with a American startup to initiate an energy sharing project for their Local Energy Market. The project aims to enable trading of energy between local businesses, consumers, the national grid and other participants in the UK. Another well-known case is the Australian startup “Power Ledger”. It uses a blockchain-based network to establish a peer to peer energy trading platform. Some of its key benefits include real-time payment, automated low cost settlement, neighborhoods trading, transparent trading information and much more. One of the most interesting features of Power Ledger is that it comprises of both a permissioned and a permissionless blockchain infrastructure. The permissionless blockchain operates on a global scale to allow international market trading, whereas the permissioned blockchain operates among local trusted peers. The platform also uses two tokens as assets: one called “POWR” — which is used to access the global P2P trading market, while the other one — “Sparkz” usually functions in the local network as the trading currency, representing the electricity price and the local “real-world” currency. In combination, the system ensures value via “POWR” tokens and facilitates exchange through “Sparkz” . Other P2P electricity trading startups and projects you could explore include: Apart from solar panels, another active field is electric vehicle charging. Sales of electricity vehicles is increasing dramatically year on year. Overall sales of Tesla’s Electric vehicle is expected to surpass 10 million by 2025, and responding to such a trend, the demand for charging stations will scale up exponentially. This puts pressure on the national electricity grid as millions of drivers demand for energy flows to charge their cars at the same time. The idea of ‘distributed’ charging stations matches well with blockchain’s characteristics, it could play a key role in improving the coordination of a distributed charging network by allowing any owner of EV charger to be a supplier. The use of blockchain smart contract enables trusted payment and transaction to be made between individual EV owners and EV charger providers who are unknown to each other. image source: unsplash The Share & Charge project initiated by Slock.it is an example of P2P EV charging station project, allowing households owning charging stations to provide charging services to drivers. It could benefit the ‘suppliers’ by recouping the cost of their investment, while broadening the geographical coverage of EV charging stations to benefit the drivers. The app has been available since the 28th of April for charging stations located in Germany.
https://medium.com/datadriveninvestor/why-is-blockchain-not-the-challenge-but-the-solution-to-sustainable-energy-e3265a42a61a
['Americana Chen']
2020-12-28 12:47:21.101000+00:00
['Energy', 'Blockchain', 'Technology', 'Sustainability', 'Renewable Energy']
1,921
Harvard Professor And His Students Raise $14 Million To Make AI Too Smart To Be Fooled By Hackers
Harvard Professor And His Students Raise $14 Million To Make AI Too Smart To Be Fooled By Hackers Digital Times Africa Nov 2, 2020·4 min read Yaron Singer climbed the tenure track ladder to a full professorship at Harvard in seven years, fueled by his work on adversarial machine learning, a way to fool artificial intelligence models using misleading data. Now, Singer’s startup, Robust Intelligence, which he formed with a former PhD advisee and two former students, is emerging from stealth to take his research to market. This year, artificial intelligence is set to account for $50 billion in corporate spending, though companies are still figuring out how to implement the technology into their business processes. Companies are still figuring out, too, how to protect their good AI from bad AI, like an algorithmically generated voice deepfake that can spoof voice authentication systems. “In the early days of the internet, it was designed like everybody’s a good actor. Then people started to build firewalls because they discovered that not everybody was,” says Bill Coughran, former senior vice president of engineering at Google. “We’re seeing signs of the same thing happening with these machine learning systems. Where there’s money, bad actors tend to come in.” Enter Robust Intelligence, a new startup led by CEO Singer with a platform that the company says is trained to detect more than 100 types of adversarial attacks. Though its founders and most of the team hold a Cambridge pedigree, the startup has established headquarters in San Francisco and announced Wednesday that it had raised $14 million in a seed and Series A round led by Sequoia. Coughran, now a partner at the venture firm, is the lead investor on the fundraise, which also comes with participation from Engineering Capital and Harpoon Ventures. Singer followed his PhD in computer science from the University of California at Berkeley, by joining Google as a postdoctoral researcher in 2011. He spent two years working on algorithms and machine-learning models to make the tech giant’s products run faster and saw how easily AI could go off the rails with bad data. “Once you start seeing these vulnerabilities, it gets scary, especially if we think about how much we want to use artificial intelligence to automate our decisions,” he says. Fraudsters and other bad actors can exploit the relative inflexibility of artificial intelligence models in processing unfamiliar data. For example, Singer says, a check for $401 can be manipulated by adding a few pixels that are imperceptible to the human eye yet cause the AI model to read the check erroneously as $701. “If fraudsters get their hands-on checks, they can hack into these apps and start doing this at scale,” Singer says. Similar modifications to data inputs can lead to fraudulent financial transactions, as well as spoofed voice or facial recognition. In 2013, upon taking an assistant professor position at Harvard, Singer decided to focus his research on devising mechanisms to secure AI models. Robust Intelligence comes from nearly a decade in the lab for Singer, during which time he worked with three Harvard pupils who would become his cofounders: Eric Balkanski, a PhD student advised by Singer; Alexander Rilee, a graduate student; and undergraduate Kojin Oshiba, who co-authored academic papers with the professor. Across 25 papers, Singer’s team broke ground on designing algorithms to detect misleading or fraudulent data and helped bring the issue to government attention, even receiving an early Darpa grant to conduct its research. Rilee and Oshiba remain involved with the day-to-day activities at Robust, the former on government and go-to-market, and the latter on security, technology and product development. Robust Intelligence is launching with two products, an AI firewall and a “red team” offering, in which Robust functions like an adversarial attacker. The firewall works by wrapping around an organization’s existing AI model to scan for contaminated data via Robust’s algorithms. The other product, called Rime (or “Robust Intelligence Machine Engine”), performs a stress test on a customer’s AI model by inputting basic mistakes and deliberately launching adversarial attacks on the model to see how it holds up. The startup is currently working with about ten customers, says Singer, including a major financial institution and a leading payment processor, though Robust will not name any names due to confidentiality. Launching out of stealth, Singer hopes to gain more customers as well as double the size of the team, which currently stands at 15 employees. Singer, who is on leave from Harvard, is sheepish about his future in academia but says he is focused on his CEO role in San Francisco at the moment. “For me, I’ve climbed the mountain of tenure at Harvard, but now I think we’ve found an even higher mountain, and that mountain is securing artificial intelligence,” he says. Source: Forbes
https://medium.com/@digitaltimes-2020/harvard-professor-and-his-students-raise-14-million-to-make-ai-too-smart-to-be-fooled-by-hackers-8dc37c084553
['Digital Times Africa']
2020-11-02 16:55:20.238000+00:00
['Artificial Intelligence', 'Technology News', 'Technology', 'Hacker', 'Tech']
1,922
Reasons Why You Need to Use a Broker for Buying Cryptocurrency
A crypto exchange broker program is not only helpful for trading but also comes with a lot of benefits of a smooth trading experience on behalf of the traders. Read on to know more. These days, it is approximately impossible to scroll through social media news feeds without watching headlines regarding cryptocurrency. The worldwide phenomenon of blockchain technology sweeping the world in a relatively short period has impacted various communities globally. The latest approval by various regulators for launching Bitcoin futures exchange happens to be a clear sign that blockchain is here to stay. Amateur users are looking for the most ideal place for purchase in Bitcoin regarding investment purposes or to utilize it as a payment system. When it comes to the digital currency, users are likely to open up Google and just proceed with the initial cryptocurrency exchange that comes on the search page. This can be considered as a big mistake as it is not how to start with crypto trading. When it comes to beginners, they need to always conduct their research as the most popular Crypto exchanges may not always be the best place for buying cryptocurrencies. In this article, there are the reasons why it is beneficial to consider a Crypto exchange broker program for buying cryptocurrencies. Read on to find more. Reduced Slippage Fees While choosing a new Crypto exchange, amateurs mainly make the common mistake of picking one based on its low trading attribute. It may appear like a logical decision and lucrative proposition for maximizing a fiat deposit within an exchange for more Crypto. However, choosing a Crypto exchange based on liquidity might result in high slippage fees that will crumble up your funds. This is where broker programs come in. Some brokers might help amateurs buy and sell throughout the most liquid Crypto exchanges (such as PayBito) to verify where the highest liquidity prevails for a particular asset and split the order throughout those changes. Beginner traders utilizing brokers can benefit by getting the lowest spread, best market rate, and minimum sleep age that can boost possibility in the long run. Greater Liquidity Pool For each asset that is sold or bought, whether it is Crypto or forex, it needs to possess someone to buy or sell from. The more traders conduct buying and selling, the more liquid a Crypto exchange happens to be considered. In the case of new Crypto exchanges, it populates each of its trading pairs and the order books with buyers and sellers for providing the users with a market to trade. This leads to distributing the trading liquidity from prevailing exchanges that might negatively affect smaller exchanges traders might be presently trading with. Here is where the Crypto exchange broker program comes in. Crypto brokers can help beginners avoid this problem by evaluating the market and acting on instruction from the trader to conduct reading according to an estimate or quoted price. By utilizing distributing buys and sells throughout the highest liquidity exchanges globally, brokers can increase their combined order book liquidity concerning a particular asset. It indicates that its users can trade crypto assets with maximum available liquidity. Increased Crypto Pairs to Trade Being stuck within the trade and not being capable of getting a fill might be a common problem on crypto exchanges offering 100 to 200 tradable assets. Based on the regular trading volume, crypto pairs on exchanges have their own order book for every asset against fiat currency and additionally Bitcoin. It indicates that less liquidity happens to be available on the crypto exchange, resulting in higher slippage and spread charges for crypto assets traded. Here, a crypto exchange broker program can be a savior. Crypto brokers come with the ability to offer assets without liquidity losses, and it happens to be a major benefit. Not only does this indicate lower spread and slippage charges, but users can also utilize a particular crypto broker trade, manage a cryptocurrency portfolio, and store assets within the particular location without requiring to register with multiple crypto exchanges to trade. Final Remarks If you are looking forward to buying Bitcoin or additional cryptocurrencies, it is recommended to do detailed research of the exchange beforehand and check slippage and spread fees associated with low liquidity order books. There happens to be an alternate option to the manipulation, deception, high spreads, hyperinflation by many crypto exchanges. The solution is to find a famous crypto broker providing better market rates and lower fees that can help increase profitability in the long run as cryptocurrency continues to increase its global adoption.
https://medium.com/@austinryan90/reasons-why-you-need-to-use-a-broker-for-buying-cryptocurrency-7165635f24e6
['Austin Ryan']
2020-12-22 13:35:02.631000+00:00
['Blockchain', 'Blockchain Technology', 'Cryptocurrency', 'Blockchain Development', 'Cryptocurrency Investment']
1,923
Computer Vision: A Step Closer to Skynet
Don’t get intimidated by the title of this post, I don’t intend to take over the world with Artificial Intelligence (Debatable, actually). I’m sure most of you remember Skynet, a self-aware Artificially Intelligent agent, back from the Terminator series. Skynet was portrayed in the movie-series to have a sinister idea of eradicating humans from the earth by creating T-800s. Contrary to fictitious believes, AI will not conquer humanity, rather it will enhance our capabilities. AI is a by-product of our intelligence and will place human civilization higher on the Kardashev scale. To mimic Human Intelligence, we need a machine with memory, processing power, godly algorithms, synthetic senses and cognitive capabilities. Let’s focus on the sense of vision that we are bestowed with by Apollo, the God of Light. (Yes, our vision is cent per cent helpless in the absence of visual spectrum of light). We constitute most of our cognition via our eyesight; Reading, writing, memorizing, recognizing objects, deciphering our environment and so on are done primarily by the use of our vision. Cameras are all around us. They are just a synthetic version of our eyes. Our brain is responsible for making sense out of the electric impulses sent out by the optic nerve. This collectively is called the Visual Cortex. So, machines have a camera but no ‘Visual Cortex’. Finally, this is where the subject of Computer Vision drops by. Simply put, it is the process of acquiring, processing and analyzing digital images. Our journey to achieving Computer Vision has seen several iterations. Early algorithms focused on locating edges, finding colours and recognizing shapes. As the tasks got complex, the methodology improved. Skip forward to the present, we have techniques such as convolutional neural networks (CNN) that have vastly improved what a machine can understand from a camera feed. From mere algorithms, Computer Vision has moved on to employ Deep Learning techniques to get more out of digital images. CNNs are inspired by the visual cortex of cats. They consist of neurons connected in neural layers. Hurray to another nature-based synthetic spoof! These CNNs have enabled the capability of namely two functionalities: Image Classification and Object Recognition. Image Classification Image classification revolves around labelling an image. Let me take the most cliche example of all time, the Cat and Dog classification problem. The task is simple: in a given data set of cat and dog images, the machine should be able to accurately label the images according to the animal represented in it. This is achieved by using convolutional neural networks such as Inception (developed at Google Inc.) where a training dataset of cats and dogs is fed to the model and the model learns from these images. How it learns, you may ask? The layers in the CNN are responsible for Feature Extraction and Feature Combination (called Feature Engineering, collectively) where each neuron is triggered just as they are triggered in our brain. Cool, isn’t it? Object Recognition What if there is both a cat and dog or maybe multiple cats and dogs in a single image? Image Classification is going to struggle with such a problem. The new objective is to localize the relative position of the cats and dogs in an image. This can be done by Object Detection algorithms such as Single Shot Detectors (SSD), Region Convolutional Neural Networks (R-CNN), RetinaNet and YOLO (You Only Look Once). These networks work by dividing an image into regions and performing predictions of these regions. We, at Integration Wizards, leverage Computer Vision techniques to solve a myriad of problems faced by our customers. Some of them are as follows: Intrusion Detection - So there is an existing CCTV network deployed in a facility but only a single security personnel to monitor them all. Error-prone and obsolete methodology in my opinion. Here, using object detection models, we detect an intruder in the view of the CCTV and generate alarms. PPE Compliance - Compliance is of utmost importance in facilities of blue-collar labour. Accidents caused due to non-compliance can be catastrophic to an organization. Here, we developed a model to detect if a person has their safety equipment on using a CCTV network, all in real-time. Fire Detection - As the name suggests, we’ve developed a model to detect an instance of fire or fire-related-smoke. Pose Estimation - So detecting humans is easy. We are taking it up a notch and estimating the pose of a human. Clubbing this with RNNs and LSTMs, we can easily predict the action performed by a person, be it walking, running, leaning, bending, et cetra. Inventory Management using Drones - Ah, how can we not include drones in our arsenal? We’ve devised an inventory management system with an autonomous indoor drone that works with Ultra Wide Band (UWB) indoor tracking technology. The drone can batch-read QRCodes or Barcodes to perform inventory checks. We live in an age where technology is rapidly pacing forward and we rather than trying to keep up, are trying to be at the forefront. Computer Vision is just one of the cognitive skills a machine possesses. We are closer than ever to achieving Skynet, yet there’s a long way left. I’ll keep churning out posts as often as I can. Until then, in classic Terminator style, I’ll be back :)
https://medium.com/@akash-james/computer-vision-a-step-closer-to-skynet-9a3692eee243
['Akash James']
2019-02-25 12:34:41.912000+00:00
['Object Detection', 'Pose Estimation', 'Machine Learning', 'Technology', 'Deep Learning']
1,924
Artificial Intelligence in Construction — TechVirtuosity
[Copyright : Pop Nukoonrat] © 123RF.com Revolutionizing Construction Construction and the methods we use are crucial to our success in modern architecture. We build houses and massive structures using our computers and we harness the that processing power to create new solutions. But artificial intelligence in construction takes things to a whole new level! It’s a tool that can help us push the boundaries further and it can do a lot to the industry as a whole. So then why haven’t we seen more innovation? The Construction Industry is Stagnating This isn’t to say that there hasn’t been a lot of improvements throughout the years, but construction has remained slower to adapt. In the past we often assumed that productivity equaled larger machines and that theory worked for a while. But now days we need something more than bigger machines, we need smarter machines and solutions. And while several other industries such as retail, medical and businesses in general have expanded, construction has fallen a bit behind. We simply need to adapt and use more technology. But what if we had more artificial intelligence in construction? Would this technology help lead us to a utopia? How Artificial Intelligence Helps Construction While it’s still early on some parts, AI has proven to show some promise in reducing costs. There is also software out there known as building information modeling, or BIM for short. AI can be trained to help suggest improvements and build solutions early on. It can also be used in risk management/mitigation, by providing safer alternatives. Construction robots are becoming more popular along with 3D printing, but add AI to it and we have a new advantage. AI can do the things that are too risky for us to risks our lives with. While using AI in this way is still new and very early, there’s a multitude of other areas it can help in. Of course, being a young technology also brings on risks for those using the technology early on… The Early Risks Involved Artificial intelligence in construction is a great solution but more technology also brings on different risks that also need to be considered. Anytime technology is involved we can typically also inherent the risk of getting hacked. If construction software or robots were to get hacked it could jeopardize an entire project. The argument is that it’s safer to have physical workers doing the actual work than to have robots or AI trying to take over. This is only partially true though. Construction steadily accounts for 20%+ of yearly deaths at work. AI poses the risk of hackers but as it stands the death toll is high right now, without the involvement in these life saving technologies. Hacking risks aside, AI isn’t perfect and does make mistakes too! This isn’t always the case but it’s important to recognize that mistakes happen with new technologies. The costs to implement AI could cost more money if it’s done wrong. But it’s not all bad! [Copyright : Kittipong Jirasukhanont] © 123RF.com Machine Learning can Mitigate Risks Machine learning is an important aspect of using artificial intelligence in construction. It allows a program to continually test and essentially “learn”. We’ve seen AI used in the field of medicine with success in the recent past which shows promise. But machine learning gives us some control. But first, in case you didn’t know what machine learning is… It’s a method used to teach an AI how to accomplish something. It is given parameters to gauge success and failure, a way to remember the results and a way to improve the numbers. Think of it like a race car, if it crashes it fails, if it completes the course it succeeds. Machine learning can take it a step further though. It can take that concept of the race car and find the most optimal way to complete the course, and that’s why it can be more productive than humans. We use machine learning to run thousands of trials and errors to succeed. This makes the AI more capable than a human which is why it can benefit construction. A single person can only try so many times, whereas an AI can run tests or simulations thousands of times endlessly. Construction can benefit from an AI that can continually learn the most efficient and safest way to build or solve problems. Machine learning can also incorporate previous knowledge in its tests, improving the outcome or immediate start of the training. As the AI grows it’ll also show us improved ways of doing things. Smart AIs Equal Smarter Solutions The future demands solutions that are safe, cheaper and ultimately faster. Construction is always a time sensitive task because a lot of it happens outside! The weather impacts it just as much as the efficiency of the workers doing it. AI provides solutions that are faster and cheaper which means there is less risk of weather delaying a project. The faster a task is completed the less likely something will go wrong, in theory. AI can help with many different areas by… Providing a solution to labor shortages through the use of software automation and solutions. Reducing risks to safety by spotting flaws or creating safer alternatives. Actively monitoring work environments and regulations. Can be used to collaborate on building plans while making smart suggestions (BIM software mentioned above). Providing analytics and statistics online for clients or workers. [Copyright : Preechar Bowonkitwanchai] © 123RF.com Artificial Intelligence in Construction Should be Embraced We talked about a lot of the areas that AI can help with but it needs to be given the opportunity. While some companies are already using this technology, there are still many more that are not quite there. AI can be used to help keep track of projects and to reassure clients. Whether it’s the city contracting the work or a variety of other businesses. Clients could benefit from seeing the progress online. AI has a lot of potential and even the user interface could have AI to learn what clients find most useful to view online. In the end, AI is going to be in the future of the construction industry. Whether we want it or not, it will help push innovation forward. But what do you think? Should we use more AI or avoid using those technologies in construction? Drop a comment below!
https://medium.com/swlh/artificial-intelligence-in-construction-techvirtuosity-124de131f26
['Brandon Santangelo']
2019-09-21 17:14:25.582000+00:00
['Construction', 'AI', 'Artificial Intelligence', 'Technology', 'Machine Learning']
1,925
Why “Yield” Is an Excellent Python Keyword — Generators and Coroutines
Suspend and Resume Execution in Coroutines Note: It’s important to know that Python is phasing out the use of generators as the basis for coroutines. Check the official documentation. However, the concept of coroutines can be easy to understand if we know that yield has the meaning of “giving way.” In this usage, it’s applicable to yield ’s second meaning: give way. It’s pretty much like the traffic “yield” sign in real life. You temporarily stop your vehicle such that other vehicles with right of the way can go first. Applying this analog to programming, it’s the basic idea of coroutines. As you can see, the key to coroutines is that the execution is suspended temporarily and it can be resumed later on when it’s applicable. Doesn’t this sound familiar to you? The generator function runs and yields the element (the execution is suspended), and when the next element is requested, it continues to run from where it was suspended (the execution is resumed). Actually, we can use the yield keyword to implement a coroutine. Let’s see the following code snippet for a simple example: We first define a coroutine function using the yield keyword. As you can see, the overall structure is similar to the generator function. However, you’ll notice that the yield keyword appears after the assignment operator (i.e. the equals sign). keyword. As you can see, the overall structure is similar to the function. However, you’ll notice that the keyword appears after the assignment operator (i.e. the equals sign). The assignment is possible because, in addition to the typical next() function that works with a generator, generators also support the send() function. function that works with a generator, generators also support the function. In the example, we send a string data (i.e. “Here’s the data that your requested.” ) back to the coroutine. Specifically, the data that we sent is set to the variable x , and we’re able to print the x value as expected. ) back to the coroutine. Specifically, the data that we sent is set to the variable , and we’re able to print the value as expected. The StopIteration exception is expected because the generator function has reached the end, which is consistent with the behavior of a generator (as shown in the last section).
https://medium.com/better-programming/why-yield-is-an-excellent-python-keyword-generators-and-coroutines-d0d9d9403171
['Yong Cui']
2020-11-04 17:48:37.297000+00:00
['Python', 'Technology', 'Software Engineering', 'Artificial Intelligence', 'Programming']
1,926
Now is the Time to Level the Digital Playing Field and Ensure Connectivity for All
Now is the Time to Level the Digital Playing Field and Ensure Connectivity for All Lili Gangas Follow Aug 5 · 4 min read The COVID-19 pandemic has exposed many disparities that have long been overlooked, few more obvious than the necessity of technology to participate in our increasingly digital world. With education transitioning to distance learning, jobs requiring telecommuting, and telehealth appointments skyrocketing, our reliance on technology and reliable internet access has increased substantially. It has become the difference between being able to earn a living, receive an education, and maintain health and well-being — or not. This increased reliance shines a spotlight on what many low-income families, often Black and Latinx households, and tribal and rural communities in California already knew: they lack equal access to the technology and reliable broadband needed to learn, work, and thrive in society, and have for decades. In California, over 1.5 million students in grades K-12 (or 25%) lack adequate access to the internet, and 60% (or 924,258) of these students are Black, Latinx, or Native American. Without access to instruction, assignments, tutoring, and social supports, students of color, low-income, and rural students lack equitable opportunities to learn. Despite numerous philanthropic efforts, the California Department of Education estimated that over 700,000 laptops and over 300,000 Wi-Fi hotspots are still needed to ensure all students are connected before going back to school in a few short weeks. Access to technology and broadband is also crucial to participation in the 21st-century workforce, to find and apply for work, access government resources, and telecommute. Yet, Black, Latinx, and Indigenous households are much less likely to have reliable broadband connectivity and technology, and face significant burdens in workforce participation and education. This was true before the COVID-19 pandemic, and those challenges have only been exacerbated. For those with access to broadband, our increased reliance on technology has also threatened to overburden antiquated networks. Video conferencing tools like Zoom require synchronous internet access with increased download and upload speeds per household. To put it simply, the more devices connected to the same internet access point, the lower quality experience may be, thus further driving the need for higher-speed networks. With trillions of dollars spent as part of pandemic recovery efforts, it is critical to prioritize digital equity by building fiber broadband networks in all zip codes that lack them and ensuring affordability. This isn’t to say that current solutions, such as creating Wi-Fi hotspots in school buses, aren’t helpful. But we must acknowledge that these investments are short-term fixes to a larger systemic issue. We must invest in fiber broadband to ensure high-speed connectivity across all of California. While broadband access is only the first step in a series of interventions needed to close the digital divide, legislators are currently working to make equitable access a reality. California Senator Lena Gonzalez is prioritizing high-speed broadband access with S.B. 1130, the “Fiber for All” bill. If passed, S.B. 1130 would unlock $300 million sitting in the California Advanced Services Fund (CASF) that can be used to build broadband infrastructure where there are currently no service providers. The bill would raise minimum standards for telecommunications companies providing broadband service to communities by requiring any broadband network funded by the state to be high-capacity fiber and open access. A competing bill introduced by Assemblymember Aguiar-Curry, A.B. 570, also aims to use resources from the CASF to fund broadband infrastructure projects and prioritize underserved communities, while prioritizing DSL instead of high-speed fiber. Broadband access is also being considered at the federal level, with the U.S. House of Representatives passing H.R. 2, the “Moving Forward Act”. The bill proposes $1.5 trillion in infrastructure funding, including $8.8 billion for a “broadband benefit” program to help low-income households and recently laid-off consumers pay for internet connections and $1.3 billion in funding for state and community digital inclusion initiatives. The Senate version of H.R. 2 allocates $100 billion for broadband access. As the tech capital of the world and an epicenter of education, innovation, and creativity, California should prioritize access to high-speed broadband for all residents. Both S.B. 1130 and A.B. 570 are being considered by legislators, representing a defining decision on whether California will move toward the future focused on digital equity or not. The COVID-19 pandemic has shown us the consequences of ignoring disparities in something as basic as connectivity. We must focus on building infrastructure to ensure California communities are connected. We must invest in high-speed broadband networks now to ensure we are set up to be equitably and competitively networked for our future. If nothing is done, the impact will be felt across our state for decades to come. Lili Gangas is the Chief Technology Community Officer at The Kapor Center.
https://medium.com/kapor-the-bridge/now-is-the-time-to-level-the-digital-playing-field-and-ensure-connectivity-for-all-ee507e95652c
['Lili Gangas']
2020-08-06 06:44:33.502000+00:00
['California', 'Technology', 'Distance Learning', 'Digital Equity', 'Digital Divide']
1,927
5G Hype vs. Reality: How Policymakers Can Harness Its Benefits
For some, 5G is a miracle network that will bring the internet’s speed, abundance, and possibility out of our pockets and throughout our physical world. For others, it’s the next national security battleground, the winner of which will be the world’s hegemonic superpower. For most, it’s probably just an ill-defined tech buzzword echoing out of podcasts, banner ads, or business headlines. All three of these associations may ring true. From these disparate and dramatic images, it can be difficult to assemble a concrete reality. What, exactly, 5G? And what does it mean for economies, societies, governments, and nations? A new briefing from the Berkeley Roundtable on the International Economy (BRIE) in support of CITRIS’ Work in the Age of Intelligent Tools and Systems (WITS) Initiative starts to fill in some of the blanks in common 5G understanding and ground its most significant issues. While 5G will be costly to build and could magnify existing unresolved cybersecurity threats, it could increase productivity across economic sectors and open up a more competitive landscape in the technology industry. As seen in many telecom industry reports, the “fifth generation” of network technology promises to deliver incredibly fast speed and response times, enabling an untold number of futuristic applications from autonomous vehicles to smart manufacturing and remote medicine. But building a network that can support these applications is tricky; equipment broadcasting at sufficiently high frequencies needs to be placed every few hundred meters, which makes for a complicated, prolonged, and hugely expensive infrastructure project. Initial service offerings labelled “5G” are being installed on 4G towers and use frequencies that, while higher than 4G, are too low to support significantly increased speed and response times; they can allow for more consistent high-speed broadband capacity, but don’t have the coverage to support the pervasive fog of internet-connected devices that 5G enthusiasts envision. Meanwhile, even the more powerful 5G-powered applications are not projected to generate significant new revenue for network providers ; this limits the return operators like AT&T and Verizon would likely see on their expensive investment in a dense, high-frequency network. Given this, these operators are unlikely to make the investment in establishing a dense high-speed network on their own. They could make the investment more manageable by sharing costs with each other, or with Google, Amazon, and other application providers who stand to benefit handsomely from fully-enveloping 5G coverage. But there’s no clear indication that operators have the inclination or wherewithal for either. This leaves governments with a dilemma: how critical is 5G to the national interest, and how far should they go in pursuit of it? One could argue that applications enabled by 5G will unleash significant, cross-sector productivity gains. Moreover, 5G could be the active ingredient in the evolution of the technology industry; the first 5G nation could be the one to develop the next generation of platforms and tools that dominate the global economy. Perhaps in service of this theory, China has been active in facilitating cooperation and cost sharing among network operators, manufacturers, and application providers. Encouraging or enabling such collaboration is one relatively low-risk way for policymakers to help make 5G a reality. Other options for governments promoting 5G include becoming a major user of the network or directly subsidizing its construction . But lack of a powerful network is not the only factor holding back deployment of the 5G technologies that could supercharge productivity; there are many reasons other than lack of network speed and stability preventing the deployment of 5G-enabled applications such as autonomous vehicles, massive IoT, and bioelectronic medical implants. If the technologies meant to use 5G lag behind the network itself, government-sponsored 5G might get little use, with little economic benefit. Seen from this perspective, 5G is a risky bet. The US government clearly wants 5G to be deployed as quickly as possible, as shown by their recent decision to auction off a section of mid band spectrum within the next year . But while selling spectrum and waiting for operators to broadcast over it might not be enough to spur construction, more active policies may not come to their intended fruition. Policymakers need to weigh these considerations carefully. Policymakers also need to consider how 5G could expand and magnify the internet’s challenges to every dimension of security. On this subject, the possibility that China’s government may be pressuring Huawei to place vulnerabilities in its increasingly-ubiquitous network equipment to facilitate political espionage has gotten the most attention. However, it’s the unintentional gaps in the security of Huawei devices that deserve more scrutiny; they draw notice to how a dense 5G network of mobile broadcast sites and a far denser jungle of devices will give bad actors millions of potentially vulnerable entry points into governments, businesses, and personal lives through which to steal and manipulate. Finally, 5G could also unsettle or even tip over the established fortresses of the digital world. For the last ten years, Google, Microsoft, and Amazon have leveraged cloud computing to shape the internet to their liking. But 5G could very well dim the power of massive datacenters and spark a multiplication of computing and storage out to base stations or remote devices themselves. If this happens, new players, whether network operators or unheard-of upstarts, could could control network infrastructure, and capture more value in the next phase in digital evolution. This possibility reinforces another idea that policymakers should bear in mind: the competitive landscape of technology firms is not fixed, and those that appear dominant today may be vulnerable to disruption tomorrow. 5G may never live up to its hype, but the questions it inspires and informs are profound. With 5G’s economic potential, policymakers could at least explore ways of fostering collaboration among network operators to get the network built, while closely monitoring the development of 5G-dependent technologies to see if more direct action might be fruitful. Policymakers could further prepare for 5G by incorporating it into the design of cybersecurity standards; simply ripping out Chinese telecom equipment, as the FCC last week effectively enjoined US companies to do , is not nearly enough to guarantee security in a 5G world. Finally, policymakers could also consider 5G when making assessments of the competitive landscape in the technology industry. Overall, understanding the pitfalls and possibilities of 5G will be critical for policymakers and other leaders as they seek to shape technology’s ongoing evolution for the benefit of all. This work was funded in part by the Ewing Marion Kauffman Foundation and led by the Berkeley Roundtable on the International Economy (BRIE) Work and Intelligent Tools and Systems (WITS) Program at UC Berkeley, with support from the CITRIS Policy Lab
https://medium.com/citrispolicylab/5g-hype-vs-reality-how-policymakers-can-harness-its-benefits-399252ae0b54
['Citris Policy Lab']
2020-01-23 08:29:00.698000+00:00
['Policy', 'Telecommunication', 'Internet of Things', '5g', 'Technology']
1,928
Zero knowledge proofs and nano node
Zero knowledge proofs and nano node Syncing in seconds thanks to ZPKs with nano nodes in Nimiq 2.0. Looking back, nano syncing in Nimiq 1.0 worked by leveraging two technologies: light clients and NiPoPoWs. How is that going to work in Nimiq 2.0 with Proof-of-Stake? Let’s get into the technical details with this blog post series. One of the big strengths of Nimiq are the nano nodes and how they allow anyone to sync and connect to the blockchain from a regular phone in just a few seconds. They enable us to deliver Nimiq’s vision of making crypto accessible to the masses. Light clients are ubiquitous in crypto and simply refer to nodes that only download the headers of the block and ignore both the transactions and the state. This evidently greatly reduces the amount of data that a user needs to download, but it still is too much for a browser or a mobile. Even Bitcoin, with its small block headers and very large block time, requires its light clients to download roughly 50 MB to sync from the genesis block. And that number is growing at a steady 4 MB per year. Calculating these numbers is simple as Bitcoin block headers are always 80 bytes. For Nimiq, with a blocktime of one minute and a header of 146 bytes, the size would be almost twenty-fold. That is not light at all 😉. NiPoPoWs, or Non-Interactive Proofs of Proof-of-Work, are short proofs that can prove that a given amount of computational power was spent in producing a chain of blocks, without the need of taking all block headers into account. In essence, it allows a light client to sync the blockchain while downloading only a small fraction of all the block headers. Without getting too technical, it does this by identifying blocks that greatly exceed their difficulty target, which happens now and then. If you want to read more on the topic, check out nipopows.com or the Nimiq 1.0 whitepaper. However, now that Nimiq is transitioning to a Proof-of-Stake-based protocol, we can no longer use NiPoPoWs, since they are specific to Proof-of-Work. So, we started to work on a novel solution to deliver Nimiq’s vision of being browser-first. It turned out, developing nano nodes for a PoS blockchain is an interesting challenge as the blockchain research community is just starting to work on this field. Thanks to the efforts of the team’s researchers, Bruno and Pascal, and their longstanding experience in scientific and blockchain research, a solution using zero knowledge proofs has been found! With this post we will explain, in some technical detail, what were the issues we faced and how we found an optimal solution. If you want to meet Bruno and Pascal, check out this video where they explain Nimiq 2.0 tech. Warning, technical content ahead! 🤓 To get the best out of this post, you should have a basic understanding of what elliptic curves, zero knowledge proofs — or ZKPs for short — and SNARKs are. And here are some resources for those interested in diving deeper: The best site for ZKPs is zkp.science. It’s a collection of resources about ZKPs. Also, this GitHub repo has some great links to further reading Or read Vitalik’s famous three post series on ZKPs: First, second, third. Albatross, the new consensus algorithm in Nimiq 2.0, will have a block time on the order of one second. That means, while Bitcoin produces 52'560 blocks per year, Nimiq 2.0 will produce 86'400 blocks per day! While this performance improvement is incredible, with so many blocks being produced, syncing the blockchain becomes a new unexpected challenge. Now, let’s have a look at how a light client works in Albatross. A regular light client in other blockchains forces the user to download all of the block headers from genesis to the current time. This is necessary in order to find out what the most recent block is. Next, the user can request Merkle proofs to whatever part she is interested in. Merkle proofs allow a light client to verify that a given transaction, or account, is part of a block without needing to download the entire block. Albatross, like many other PoS algorithms, picks a random set of nodes to be validators — called the validator set — and lets them produce a fixed number of blocks — which is called an epoch. At the end of an epoch, a new set of nodes will be randomly picked. In Albatross, macro blocks mark the end of an epoch and the beginning of another one, it’s where the validator set changes. Between any two macro blocks, the validator set doesn’t change, that means that we can safely skip all of the micro blocks between two macro blocks. So, a light client in Albatross would download only the macro block headers since genesis to the most recent one and all of the micro block headers produced in the current epoch. Then, once again, the user can ask for Merkle proofs for parts of the state. Skipping the micro blocks is already a huge improvement. Macro blocks might only happen once per day — the exact frequency is still to be decided — while micro blocks happen tens of thousands of times per day. Let’s do some math. Considering that a block header for a PoS chain needs to include the public keys of each validator and assuming 1024 validators per epoch, that will give us 290kB for the public keys only. Then, a quick calculation — assuming a macro block header of 300kB and one macro block being produced per day — shows us that the required download would grow to 109MB within just one year. That is too much for a web and mobile environment! This got us on the journey of finding a better solution: Zero knowledge proofs, more specifically zk-SNARKs! Zk-SNARKs, or Zero-Knowledge Succinct Non-Interactive Argument of Knowledge, are probably the most famous type of zero knowledge proofs and have found a lot of use cases in blockchain recently. As a zero knowledge proof, they are known for their ability to prove arbitrary statements without revealing any other information other than that the statement is true. This property is leveraged by Zcash, which is also arguably the first blockchain to make extensive use of zero knowledge proofs. However, we are not interested in the zero knowledge property of SNARKs but rather on its succinctness. Succinctness refers to the ability of SNARKs to take a statement of any size and prove its correctness with a very small proof that is very quick to verify. A statement in this case can be a simple calculation or an entire program that returns true or false. So, in other words, this allows us to do some large calculation and then produce a proof that the calculation was performed correctly. Then, anyone can simply check the proof instead of redoing the same calculation. SNARKs are exceptionally succinct, producing proofs with sizes of a few hundreds bytes that take only milliseconds to verify. If you want to dive deeper, consider reading this article. Normally, a light client needs to download all of the macro block headers to verify that the last macro block header it received is indeed authentic. All of those intermediate macro block headers are just a way of getting from the genesis macro block to the most recent macro block. So here lies the optimal solution: Using the succinctness of SNARKs to compress all that computation into a small proof! That leads us to nano nodes. While light nodes need to download all of the macro block headers, nano nodes only ask for the most recent one and a SNARK attesting to the fact that there is a valid chain of macro block headers that start with the genesis block and end in the most recent block. Note, we can assume that it has the genesis macro block header already as it is public knowledge. The SNARK is enough for the node to be convinced that the macro block header that it received is valid. After that, it just needs to download the headers of the micro blocks in the current epoch plus whatever Merkle proofs it needs. This ideal solution would mean that a user can sync with the blockchain in a few seconds maximum. The data that needs to be downloaded would include the last macro block header of ~300kB, tiny 200 bytes for the SNARK, about 400 bytes for each micro block headers in the current epoch and ~200 bytes for all the necessary Merkle proofs. So, in total, we’re talking about ~400–500 kB to sync. That is completely doable in a mobile/browser environment! But the best part is that the syncing time will never increase! Indeed, it’s a fixed size, no matter how large the Nimiq blockchain grows, it will stay the same! However, while verifying SNARKS is very quick, creating them is a big effort. So, while it’s theoretically possible to create one new SNARK verifying the entire chain of macro blocks at the end of each epoch, it would put a big burden on the validator. Instead, we are using a strategy where our SNARK proof will be extended with each new macro block without growing the size of the proof. We can do this by using a very recent, very promising technique called “recursive SNARKs”. This has been a long and technical post, thanks for coming with us on this journey. Let’s take a break and explore the specifics on how to achieve this in part 2 of this series. Pura vida, Team Nimiq
https://medium.com/nimiq-network/zero-knowledge-proofs-and-nano-nodes-nimiq-ec1713fd332e
['Team Nimiq']
2020-08-28 18:55:32.624000+00:00
['Nimiq', 'Blockchain Technology']
1,929
Weekly Blockchain News with Mammycrypto Dec, 5th 2020
December is here, and we have realized that 2020 has passed by us like a mad train. Despite the chaos, the crypto markets not only survived one of its biggest challenges, the covid-19 pandemic fear, but it proved to be a hedging solution for many. There is a current need for an easy way to transact. And the world is currently feeling the pressure to adapt to a digital financial spectrum. Bitcoin together with other cryptocurrencies has been proving its usefulness and is being adopted by large corporations such as PayPal and attracting the attention of large investors and investment funds such as Michael Saylor and Jp Morgan. Who knew everybody one day would jump on this train? Nobody was sure but we certainly hoped that the world would eventually accept and endorse this amazing technology. This week, the crypto markets have been moving sideways and in the last few days fluctuating between the highs $18000s to the low $19000s. Truly impressive to see the price of Bitcoin standing steady close to its all-time high. The pressure is high! Super high, and we can expect some volatility in the course of more developments and headlines. The articles below are hand-picked for you, one article to read for each day of the week. They will help you understand the environment we are in and some of the catalyst topics that can translate into market actions as it develops. Check out the Major headlines in this quick 3-minute video below. Also, access each article provided in the video with the links below. I certainly recommend the read. Sunday, Nov 29th 2020: Pizza Hut Venezuela Now Accepts Crypto Payments - https://www.coindesk.com/venezuela-pizza-hut-bitcoin-dash Monday, Nov 30th 2020: Venezuelan army starts mining Bitcoin to make ends meet — https://cointelegraph.com/news/venezuelan-army-starts-mining-bitcoin-to-make-ends-meet Tuesday, Dec 1st 2020: The Libra Association Rebranded Itself to Diem https://decrypt.co/50063/the-libra-association-rebrands-itself-to-diem Wednesday, Dec 2nd 2020: Visa will facilitate USDC payments, thanks to fresh partnership — https://cointelegraph.com/news/visa-will-facilitate-usdc-payments-thanks-to-fresh-partnership Thursday, Dec 3rd 2020: S&P Dow Jones Indices to Launch Crypto Indexes in 2021- https://www.coindesk.com/sp-dow-jones-indices-to-launch-crypto-indexes-in-2021 Friday, Dec 4th, 2020: New Fund Exposes Canadian Traditional Investors to Ethereum -https://decrypt.co/50496/new-fund-exposes-canada-traditional-investors-ethereum Saturday, Dec 5th, 2020: Enjoy your Saturday! www.mammycrypto.com Social Media links: https://cointr.ee/mammycrypto
https://medium.com/@mamaecrypto/weekly-blockchain-news-with-mammycrypto-dec-5th-2020-eff28a6c370e
['Mamaecrypto Michelle M.']
2020-12-07 00:17:22.741000+00:00
['Cryptocurrency News', 'Technology News', 'Bitcoin Price', 'Crypto Trends', 'Blockchain Development']
1,930
Dear Apple, Please Make a Next-Gen Apple Cinema Display for ‘Normal People’
Pro Display XDR Alternatives The alternatives to the XDR are less appealing. While Apple uses aluminum, others still use plastic. Sure, some screens look good, but you can not go beyond good with plastic. Why aren’t more manufacturers using premium materials? The main problem with the XDR for the average consumer is that it costs $7.000. This is a screen most people should stay away from as it is targeted at high-end professional users. So where does that leave the average Joe? Nowhere really. There are no options and you have to look elsewhere. Even Apple lists LGs displays as options on their site. I’m not sure this is a sign they won’t bother with displays for you and me anymore. After days of research, there are a few alternatives you can consider. Good luck learning the model names though. BenQ One of the best options offered by BenQ is the Designer Professional Monitor with 31.5-inch, 4K UHD, Display P3 |PD3220U. With its similar size and great panel, it is a valid competitor. The display is high rez (although not 6k) with a 10-bit HDR panel. You can daisy-chain it to set up multiple monitors and it comes with a puck with buttons so you can quickly change color modes (if you’re a fan of more clutter on your desk) ASUS ASUS has a lineup targeted towards creative users. They have a whole series called ProArt. It is within this lineup we find some of the best alternatives to the XDR. The ASUS ProArt Display PA32UCX-K 4K HDR IPS Mini LED Professional Monitor — 32” There is a variety of models to choose from. The K model f.ex. has the calibration tool included. LG LG has a few options. Apple sells the 27-inch LG UltraFine 5K Display on their site and it is tailored for the mac. It doesn’t even have hardware menus as the OS will help you with the settings. It is packed with a web camera, microphones, speakers, and several USB-C ports. The thunderbolt plug also powers your MacBook. If you are looking for more screen real estate, you can have a look at the 34'’ LG 34WK95U-W UltraWide® 5K2K Nano IPS LED Monitor with HDR 600. Working with widescreen is amazing. The productivity levels are boosted and you don’t get a screen border in the middle of your view. Another popular option in the LG camp is the Ergo series. The LG 32UN880-B 32 Inch UltraFine™ Display Ergo 4K HDR10 Monitor is something to consider. The stand looks fantastic.
https://medium.com/macoclock/dear-apple-please-make-a-next-gen-apple-cinema-display-for-normal-people-1d87d40dc48a
['Martin Andersson Aaberge']
2020-12-12 06:29:27.773000+00:00
['Technology', 'Apple', 'Content Creation', 'Hardware', 'Creative']
1,931
Rethinking Separation of Concerns with React
Angular and React Separation of Concern Rethinking Separation of Concerns with React And how it differs from Angular Understanding the true nature of separation of concerns is one of those topics that separates newbies from the seasoned developer. There’s more to code than just getting the syntax right. Beautiful code is one that is modular, loosely coupled and extremely easy to maintain. I’ve been thinking about long term feasibility of React and SoC over the past few days since my original deep dive into the framework. When Angular 1 came out, everyone was excited because of how easy it was to do things. But as projects grew, complications started to arise — especially when there are multiple people working on a single code base. One could argue that it is the nature of programming — and perhaps it is, but it doesn’t have to be. Angular 2 came out and forced a stronger separation of concerns through explicitly requiring different types of files for different functionality. HTML lived in one file. Your component logic lived in another file. Relevant CSS is automatically localized to a single file and your services in another file. React challenges this way of thinking by encouraging you to put everything into one space. What is separation of concern? separation of concerns (SoC) is a design principle for separating a computer program into distinct sections, such that each section addresses a separate concern — Wikipedia Separation of Concern is supposed to help keep the code organized. A ‘concern’ is a particular set of information that is related to each other within the immediate vicinity. In short, separation of concern is a classification method. If you think about your local supermarket, each item categorized into item types. The sorting of all oranges, apples, and bananas into their different and grouped arrangement makes it easier for shoppers to chose which fruit they want. However, efficiency and sales are increased when related products are grouped together — like pasta and pasta sauce or rice and butter chicken mix. There is less walking required and the grouping of ingredients makes it easier to determine what else you need. From what I understand, separation of concern is all about organization and the conventional philosophy goes — group like and like together. Angular does this through file types. They don’t mix JavaScript and HTML together like React. However, this isn’t the only style available when it comes to organizing code and separating out their domain of concerns. Just because it’s a popular convention, doesn’t mean its right React challenges the traditional models of thinking towards separation of concern. The diagram below illustrates how MVC frameworks often structure their files and content. Separation of concern is determined by the file type and may be grouped together in a file of sorts. MVC styled separation of concern This structure is common in templating styled frameworks. However, the issue with this structure is that when problems or new features arise, it’s often implemented horizontally. The point of SoC is so that code remains maintainable and is mentally digestible by the developer. There’s no point of splitting code into a million different files if no one can understand it. We’re only human and have a finite short term memory computing space. SoC is supposed to reduce the load by grouping like for like together. Templating systems SoC The idea behind grouping like for like is also to reduce code coupling and increase cohesion. This means that a piece of code is independent from each other and can easily be interchanged like lego parts. However, that is often not the case — especially in templating frameworks like Angular. In large projects, this can actually increase code interdependency if you’re not careful. As we are forced to move horizontally and across multiple files, the structure and method behind this style of SoC make the code harder to maintain and understand. It gets even worse when an unmoderated junior enters the fray and it’s not their fault, not really. Challenging traditional thinking for Separation of Concern When Angular 1 came out, it was groundbreaking because it made view binding work like magic. It just worked. Then projects grew and grew, and everything started to fall apart. The partials started to get used everywhere and the originally separated templates became unwittingly interconnected. If you wanted to change something, you also needed to trace horizontally across the files and figure out what else is affected. Things get messy and soon, you’re more worried about breaking the code than writing it. React challenges the traditional method of file silos by flipping it on its side and group their code based on relationship to each other. Language is not a hard line barrier but is seen as one cohesive and logically related block. With React, there is no templating — just components that can be created and destroyed as required. relationship based SoC When code is grouped together based on relationship rather than an arbitrary requirement, processing the block becomes easier. The original mentality behind separation of concerns was created by the early pioneer in computing scientist Dijkstra in 1974, and in short, the purpose of SoC is the “effective ordering of one’s thoughts” React does this by turning grouping code into digestible blocks of code where the flow of logic can be easily traced. We’ve since evolved from the long single files and coding styles of the 80s and 90s — but that doesn’t mean that the current and popular methods are 100% correct for our needs either. That’s why frameworks are created and the ones that solve major problems experienced in other frameworks are the ones that thrive. Someone will still have to play the code janitor The point of frameworks is that they provide a scaffold for how you write your code. React takes this scaffold away and you’re free to structure the code however you want. SoC is designed for you when it comes to Angular. React lets you design your own and there is no limitation on how you arrange your code. You’re still going to need a code janitor nevertheless. It doesn’t matter how good a framework is, if not used properly, code will still fall apart. There still needs to be that one person that takes a step back, create the required conventions for the project and enforce it on everyone else. The code janitor plays an important role of keeping the code organized and ensure that modules and components remain cohesive rather than tacked on. The code janitor usually comes in the form of the tech lead. There is usually a code review to ensure that the code fits the way its supposed to — structurally, stylistically and functionally. No one is perfect and someone is bound to inject or do something weird sooner or later. Projects usually fall apart in the long run without a code janitor to keep everyone in line and clean after their own messes. The thing you must also acknowledge and remember is that all major and popular frameworks solves a problem of sorts — but the most effective one is the one that solves your problem. Every project and every team has their own personal styles and issues that they will encounter. Separation of concern works best when it serves the needs and future growth of the project, along with the team’s ability to understand and how much documentation they’re capable of processing.
https://medium.com/madhash/rethinking-separation-of-concerns-with-react-7e5de1b5c0f7
['Aphinya Dechalert']
2019-12-22 09:29:24.402000+00:00
['Technology', 'JavaScript', 'Startup', 'Productivity', 'React']
1,932
Why So Many Product Ideas Fail?
Now let’s analyze the top five product failure reasons in little more detail to understand them better. Reason # 1: Not Viable — Organizations don’t want 50% (half of all) of new ideas are rejected internally within the organization as they are not viable. Meaning — organizations don’t want them. There are several underlying reasons here. Few important ones are — idea isn’t aligned with the overall organizational strategy/direction; idea isn’t profitable for the business; lack of executive/leadership alignment on the value of the idea. Reason # 2: Not Feasible — Organizations can’t build at all 20% (one-fifth of all) of new ideas are rejected internally as organizations are unable to build them. Users may like those ideas but organizations either may not have the resources or the time to build them. At the time when an idea originates, the business value isn’t greatly known. This sometimes end up in our decision to deploy resources on other ideas that later prove to be of less business value. Reason # 3: Not Valuable — Users don’t want 12% (little more than one-tenth of all) of new ideas/products fail pre-launch or post-launch as users don’t see any value in them. This is common in organizations where product team is either spending too little time with the users (external) or too much time with the internal people within the organization. Key to success here is to spend as much time as possible with the users who will be using the product after launch. The practice of building the product and features driven solely by discovery conversation with the users is called “outside-in product development” and is a proven methodology for success. Reason # 4: Not Usable — Users can’t use 8% (little less than one-tenth of all) of new products fail after launch as users find them complex to use. This is the area where product designers are valuable as they work with the users and design the product in such a way that it is easy to use. A good interaction and workflow design are the key here. Bigger isn’t always better. Meaning — don’t load the product with a ton of features. Adding a new feature by disrupting the workflow of the previous feature may do more harm than good. Reason # 5: Not Timely — Organizations can’t build on time 5% of new products fail post-launch as they are launched late and unable to attract buyers. This happens when we take too long to productize and launch the idea. Time to market should be quick — good enough, soon enough. If we are not quick enough, our competitors may launch a similar or a better product sooner than us unless we are aiming for a disruptive innovation in an already saturated market (e.g. Uber).
https://medium.com/product-center-of-excellence/why-so-many-product-ideas-fail-6a90af06c15a
['Harsh Vardhan']
2020-09-27 02:40:24.577000+00:00
['Technology', 'Startup', 'Product Management', 'İnnovation', 'Leadership']
1,933
Packet Switching Technique and Different Switching Techniques
“There was a time when people felt the internet was another world, but now people realize it’s a tool that we use in this world.” ~Tim Bemers-Lee What is Packet Switching? The packet switching is a switching technique in which the message is sent in one go, but it is divided into smaller pieces, and they are sent individually. The message splits into smaller pieces known as packets and packets are given a unique number to identify their order at the receiving end. Every packet contains some information in its headers such as source address, destination address and sequence number. Packets will travel across the network, taking the shortest path as possible. All the packets are reassembled at the receiving end in correct order. If any packet is missing or corrupted, then the message will be sent to resend the message. If the correct order of the packets is reached, then the acknowledgment message will be sent. Advantages Of Packet Switching: Cost-effective: In packet switching technique, switching devices do not require massive secondary storage to store the packets, so cost is minimized to some extent. Therefore, we can say that the packet switching technique is a cost-effective technique. In packet switching technique, switching devices do not require massive secondary storage to store the packets, so cost is minimized to some extent. Therefore, we can say that the packet switching technique is a cost-effective technique. Reliable: If any node is busy, then the packets can be rerouted. This ensures that the Packet Switching technique provides reliable communication. If any node is busy, then the packets can be rerouted. This ensures that the Packet Switching technique provides reliable communication. Efficient: Packet Switching is an efficient technique. It does not require any established path prior to the transmission, and many users can use the same communication channel simultaneously, hence makes use of available bandwidth very efficiently. Disadvantages Of Packet Switching: Packet Switching technique cannot be implemented in those applications that require low delay and high-quality services. The protocols used in a packet switching technique are very complex and requires high implementation cost. If the network is overloaded or corrupted, then it requires retransmission of lost packets. It can also lead to the loss of critical information if errors are nor recovered. Different Switching Techniques Multiprotocol Label Switching (MPLS) Multiprotocol Label Switching (MPLS) is data forwarding technology that increases the speed and controls the flow of network traffic. Multiprotocol label switching (MPLS) is a technique for speeding up network connections that was first developed in the 1990s. The public Internet functions by forwarding packets from one router to the next until the packets reach their destination. MLPS, on the other hand, sends packets along predetermined network paths. Ideally, the result is that routers spend less time deciding where to forward each packet, and packets take the same path every time. Pros of using MPLS: Service providers and enterprises can use MPLS to implement QoS by defining LSPs that can meet specific service-level agreements (SLAs) on traffic latency, jitter, packet loss and downtime. It’s good for real-time applications that can’t tolerate latency, such as video, voice and mission-critical data. Different types of data can be preprogrammed with different priorities and classes of service. Organizations can assign different percentages of their bandwidth to various types of data. Cons of using MPLS: Cost: MPLS is more expensive than regular Internet service. MPLS is more expensive than regular Internet service. Long setup time: Setting up complicated dedicated paths across one or more large networks takes time. LSPs have to be manually configured by the MPLS vendor or by the organization using MPLS. This makes it difficult for organizations to scale up their networks quickly. Setting up complicated dedicated paths across one or more large networks takes time. LSPs have to be manually configured by the MPLS vendor or by the organization using MPLS. This makes it difficult for organizations to scale up their networks quickly. Lack of encryption: MPLS is not encrypted; any attacker that intercepts packets on MPLS paths can read them in plaintext. Encryption has to be set up separately. MPLS is not encrypted; any attacker that intercepts packets on MPLS paths can read them in plaintext. Encryption has to be set up separately. Cloud challenges: Organizations that rely on cloud services may not be able to set up direct network connections to their cloud servers, as they do not have access to the specific servers where their data and applications live. SD-WAN (Software-Defined Wide-Area Network) As the name states, software-defined wide-area networks use software to control the connectivity, management and services between data centers and remote branches or cloud instances. One of SD-WAN’s chief features is the ability to manage multiple connections from MPLS to broadband to LTE. Another important piece is the ability to segment, partition and secure the traffic traversing the WAN. SD-WAN’s driving principle is to simplify the way big companies turn up new links to branch offices, better manage the way those links are utilized — for data, voice or video — and potentially save money in the process. Benefits of using SD-WAN: Lower costs Increased bandwidth Higher performance More uptime Better performance at small, remote, or international sites Quicker provisioning and rollout Thank You! Open for improvements :)
https://medium.com/@ravindra-utkarsh/packet-switching-technique-and-different-switching-techniques-44b150a0d018
['Utkarsh Ravindra Kushwaha']
2020-12-19 08:35:48.074000+00:00
['Information Technology', 'Packet Switching', 'Sd Wan', 'Sd Wan Technology', 'Mpls']
1,934
Evolution of robots denies anthropocentric theories
Creationism, flat-Earth theory, geocentricity, and anthropocentricity are the major false conceptions from religions. Though the other three theories don't have so much of impact in human mind and behavior, the last one, anthropocentricity has impacted the way humans pursue the whole universe. When we started to believe that everything in the universe is created for our wellbeing, we have grown selfish and greedy over the other living beings. When we say that Earth is a resourceful planet, it clearly shows our mentality that we see other living beings as resources to satisfy our own needs. Anthropocentricity has been the reason for wars, slavery, cruelty against animals, famines, deforestation, extinction of several creatures, global pollutions and other man-made disasters. It is completely necessary to redefine our conceptions about the world. It is so required to accept and spread the theory of evolution throughout the planet. When asking about the proof for theory of evolution, the way we conceptualised robots from the beginning has provided essential evidence about the behavior of life in the world. Let's try it with a small thought experiment. Close your eyes. Think about a car. Slowly, drive the car across a highway. Cross the large industries and technological companies. Take the road to your home. Reach your home. Park the car. Walk to your home. Open the door. Now there's a robot waiting for your orders right at your hall. No matter what type of car, industries, roads, or house you would have imagined. Ultimately you would have imagined a humanoid that would take an order from you, when you had to imagine a robot. That's where anthropocentricity begins in humans. When we look for a replacement to do our work, we literally imagine a person instead of imagining a process of replacement to our work. This was the initial mentality when people thought that robots would replace humans at work. Movies such as “Terminator” essentially proved this mentality. As time progressed, as robotics started to take it's first steps into automation business, the robots where nothing but specialised machines to do the tasks rather than actual humanoids, which could do every task in a humanly manner. As time progressed, as robotics evolved, we moved into concepts of AI as a set of softwares rather than creating a complex mechanical device, which could replicate the actions of human brain. Today robotics is present in every industry. It is present as: rovers in exploration process, static machines in manufacturing process, as drones in security and defence works, and complex algorithm in automated cars, chatbots etc. Robots today are very specialised to do specific tasks. Hence, they take different forms. We don't need a humanoid to drive a car for us, or do a chat with us, or do an industrial process. Every robot is specific, custom-made for its own purpose, and cannot be interchangeable with a robot specialized for some other task. This being a reality, scientists in the recent past have realised the need for humanoids and started researching on them. The reason for the same being that specialised robots need specialised environment for their operations. Meanwhile, humanoids can provide agility, adaptability, and interoperability irrespective of the environment they work on. So, if we have seen the evolution of robots in implementation, we could actually see the transformation of robots from basic forms designed for simple applications in a given work environment to sophisticated humanoids that could do complicated applications in various environments. Though the robots are evolved, the basic robots are still in existence due to their cost efficiency and application specifications. It kind of reflects the theory of evolution in reality and shows us that the world is not created for humans, but humans have evolved from the basic creatures through various intermediate creatures, which may or may not exist today. In another view, if we think about imaginations of robots, it is seen that we had started by imagining a humanoid and gradually drifted to application specific mechanical and software robots for specialised applications. Both these conclusions intersect as evidences for anthropocentricity being a false conception about the world.
https://medium.com/@viswanarayanan.vijai/creationism-flat-earth-theory-geocentricity-and-anthropocentricity-are-the-major-false-e685880bde74
['Viswa Narayanan']
2020-12-12 18:34:08.579000+00:00
['Flat Earth', 'Technology', 'Evolution', 'AI', 'Philosophy']
1,935
Structural Scaffolds for Citation Intent Classification in Scientific Publications
Source: Cohan et al. (2019) Structural Scaffolds for Citation Intent Classification in Scientific Publications This post is a paper summary highlighting the main ideas of the paper “Structural Scaffolds for Citation Intent Classification in Scientific Publications” by Cohan et al. (2019) arXiv Github Introduction Machine reading and automated analysis of scientific literature have increasingly become important due to information overload. Citations are typically used to measure the impact of scientific publications (Li and Ho, 2008)[1]. Citation Intent Classification is the task of identifying why an author cited another paper. The automatic identification of citation intent could also help users in doing research. FIGURE 1 shows an example of two citation intents. Some citations indicate direct use of a method, while others may acknowledge prior work or compare methods or results. Existing models are based on hand-engineered features, which may not sufficiently model signals in the text (e.g. linguistic patterns or cue phrases). Recent advances in Natural Language Processing (NLP) have introduced large, contextual representations that are obtained from textual data without the need for manual feature engineering. Cohan et al. (2019)[2] introduce a novel framework to include structural knowledge into citations as well as a new dataset of citation intents: SciCite. Figure 1: Citation intent example. Source: Cohan et al. (2019) Dataset SciCite is five times larger, contains fewer but more general categories, and covers scientific literature from more general domains than existing datasets such as ACL-ARC (Jurgens et al., 2018)[3]. FIGURE 1 compares the datasets. The papers for SciCite were sampled from the Semantic Scholar corpus. The authors chose more general categories as some are very rare and would not have been enough for training. Citations were extracted using science-parse. The ACL-ARC dataset, which consists of Computational Linguistics papers, was annotated by domain experts in NLP. The training set for SciCite was crowdsourced using the Figure Eight platform, while the test set was annotated by an expert annotator. Table 1: SciCite vs ACL-ARC. Source: Cohan et al. (2019) Model The proposed neural multitask framework consists of a main task (Citation intent) and two structural scaffolds, or auxiliary tasks: section title and citation worthiness. The input x is the set of tokens in the citation context, which are encoded by concatenating non-contextual word representations GloVe (Pennington et al., 2014)[4] with contextualized embeddings ELMo (Peters et al., 2018)[5] (Eq. 1). Eq. 1 The encoded tokens then get fed into a bidirectional long short-term memory (Hochreiter and Schmidhuber, 1997)[6] network with hidden size d₂, which results in the contextual representation of each token w.r.t. the entire sequence (Eq. 2). Eq. 2 Finally, an attention mechanism is added, which produces a vector representation of the input sequence (Eq. 3). w is a parameter served as the query vector for dot-product attention. Eq. 3 The citation worthiness task is to predict whether a sentence needs a citation. The hypothesis is that language in sentences with citations is different from regular sentences in scientific work. Sentences with citations are positive samples and sentences without citation markers are negative samples. This task could also be used in a different setting (e.g. paper draft aid). The section title task is to predict the title of the section in which the citation appears. The hypothesis here is that citation intent is relevant to its section. Contrary to the other tasks, the authors use a large number of scientific papers to generate the training data for this task. In the multitask framework, a Multi-Layer Perceptron (MLP) followed by a softmax layer is used for each task (Eq. 4). The class with the highest probability is then chosen. Eq. 4 Figure 2 shows an overview of the proposed model. Figure 2: Model overview. Source: Cohan et al. (2019) Preprocessing For the citation worthiness task, citation markers were removed in order for the model to not “cheat” by simply recognizing citations in sentences. For sentence title, citations and their contexts were sampled from the corpus. Section titles were normalized with regular expressions to the following general categories: introduction, related work, method, and experiments. Titles that did not map to these were removed. The table below shows the total number of instances for each of the datasets and tasks. Experiments The proposed model, trained with standard hyperparameters, is compared to a strong baseline and a state-of-the-art model. The first baseline is a BiLSTM with an attention mechanism (with and without using ELMo embeddings) that only optimizes for the citation intent classification task. This is meant to show if the structural scaffolds and contextual embeddings, in fact, improve performance. The second baseline is the model used by Jurgens et al. (2018)[3], which had the best-reported results on ACL-ARC. The authors incorporate a diverse set of features (e.g. pattern-based, topic-based, prototypical argument) and train a Random Forest classifier. Results The results on both the ACL-ARC (FIGURE 2) and SciCite (FIGURE 3) datasets indicate that the inclusion of structural scaffolds improves performance on all of the baselines. The performance differences for the respective datasets is partly due to the different dataset sizes. Each auxiliary task contributes slightly over the baseline, while the combination of both tasks shows a large improvement on ACL-ARC and a marginal improvement on SciCite. The addition of contextual embeddings further increases performance by about 5% macro F1 on both datasets (including on the baselines). Table 2: Results on ACL-ARC. Source: Cohan et al. (2019) Table 3: Results on ACL-ARC. Source: Cohan et al. (2019) Figure 3 shows an example sentence from ACL-ARC for which the correct label is Future Work. The best-proposed model predicts this correctly, attending over more of the context, while the baseline predicts Compare. The attention is stronger on “compare” for the baseline, ignoring the context of its use. Figure 3: Example from ACL-ARC. Source: Cohan et al. (2019) When looking at each of the categories independently, categories with more instances show higher F1 scores on both datasets (FIGURES 4 and 5). Recall seems to suffer from a limited number of training instances. Table 4: Per category classification results on ACL-ARC. Source: Cohan et al. (2019) Because the categories in SciCite are more general, there are more training instances for each. The recall on this dataset is accordingly higher. Table 5: Per category classification results on SciCite. Source: Cohan et al. (2019) Conclusion Cohan et al. (2019)[2] show that structural properties of scientific literature can be useful for citation intent classification. The authors argue that relevant auxiliary tasks can help improve performance in multitask learning. The main contributions of this work are the following: (i) A new scaffold framework for citation intent classification. (ii) A new state-of-the-art of 67.9% F1 on ACL-ARC (an increase of 13.3%). (iii) A new dataset of citation intents: SciCite. Future Work While the current work uses ELMo, Beltagy et al. (2019)[7] show that incorporating BERT, a large pre-trained language model, which they fine-tuned on scientific data, increases performance further. A possible extension could be to adapt the model to other domains (e.g. Wikipedia). Further reading SciBERT: A Pretrained Language Model for Scientific Text — Beltagy et al., 2019 ScispaCy: Fast and Robust Models for Biomedical Natural Language Processing — Neumann et al., 2019 Original post here. References [1]: Zhi Li and Yuh-Shan Ho. 2008. Use of citation per publication as an indicator to evaluate contingent valuation research. Scientometrics. [2]: Arman Cohan, Waleed Ammar, Madeleine van Zuylen, and Field Cady. 2019. Structural scaffolds for citation intent classification in scientific publications. In NAACL-HLT, pages 3586–3596, Minneapolis, Minnesota. Association for Computational Linguistics. [3]: David Jurgens, Srijan Kumar, Raine Hoover, Dan McFarland, and Dan Jurafsky. 2018. Measuring the evolution of a scientific field through citation frames. TACL, 6:391–406. [4]: Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In EMNLP, pages 1532–1543. [5]: Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke S. Zettlemoyer. 2018. Deep contextualized word representations. In NAACL-HLT. [6]: Sepp Hochreiter and Jurgen Schmidhuber. 1997. Long short-term memory. Neural Computation. [7]: Iz Beltagy, Arman Cohan, and Kyle Lo. 2019. SciBERT: A Pretrained Language Model for Scientific Text. CoRR, abs/1903.10676.
https://medium.com/dair-ai/structural-scaffolds-for-citation-intent-classification-in-scientific-publications-e5acd2f0ebf9
['Tornike Tsereteli']
2020-11-11 08:56:55.535000+00:00
['Artificial Intelligence', 'Technology', 'Machine Learning', 'Deep Learning', 'NLP']
1,936
50 Smart gadgets under $50
Want to treat yourself with some of the best gadgets available at an affordable price? Presenting our weekly daily digest with 50 such amazing tech gadgets that you can get for less than or $50. From smart home tech to everyday carry gear, this list has it all to treat the gadget lover in you. Looking forward to doing some gadget shopping this weekend? We’ve rounded up our favorite gadgets under $50 in this blog. From everyday home office tech to innovative smart gadgets, this list has it all. Related: 23 Smart home gadgets to cover every corner of your home Ideally, if we could, we’d get all of them. After all, why wouldn’t we want to treat ourselves without burning holes in our pockets! GENKI ShadowCast Console Link Play your console games right on your laptop with the GENKI ShadowCast Console Link. This link works with Switch, PS4, PS5, and Xbox and is priced at under $40. Masta Box Fidget Controller Gadget Satisfy your need to fidget, in an office-appropriate way, with the Masta Box Fidget Controller Gadget. It costs $42 and is also a mouse, laser pointer, presenter, microSD card reader, and more. Orbitkey Fast USB 3.0 Chip An Orbitkey Fast USB 3.0 Chip is a quick way to store your files. It comes in 8 and 32 GB sizes, is slim enough to take anywhere, and you can get it for under $20. It’s a great item on our 50 smart gadgets under $50 list. Razer Seiren Mini Portable Microphone A Razer Seiren Mini Portable Microphone gives you all the features of a full-sized mic in a small form factor. If you’re looking for a great entry-level mic, this one’s the way to go for $50. Satechi USB-C Compact Wireless Charging Dock Power up your AirPods and AirPods Pro on your laptop with the Satechi USB-C Compact Wireless Charging Dock. With a five-watt delivery speed, your AirPods will be up and running in no time. The best part is, it only costs $29.99. Native Union Belt Watch Magnetic Apple Watch Charging Cable A Native Union Belt Watch Magnetic Apple Watch Charging Cable is a beautiful thing. This aesthetically-pleasing cord is durable and MFi certified by Apple, so it’ll work perfectly with your Apple Watch. Get it for $49.99. Flo by Moen Smart Water Detectors You won’t have to worry about waking up to a flooded basement when you have the Flo by Moen Smart Water Detectors. These gadgets attach to your water lines and send you an alert when it detects something wrong, all for $49.99. Satechi USB-C PD Audio Adapter With the Satechi USB-C PD Audio Adapter, you get the best of audio and charging. Its clear sound and USD-C PD charger make it easy to use anywhere, which is why we included it on this roundup of 50 smart gadgets under $50. XL Speaker Giant Earbud You’ll certainly make a statement with the XL Speaker Giant Earbud. This novelty speaker adds a cool vibe to your desk and packs serious sound for $39.99. Urbanista Sydney Pocket-Size Speaker The Urbanista Sydney Pocket-Size Speaker lets you get your groove on at the beach, in the park, or at the pool for a solid five hours. You’ll love the pretty colors: Rose Gold, Olive Green, White Mist, and Midnight Black, and the price, $49.99. Samsung Galaxy SmartTag & SmartTag+ Finders You won’t have to tear the house apart searching for lost items when you have Samsung Galaxy SmartTag & SmartTag+ Finders. These smart tags attach to pretty much anything, including your pup. The Nearby Find feature helps you locate what you’ve misplaced for only $29.99. Edimax EW-7438RPn Mini Compact Wi-Fi Range Extender An Edimax EW-7438RPn Mini Compact Wi-Fi Range Extender will help you say goodbye to spotty connection and dead zones. This wall-mounted Wi-Fi repeater offers up to 300 Mbps wireless speeds and only costs $29.99. NOCO GENIUS1 Smart Vehicle Charger You’ll want to have the NOCO GENIUS1 Smart Vehicle Charger by your side. It’s a battery charger, maintainer, and desulfator in one device. It’s a must-have car item on our 50 smart gadgets under $50 roundup. Clevermade LockBox Package Locker Make sure no one else gets to your Amazon deliveries before you do with the Clevermade LockBox Package Locker. The lid opens with hydraulics, and you can anchor the entire box to the wall, for as little as $50. WHOOP Strap 3.0 Activity Tracker Wearable A WHOOP Strap 3.0 Activity Tracker Wearable gives you all the stats you need about your health and fitness levels. For $30, you can use it to track your sleep, training intensity, and recovery time. Dissim Inverted Lighter Then, there’s the Dissim Invert Lighter. This handy gadget for $48 can help you ignite things without burning your finger. Its angled port lets you reach even low-wick candles. Satechi USB-C iPhone 12 Magnetic Wireless Charging Cable Make it even easier to wirelessly charge your iPhone Pro with the Satechi USB-C iPhone 12 Magnetic Wireless Charging Cable. It looks stylish, snaps to your phone with magnets, and provides a 7.5-watt charge. Get it for $29.99. Swidget Smart home Plugin Inserts The Swidget Smart home Plugin Inserts turn your ‘dumb’ appliances into smart ones for under $20. It’s a pretty good deal and is a great way to make your home smarter. Zens Wireless Charger Station A Zens Wireless Charger Station lets you build your own wireless charging station for under $50. This series of single chargers connect together so that you can create the ideal charging setup. Kangaroo Motion + Entry Sensor Home Security Device Another must-have item on our 50 smart gadgets under $50 roundup is the Kangaroo Motion + Entry Sensor Home Security Device. Just set it up at a point of entry, and you’ll be notified whenever someone comes in. Priced at under $30, it’s a steal. Wyze Watch Keep an eye on your blood oxygen levels, heart, rate, sleep duration, and more with the Wyze Watch. It’s one of the most affordable smartwatches out there since it costs less than $20. You’ll want this item on our roundup of 50 smart gadgets under $50. SteelSeries QcK Prism Cloth Medium Gaming Mouse A SteelSeries QcK Prism Cloth Medium Gaming Mouse will impress the gamer in your life without blowing a hole in your budget. Get its brilliant two-zone RGB illumination for under $50. Ring Mailbox Secure Sensor You have sensors for your front door, so why not for your mailbox? The Ring Mailbox Secure Sensor sends you notifications anytime it detects motion. It’s a steal at $29.99. Wyze Thermostat Smart Heating & Cooling System It’s super simple to control your home’s heating and cooling with the Wyze Thermostat Smart Heating & Cooling System. It works with voice assistants like Alexa and will only set you back $50. Lenovo Smart Clock Essential Voice-Assistant Clock Sure, the Lenovo Smart Clock Essential Voice-Assistant Clock tells the time, but it can also check your agenda, manage your smart home, and play music. It can also charge your phone, and thanks to voice assistance, this clock is easy to use. Snag it for $49.99. Wyze Video Doorbell Yes, you should be able to see who’s at your front door before you open it, and the Wyze Video Doorbell can help. The full HD display and 3:4 aspect ratio let you see everything clearly, too. Preorder it from Wyze for $29.99. Tapplock Lite Fingerprint Padlock Your suitcases and bags will be a lot safer with the Tapplock Lite Fingerprint Padlock. It unlocks only with your fingerprint and opens in a snappy 0.8 seconds. It’s a great gadget for safety on our list of 50 smart gadgets under $50. Nanoleaf Essentials Smart Lighting Basics A Nanoleaf Essentials Smart Lighting Basics is a surefire way to amp up the fun and enhance the lighting in your house. The collection includes a bulb and two different light strips. Each gives you 16 million different colors and tunable white hues. Belkin USB-C PD GaN Wall Charger Need a quick way to charge your devices? Try the Belkin USB-C PD GaN Wall Charger. It uses the power of GaN and provides a whopping 60 watts of power. Get it for $49.99. 8Bitdo TG16 2.4G Wireless Gamepad An 8Bitdo TG16 2.4G Wireless Gamepad will help you zoom past the competition. This completely lag-free gamepad works with your Switch, CoreCrafx, and PC Engine Mini. It’s a sweet deal at $24.99. almost object Stretch Clock by Nicolas Baker Looking for one-of-a-kind art for your home, office, or home office? The almost object Stretch Clock by Nicolas Baker is it. It’s a stretchy clock that you can place around any object, letting you create a timepiece anywhere for only $35. Microsoft Number Pad Keyboard Accessory A Microsoft Number Pad Keyboard Accessory is another useful product on our list of 50 smart gadgets under $50. It lets you type numbers faster and connects to your devices via Bluetooth. Chipolo ONE Ocean Edition Key Finder And the Chipolo ONE Ocean Edition Key Finder helps you find your misplaced items in an eco-friendly design. Get it for $29. Microsoft Surface Mobile Mouse If you need a good, affordable mouse, go for the Microsoft Surface Mobile Mouse. It has a one-year battery life, streamlined design, and is priced at $34.99. And that’s why it made our 50 smart gadgets under $50. Beats Flex All-Day Wireless Earphones The Beats Flex All-Day Wireless Earphones let you dance to your favorite beats all day. That’s thanks to their 12 hours of listening time. You can also expect clear calls and advanced acoustics. They go for $49.99. Belkin MagSafe Car Vent Mount PRO for iPhone 12 It’s way easier to use your iPhone 12 in the car with the Belkin MagSafe Car Vent Mount PRO for iPhone 12. This car mount has a powerful magnetic module that keeps your 2020 iPhone from falling off the dashboard. Get it for $39.99. Cable Matters USB 3.1 Ethernet Switch 4-Port Gigabit Internet Adapter A Cable Matters USB 3.1 Ethernet Switch 4-Port Gigabit Internet Adapter helps your gadgets and peripherals work faster. It connects up to four devices and even works with your Nintendo Switch. It costs $27.99. ASTRO A03 In-Ear Monitors Gaming Earbuds The ASTRO A03 In-Ear Monitors Gaming Earbuds prove that you can totally get a great pair of gaming earbuds for under $50. They give you the immersive sound you crave for gaming, music, and even work. Wyze Headphones And the Wyze Headphones show that it’s possible to get noise-canceling headphones for under $50. This set gives you 20 hours of battery life, voice-isolating microphones, and four-noise canceling microphones. Lexon Flilp+ LCD Alarm Clock Morning wakeups are more manageable with the Lexon Flilp+ LCD Alarm Clock. It features a reversible display, so it’ll continue to show the time even upside down. You can also use it as a nightlight, and it goes for $39.00. Petcube Cam Pet Monitoring Camera The Petcube Cam Pet Monitoring Camera makes visits to the vet easier for just $39.99. Its virtual vet feature ensures your pet gets timely checkups. This is one cool pet gadget on our roundup of 50 smart gadgets under $50. Amazon Echo Dot 4th-Generation Smart Speaker Of course, you have a new smart speaker on your wishlist. And the Amazon Echo Dot 4th-Generation Smart Speaker only costs $49.99. This one boasts a small form factor and Amazon’s classic fabric covering. Philips Hue E12 Candelabra Bluetooth Bulb The flame-shaped Philips Hue E12 Candelabra Bluetooth Bulb is a smart addition to any home. It works over Bluetooth, and you can control it via Alexa or Google assistant. It only sets you back about $24.99. VOAMOKO USB-C Numeric Keypad Bluetooth Numpad With the VOAMOKO USB-C Numeric Keypad Bluetooth Numpad, you’ll type those numbers faster than ever. Its light weight and compact size make it easy to store in your laptop bag. Buy it for $48.99. AmazonBasics Humidifier Aroma Diffuser Dry air in your home is no fun. You can combat it with the AmazonBasics Humidifier Aroma Diffuser. This device adds beneficial moisture to the air and, if you choose, relaxing scents like lavender and jasmine. It’s a steal at $31.99. Satechi Aluminum Wireless Multimedia Remote Control A Satechi Aluminum Wireless Multimedia Remote Control streamlines the process of operating your Bluetooth devices. From your laptop to your home theater, all actions are just a button away, for only $39.99. JBL Go 3 Compact Speaker It’s nice to hear some tunes while you’re at the beach, and that’s why the JBL Go 3 Compact Speaker is such a great product on our list of 50 smart gadgets under $50. It’s waterproof, looks great, and is easily portable. Genki Audio Lite Slim Bluetooth Adapter A Genki Audio Lite Slim Bluetooth Adapter gives your Nintendo Switch wireless audio. So you can enjoy immersive sound as you fish in the ocean while you play Animal Crossings. This gadget costs $34.99, and it works with all Bluetooth headphones and speakers. Tractive Pet GPS Tracker Keep tabs on Fluffy, at a price point that’ll make you smile, with the Tractive Pet GPS Tracker. This LTE GPS tracker costs $49.99 and has international coverage, so you can see where your furry friend is from anywhere in the world. elago AirPods Charging Station Silicone Stand Keep your AirPods protected while they charge with the elago AirPods Charging Station Silicone Stand. With hard-wearing silicone, it gives you all-over protection and costs just $10.99. Well, there you have it, our list of 50 smart gadgets under $50. No matter your budget, you’re sure to have found a great smart gadget or two on this list that can make life so much easier and using your gadgets more seamless. Let us know which of these devices you like the most in the comments. Want more tech news, reviews, and guides from Gadget Flow? Follow us on Google News, Feedly, and Flipboard. If you’re using Flipboard, you should definitely check out our Curated Stories. We publish three new stories every day, so make sure to follow us to stay updated! The Gadget Flow Daily Digest highlights and explores the latest in tech trends to keep you informed. Want it straight to your inbox? Subscribe ➜
https://medium.com/the-gadget-flow/50-smart-gadgets-under-50-4268cf93211a
['Gadget Flow']
2021-02-13 18:52:10.407000+00:00
['Tech', 'Internet of Things', 'Gadgets', 'Smart Home', 'Technology']
1,937
3 Reasons Why People Fail CompTIA Exams
Take a CompTIA exam, they said. It will be fun, they said. Photo by Nick Morrison on Unsplash Studying the Wrong Material You do no need to spend thousands of dollars or spend months in a formal class to pass CompTIA exams. Between the end of April 2019 and the first week of December in 2019 I studied for and passed the A+, Network+, and Security+ while also working on a master’s degree. The only two sources relevant to pass are CompTIA’s Certmaster Practice (have to pay) and ProfessorMesser.com’s free study videos. That is all I used and have had plenty of students pass using only those two sources. Please do not waste your time and money on secondary sources that are not straight from CompTIA that charge you an arm and a leg for the privilege of learning their interpretation of what is on the exam. CompTIA does not care what they say, so you should not either. The point is to pass the certification not to read a novel about technology. You can do that on your own time. CompTIA does a very good job replicating the style of questions asked on the real exam with their own material and practice exams. Secondary sources created by other people will never be able to get as close to the real exam as CompTIA. Half of the battle is studying the way CompTIA phrases the questions because it is truly confusing if you do not know exactly what they are asking.
https://medium.com/an-idea/3-reasons-why-people-fail-comptia-exams-59e02d8b4a75
['Tyler Harris']
2020-10-12 22:54:46.521000+00:00
['Tech', 'Comptia', 'Comptia Certifications', 'Technology', 'Education']
1,938
MJAC — Crypto Compare Blockchain Summit, London. 30th November 2018
MJAC announces added speakers, ICOs and Power Snooker Exhibition Games at blockchain conference - Showcasing innovators and thought leaders of the blockchain industry alongside sporting personalities - Obi Nwosu, CEO of Coinfloor, and Nick Chong of Liquid by Quoine will be delivering the two keynote speeches at the MJAC & CryptoCompare London Blockchain Summit. The event, which is taking place on November 30th 2018, will see a host of key industry players discussing and debating the hot topics from within the blockchain and cryptocurrency spheres. Claire Wells and Marieke Flament from global crypto finance company Circle will be having a Fireside Chat where they will discuss Opportunities and Threats for the Tokenisation of Everything. Iqbal V. Gandham of eToro will be asking: Will STOs replace IPOs? during his address and Steve Swain, CEO of Lendingblock, will be on stage discussing how securities lending will lead to the institutional adoption of cryptocurrencies. David Fauchier, CEO/CIO at Cambrial, will be discussing: The state of crypto funds: challenges, opportunities, what next? and Saar Levi, CEDEX CEO will be looking at the tokenization of new assets classes and what this means for the financial markets. The schedule includes a number of panel discussions. Edd Carlton of BlockEx will be joined by Marina Titova from NKB Group, Amir Ness of Elevate Group and Joseph-Daniel Millwood of Coinbase to discuss: Crypto and how it has evolved as an asset class over 2018. The Regulatory Panel Discussion, moderated by Teana Baker-Taylor, Advisory Council at Global Digital Finance, comprises of Dan Morgan from Ripple, Bryony Widdup of DLA Piper and Ruth Wandhöfer of Coinfirm. The Institutional Trading in the Crypto Market panel includes moderator Simon Taylor of 11:FS who is joined by Ryan Radloff, CEO of CoinShares; Hansen Wang of Melonport, Lucas Friss from Cumberland and Martin Hoffgen of Elation Capital. The ICO section of the agenda will see presentations from names including: Gigzi, Creative 3D Web and PowerSnooker; the latter being presented by Dean Christy, PowerSnooker Ambassador and Former Head of Corporate Governance at Prudential. Additionally, MJAC will host a global selection of eminent and emerging companies as exhibitors. Delegates will have an opportunity to enter a PowerSnooker prize draw with six lucky winners getting to play against snooker legend Tony Knowles and boxing superstar Johnny Nelson on a full-sized professional snooker table at the event. This exhibition game will be played during the drinks reception, which will start at 18.00. Confirmed speakers: Claire Wells — Legal & Business Affairs Director for Europe at Circle Marieke Flament — Global CMO and EMEA Managing Director at Circle Saar Levi — CEDEX CEO and co-founder Amir Ness — Founder of Elevate Group Teana Baker-Taylor — Advisory Council at Global Digital Finance Obi Nwosu, CEO and co-founder Coinfloor Ltd Lucas Friss — Head of Business Development, Europe Cumberland Marina Titova — Head of ICO Advisory at NKB Group Edd Carlton — Head of OTC Trading at BlockEx Simon Taylor — Co-founder and Blockchain Practice Lead at 11:FS Joseph-Daniel Millwood — EU Marketing Lead at Coinbase Cassius Kiani — Founder of Atlas Neue Dan Morgan — Head of Regulatory Relations, Europe at Ripple Steve Swain — Co-founder & CEO of Lendingblock Iqbal V. Gandham — Managing Director at eToro (UK Region) Hansen Wang — Head of External Relations at Melonport Martin Hoffgen — Director at Elation Capital Ryan Radloff — CEO at CoinShares Ruth Wandhöfer — Head of Industry and Regulatory Affairs at Coinfirm Dean Christy — PowerSnooker Ambassador David Fauchier — CEO/CIO Cambrial Bryony Widdup, Partner at DLA Piper Nick Chong — Head of North America for Liquid by Quoine “This summit, with its abundance of high calibre participants, will help people keep abreast of what is happening in the rapidly evolving crypto and blockchain space,” said Matthew Collom of MJAC. “Delegates can expect to see both a retail and institutional focus and witness the crucial issues being hotly debated.” The MJAC & CryptoCompare London Blockchain Summit is taking place at Old Billingsgate, 1 Old Billingsgate Walk, London, EC3R 6DX on Friday, November 30th, 2018. For the latest details and to book tickets go to http://www.mjac.io. Use code: MJAC30 for a 30% discount. To find out about exhibitor, ICO packages and sponsorship opportunities call: 07980 868 676 or email: blockchain@advfn.com. To enquire about press passes email: francescad@advfnplc.com About MJAC Disruptive Investment Conferences MJAC InvestorsHub Disruptive Investment Conferences are conferences and expos focusing on disruptive technologies including: blockchain, lithium and medical marijuana. Next event: MJAC & CryptoCompare London Blockchain Summit, London, November 30th, 2018. About CryptoCompare At CryptoCompare we pride ourselves in showing the most accurate live prices, charting and market analysis from 65 of the top crypto exchanges globally. Our content editors keep the mining data, be it company, mining equipment or mining contracts up to date with the latest pricing and products available on the market. Our wallets section has a comprehensive list of wallets that will help you find the best wallet for your needs. We have over 217,663 trading pairs and over 4,688 crypto currencies or crypto assets in an easy to understand and intuitive format. View all posts by btcbros
https://medium.com/btcbros/mjac-crypto-compare-blockchain-summit-london-30th-november-2018-cdc21af6b7a9
['Btc Bros']
2018-11-19 11:11:08.627000+00:00
['Cryptocurrency Investment', 'Bitcoin', 'Blockchain', 'Blockchain Technology', 'Cryptocurrency']
1,939
[S5E1] The Good Fight Series 5 Episode 1 (Full Episode)
➕Official Partners “TVs” TV Shows & Movies ● Watch The Good Fight Season 5 Episode 1 Eng Sub ● The Good Fight Season 5 Episode 1 : Full Series ஜ ۩۞۩ ஜ▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭ஜ ۩۞۩ ஜ ஜ ۩۞۩ ஜ▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭▭ஜ ۩۞۩ ஜ The Good Fight — Season 5, Episode 1 || FULL EPISODES : Picking up one year after the events of the final broadcast episode of “The Good Wife”, an enormous financial scam has destroyed the reputation of a young lawyer, Maia Rindell, while simultaneously wiping out her mentor and godmother Diane Lockhart’s savings. Forced out of her law firm, now called “Lockhart, Deckler, Gussman, Lee, Lyman, Gilbert, Lurie, Kagan, Tannebaum & Associates”, they join Lucca Quinn at one of Chicago’s preeminent law firms. . The Good Fight 5x1 > The Good Fight S5xE1 > The Good Fight S5E1 > The Good Fight TVs > The Good Fight Cast > The Good Fight Online > The Good Fight Eps.5 > The Good Fight Season 5 > The Good Fight Episode 1 > The Good Fight Premiere > The Good Fight New Season > The Good Fight Full Episodes > The Good Fight Season 5 Episode 1 > Watch The Good Fight Season 5 Episode 1 Online Streaming The Good Fight Season 5 :: Episode 1 S5E1 ► ((Episode 1 : Full Series)) Full Episodes ●Exclusively● On TVs, Online Free TV Shows & TV The Good Fight ➤ Let’s go to watch the latest episodes of your favourite The Good Fight. ❖ P.L.A.Y ► https://cutt.ly/En7VYtV The Good Fight 5x1 The Good Fight S5E1 The Good Fight TVs The Good Fight Cast The Good Fight Online The Good Fight Eps.5 The Good Fight Season 5 The Good Fight Episode 1 The Good Fight Premiere The Good Fight New Season The Good Fight Full Episodes The Good Fight Watch Online The Good Fight Season 5 Episode 1 Watch The Good Fight Season 5 Episode 1 Online ⭐A Target Package is short for Target Package of Information. It is a more specialized case of Intel Package of Information or Intel Package. ✌ THE STORY ✌ Its and Jeremy Camp (K.J. Apa) is a and aspiring musician who like only to honor his God through the energy of music. Leaving his Indiana home for the warmer climate of California and a college or university education, Jeremy soon comes Bookmark this site across one Melissa Heing (Britt Robertson), a fellow university student that he takes notices in the audience at an area concert. Bookmark this site Falling for cupid’s arrow immediately, he introduces himself to her and quickly discovers that she is drawn to him too. However, Melissa holds back from forming a budding relationship as she fears it`ll create an awkward situation between Jeremy and their mutual friend, Jean-Luc (Nathan Parson), a fellow musician and who also has feeling for Melissa. Still, Jeremy is relentless in his quest for her until they eventually end up in a loving dating relationship. However, their youthful courtship Bookmark this sitewith the other person comes to a halt when life-threating news of Melissa having cancer takes center stage. The diagnosis does nothing to deter Jeremey’s love on her behalf and the couple eventually marries shortly thereafter. Howsoever, they soon find themselves walking an excellent line between a life together and suffering by her Bookmark this siteillness; with Jeremy questioning his faith in music, himself, and with God himself. ✌ STREAMING MEDIA ✌ Streaming media is multimedia that is constantly received by and presented to an end-user while being delivered by a provider. The verb to stream refers to the procedure of delivering or obtaining media this way.[clarification needed] Streaming identifies the delivery approach to the medium, rather than the medium itself. Distinguishing delivery method from the media distributed applies especially to telecommunications networks, as almost all of the delivery systems are either inherently streaming (e.g. radio, television, streaming apps) or inherently non-streaming (e.g. books, video cassettes, audio tracks CDs). There are challenges with streaming content on the web. For instance, users whose Internet connection lacks sufficient bandwidth may experience stops, lags, or slow buffering of this content. And users lacking compatible hardware or software systems may be unable to stream certain content. Streaming is an alternative to file downloading, an activity in which the end-user obtains the entire file for the content before watching or listening to it. Through streaming, an end-user may use their media player to get started on playing digital video or digital sound content before the complete file has been transmitted. The term “streaming media” can connect with media other than video and audio, such as for example live closed captioning, ticker tape, and real-time text, which are considered “streaming text”. This brings me around to discussing us, a film release of the Christian religio us faith-based . As almost customary, Hollywood usually generates two (maybe three) films of this variety movies within their yearly theatrical release lineup, with the releases usually being around spring us and / or fall respectfully. I didn’t hear much when this movie was initially aounced (probably got buried underneath all of the popular movies news on the newsfeed). My first actual glimpse of the movie was when the film’s movie trailer premiered, which looked somewhat interesting if you ask me. Yes, it looked the movie was goa be the typical “faith-based” vibe, but it was going to be directed by the Erwin Brothers, who directed I COULD Only Imagine (a film that I did so like). Plus, the trailer for I Still Believe premiered for quite some us, so I continued seeing it most of us when I visited my local cinema. You can sort of say that it was a bit “engrained in my brain”. Thus, I was a lttle bit keen on seeing it. Fortunately, I was able to see it before the COVID-9 outbreak closed the movie theaters down (saw it during its opening night), but, because of work scheduling, I haven’t had the us to do my review for it…. as yet. And what did I think of it? Well, it was pretty “meh”. While its heart is certainly in the proper place and quite sincere, us is a little too preachy and unbalanced within its narrative execution and character developments. The religious message is plainly there, but takes way too many detours and not focusing on certain aspects that weigh the feature’s presentation. ✌ TELEVISION SHOW AND HISTORY ✌ A tv set show (often simply Television show) is any content prBookmark this siteoduced for broadcast via over-the-air, satellite, cable, or internet and typically viewed on a television set set, excluding breaking news, advertisements, or trailers that are usually placed between shows. Tv shows are most often scheduled well ahead of The War with Grandpa and appearance on electronic guides or other TV listings. A television show may also be called a tv set program (British EnBookmark this siteglish: programme), especially if it lacks a narrative structure. A tv set Movies is The War with Grandpaually released in episodes that follow a narrative, and so are The War with Grandpaually split into seasons (The War with Grandpa and Canada) or Movies (UK) — yearly or semiaual sets of new episodes. A show with a restricted number of episodes could be called a miniMBookmark this siteovies, serial, or limited Movies. A one-The War with Grandpa show may be called a “special”. A television film (“made-for-TV movie” or “televisioBookmark this siten movie”) is a film that is initially broadcast on television set rather than released in theaters or direct-to-video. Television shows may very well be Bookmark this sitehey are broadcast in real The War with Grandpa (live), be recorded on home video or an electronic video recorder for later viewing, or be looked at on demand via a set-top box or streameBookmark this sited on the internet. The first television set shows were experimental, sporadic broadcasts viewable only within an extremely short range from the broadcast tower starting in the. Televised events such as the 2020 Summer OlyBookmark this sitempics in Germany, the 2020 coronation of King George VI in the UK, and David Sarnoff’s famoThe War with Grandpa introduction at the 9 New York World’s Fair in the The War with Grandpa spurreBookmark this sited a rise in the medium, but World War II put a halt to development until after the war. The 2020 World Movies inspired many Americans to buy their first tv set and in 2020, the favorite radio show Texaco Star Theater made the move and became the first weekly televised variety show, earning host Milton Berle the name “Mr Television” and demonstrating that the medium was a well balanced, modern form of entertainment which could attract advertisers. The firsBookmBookmark this siteark this sitet national live tv broadcast in the The War with Grandpa took place on September 5, 2020 when President Harry Truman’s speech at the Japanese Peace Treaty Conference in SAN FRAThe Good Fight CO BAY AREA was transmitted over AT&T’s transcontinental cable and microwave radio relay system to broadcast stations in local markets. ✌ FINAL THOUGHTS ✌ The power of faith, love, and affinity for take center stage in Jeremy Camp’s life story in the movie I Still Believe. Directors Andrew and Jon Erwin (the Erwin Brothers) examine the life span and The War with Grandpas of Jeremy Camp’s life story; pin-pointing his early life along with his relationship Melissa Heing because they battle hardships and their enduring love for one another through difficult. While the movie’s intent and thematic message of a person’s faith through troublen is indeed palpable plus the likeable mThe War with Grandpaical performances, the film certainly strules to look for a cinematic footing in its execution, including a sluish pace, fragmented pieces, predicable plot beats, too preachy / cheesy dialogue moments, over utilized religion overtones, and mismanagement of many of its secondary /supporting characters. If you ask me, this movie was somewhere between okay and “meh”. It had been definitely a Christian faith-based movie endeavor Bookmark this web site (from begin to finish) and definitely had its moments, nonetheless it failed to resonate with me; struling to locate a proper balance in its undertaking. Personally, regardless of the story, it could’ve been better. My recommendation for this movie is an “iffy choice” at best as some should (nothing wrong with that), while others will not and dismiss it altogether. Whatever your stance on religion faith-based flicks, stands as more of a cautionary tale of sorts; demonstrating how a poignant and heartfelt story of real-life drama could be problematic when translating it to a cinematic endeavor. For me personally, I believe in Jeremy Camp’s story / message, but not so much the feature. FIND US: ✔️ https://cutt.ly/En7VYtV ✔️ Instagram: https://instagram.com ✔️ Twitter: https://twitter.com ✔️ Facebook: https://www.facebook.com
https://medium.com/@the-good-fight-s05e01-ep-1/s5e1-the-good-fight-series-5-episode-1-full-episode-efc3fe9fd4ad
['The Good Fight', 'Previously On']
2021-06-24 07:30:35.871000+00:00
['Politics', 'Technology', 'Covid 19']
1,940
The quest for quantum energies in computing
The beginning of 2019 was marked with IBM announcing its first commercial quantum computer at the 2019 Consumer Electronics Show (CES). Even though experts are quite skeptical about its power and viability, it is a big step forward in the computer building technology. In this post we’ll discuss why it is so important and what we can expect from it. What is a quantum computer? A quantum computer is a computation device that makes direct use of quantum mechanical phenomena, such as superposition and entanglement, to perform operations on data. Quantum computers differ from traditional computers based on transistors and more innovative computing technologies such as DNA, optical, or biocomputers. How does it work? Quantum computers take advantage of the ability of subatomic particles to exist in more than one state at any time. This ability allows operations to be done much more quickly and use less energy than classical computers. While in classical computing, a single piece of information is a bit existing in two states — 1 or 0, quantum computing stores information in ‘qubits’. These quantum systems also have two states, but they can storage capacity is not limited to 1 or 0, because they can exist in any superposition of these values. To manipulate information, quantum computers leverage different physical phenomena — superposition, entanglement, and interference. Superposition is a combination of states we would ordinarily describe independently. In our everyday life, a superimposition will be the sound of two notes played at once. In quantum superposition, the difference is what happens when you perform a measurement. Even though the system exists in a perfectly defined superposition state, when you perform certain measurements on it, you get random outcomes. So, the magic is observed as a special kind of quantum randomness. Entanglement is a famously counter-intuitive quantum phenomenon that we never see in the classical world. Entangled particles behave together as a system in ways that cannot be explained using classical logic. In the simplest terms, they cannot be described independently of each other. If we measure one of them, we might observe that it behaved randomly, but it tells us what to expect when measuring the other particle, in the same way. Moreover, this perfect correlation remains true even if you measure entangled particles at opposite ends of the galaxy, so it is crucial for speeding up computation. Interference is another challenging principle of quantum theory. When particles exhibit interference, they are behaving like waves: when two waves are in phase, their amplitudes add, and when they are out of phase, their amplitudes cancel. Quantum algorithms Just like in classical computing, you need an algorithms — a set of instructions that are supposed to solve the problem and you need a machine to execute these instructions. The fact that quantum computers can create superpositions, entanglement, and other quantum effects means that we have to write algorithms in a new way incorporating these phenomena. We suggest you to watch this explanation by IBM researcher David Gosset to conceptualize it faster and more effectively: What are the benefits? At present, there exist challenges above a certain size and complexity that we cannot solve since we don’t have enough computational power on Earth to tackle them. For them we need a new kind of computing: one whose computational power also scales exponentially as the system size grows. If large-scale quantum computers can be built, they will solve certain problems exponentially faster than any of our current computers. For example, integer factorization is believed to be computationally infeasible with for large integers that are the product of only a few prime numbers. Contrary to classical computers, a quantum computer could solve this problem efficiently using Shor’s algorithm to find its factors. At present, it is the difficulty of integer factorizations that is the basis of most of public key ciphers. Therefore, a quantum computer will potentially “break” many of the current cryptographic systems. Shor’s algorithm From a less sinister perspective, quantum computers can be used to model complicated chemical reactions. Since Google and IBM have already presented quantum devices to simulate a hydrogen molecule and then even more complex molecules, there is hope that quantum simulations will eventually be used to design entirely new molecules for use in medicine. Furthermore, the Holy Grail for quantum chemists is to model the Haber-Bosch process — a way of artificially producing ammonia. Do we have a winner? For quite a time it was believed that Google was leading the race. In a Nature article published in March 2017, it revealed ambitious plans to commercialize quantum technology in the next five years. At the same time, Google intended to achieve the ‘quantum supremacy’ — the point where a quantum computer can crunch sums that a classical computers fail hopelessly — with a 49-qubit computer by the end of 2017. In November 2017, however, Google’s leadership was shaken by IBM that announced it had built a 50-qubit quantum computer. Yet, it was far from stable, and the system could only hold its quantum microstate for 90 microseconds, far from the times needed to make quantum computing viable. However, since 2016, long before creating a standalone computer, IBM has been offering a free and publicly available IBM Q Experience that currently has over 100,000 users, who have already run more than 6.7 million experiments resulting in publishing more than 130 research papers. The IBM Q Network has recently included such additions as Argonne National Laboratory, CERN, ExxonMobil, Fermilab, and Lawrence Berkeley National Laboratory. Similar efforts were taken by such startups as Rigetti focusing on the stability of its own systems and the Canadian company D-Wave that created what it was calling a 2,000-qubit system. Finally, this January IBM seems to win the race with launching, Q System One, the first standalone commercial quantum computer. But is it a real victory? Even though IBM describes it as “the first fully integrated universal quantum computing system designed for scientific and commercial use”, it is merely a 20-qubit machine that is not powerful enough for most of the commercial applications. IBM itself stresses that this is a first attempt and that the systems are “designed to one day tackle problems that are currently seen as too complex and exponential in nature for classical systems to handle.” However, the most important features of it are stability, repeatability and flexibility, and if the underlying architecture of the Q System One is scalable, new processors can be popped in when they’re developed without new engineering challenges. It is true that right now the Q System One is the quantum equivalent of the room-sized computers of the 1950s. And of course, it is true that getting quantum computing into the hands of consumers will require a big breakthrough to handle high temperatures, at the very least. Still, quantum computers can one day revolutionize many disciplines, such as materials and drug discovery, AI and complex systems optimization, and the present-day IBM announcement is a big step from the fantasy world to our reality.
https://medium.com/sciforce/the-quest-for-quantum-energies-in-computing-ac40d44117c9
[]
2019-02-01 16:53:37.903000+00:00
['Science', 'Quantum Computing', 'Computer Science', 'Technology', 'Algorithms']
1,941
Can computer kill
I read a story today. A primary school headmaster killed himself by burning himself with petrol. The reason is his inability to cope with pressure of using computer which were mandated by his higher officials. The goal of using technology is to help people, but why is the computers are causing lot of stress which are leading to sucides. Imagine you are in your fifties and doing a job like teaching kids for most of your life and there is a great pressure to digitize everything. The amount of experience you have is not going to save you. The accounts or auditing details which were used to write in note books might be going to be online. There is pressure from higher officials on the details not present in the system. The trust deficiency in Indians are taking shape to this kind of apps using the computers and helding people accountable for not using it. We can easily say one has to skill up in this modern world. Is the fault always lies in the people who were not able to use technology? Can the people who design this systems can do better job by taking into consideration the user who don’t know how to use computer? I have exprerinced lot of stress while visiting my relatives place which is adjacent to my house. The watchmen of apartments are asking to install a app and show a pass code when I am going to visit my relatives house. Even I visit there 10 times a day, they ask me to show passcode. Imagine the kind of stress it adds for people who are not using smart phones and treated as thugs who are trying to enter a place which are forbidden. There is nothing wrong with using technology. But my conern how the technologies occupied central place instread of solving problems. I think technology is used as watch dogs to control by apartment owners and government officials. I think the reason is not everyone is taken into account when designing a system. A process should not be biased for people who know how to use computer. Technologies should make life simple for everyone but not include only young and discounting the old. And more importantly technologies should not be forced on people. https://www.eenadu.net/crime/newsarticle/general/0302/120153294 The article is in telugu. You can use the google translate if you are not familiar with telugu.
https://medium.com/@visu-rohini/can-computer-kill-be141f914494
['Vishnu Sandhi']
2020-12-13 10:01:43.901000+00:00
['Technology', 'Teachers', 'Distress', 'Suicide', 'Andhra Pradesh']
1,942
Emulating the move functionality in S3 using Spark
I was recently working on a scenario where I had to move files between buckets using Spark. Now, Spark does not have native support for S3 but uses the Hadoop FileSystem API to treat S3 as a distributed FileSystem. The API is pretty good and can handle most of the use cases of S3 from Spark pretty well, with many methods provided to solve frequent use cases. But surprisingly there is not direct method to move files from one bucket to another. It has to be done using the copy() function provided and has some intricacies which must be kept in mind before starting the movement across buckets. The difference between move and copy is that move transfers the data from source to destination while copy keeps the original data at the source and only makes an additional copy at the destination. The first thing to remember in Spark while moving files is the configuration that must be done for the FileSystem. The FileSystem has to be declared as s3native and the Access keys and Secret keys for AWS have to be entered. The “mapreduce.fileoutputcommitter.algorithm.version” parameter is optional, but it highly recommended to use this as it speeds up the write process from Spark to S3. Before this parameter, Spark used to write the data from executor to the driver and then from the driver to S3. This parameter bypasses the transfer to driver and directly writes from executors to S3, increasing parallelism many folds and consequently the speed of writing. sparkSession.sparkContext.hadoopConfiguration.set("fs.s3a.impl", "org.apache.hadoop.fs.s3native.NativeS3FileSystem") sparkSession.sparkContext.hadoopConfiguration.set("fs.s3a.awsAccessKeyId", "XXXXXXXXXXX") sparkSession.sparkContext.hadoopConfiguration.set("fs.s3a.awsSecretAccessKey","XXXXXXXXXXX") sparkSession.sparkContext.hadoopConfiguration.set("mapreduce.fileoutputcommitter.algorithm.version", "2") Once the configuration has been set, it is time to code the process. Here each bucket of S3 is treated as a separate FileSystem and thus the FileSystems to each bucket have to be created separately. An example of rawPath could be “s3://test-bucket/data”. The FileSystem.get() method expects the raw path and the configuration which we declared above. A thing to remember while using the copy() method is that it copies one file at a time. Hence you have to give the full source path of a file and the full destination path of the file you want to move. This can be done by listing the files in the source FileSystem, iterating over them and storing their paths in a List. Now it’s trivial to set the destination path by iterating over the paths above and replacing the source bucket name with the destination bucket name. This ensures that the destination bucket has the same directory structure as the source bucket. Finally, the copy() method is called for each source path and the relevant parameters are passed to it as shown below. Even though the parameters given to copy() are pretty clear, I’ll explain what the ‘deleteSource’ and ‘overwrite’ parameters are for. ‘deleteSource’ deletes the files from the source bucket once the files have been copied and ‘overwrite’ replaces any previous file with the same exact path in the destination bucket. def copyS3DataBetweenBuckets(sparkSession: SparkSession, rawPath: String, processingPath: String, deleteSource : Boolean, overwrite : Boolean): Unit = { import org.apache.hadoop.fs.{FileSystem, Path} val srcPath: Path = new Path(rawPath) val srcFs = FileSystem.get( srcPath.toUri, sparkSession.sparkContext.hadoopConfiguration) val dstPath: Path = new Path(processingPath) val dstFs = FileSystem.get( dstPath.toUri, sparkSession.sparkContext.hadoopConfiguration) val paths: ListBuffer[Path] = ListBuffer() val fileStatusListIterator: RemoteIterator[LocatedFileStatus] = srcFs.listFiles(new Path(rawPath), true) while (fileStatusListIterator.hasNext) { val fileStatus: LocatedFileStatus = fileStatusListIterator.next() paths += fileStatus.getPath } paths.foreach { path => val finalPath = processingPath + path.toUri.getRawPath FileUtil.copy(srcFs, path, dstFs, new Path(finalPath), deleteSource, overwrite, sparkSession.sparkContext.hadoopConfiguration) } } Hadoop FileSystem in Spark emulates the move() option across S3 buckets by utilising the copy() function. Though a little prior setup has to be done to ensure the move behaviour, the end result is completely satisfactory. P.S — Ensure the hadoop-aws jar needed for the FileSystem API is above 2.7.0, if setting the “mapreduce.fileoutputcommitter.algorithm.version” parameter to 2. For more stories check out my publication- https://medium.com/polar-tropics
https://medium.com/polar-tropics/emulating-the-move-functionality-in-s3-using-spark-7b13b18388a1
['Shitij Goyal']
2017-10-04 17:22:38.568000+00:00
['Apache Spark', 'Aws S3', 'Hadoop', 'Technology', 'Big Data']
1,943
The Pros And Cons of Apple M1 Chip
Pros: The M1 Chip Is Faster Researchers have tested these chips in their labs through their test machines. They have measured every single up and downside of M1 chimp while it’s being tested on its pace. And the results were impressive. The new M1 chip is faster than any of the other processing chips that Apple was using before. Apple’s manufacturing team has also included a memory (SOC) inside the M1 chip which can speed up the transmission of data. Aside from factory tests. People all over the world are being a positive signboard that the M1 chip is actually faster. No Thermal Throttling The M1 chips are super silent, the fan never kicks in, even during the high-performance tasks. While Intel processing base Mac models will start screaming out whenever you start high performing tasks. But the M1 chips are a lot more smoother and silent than other processors out there. M1 Chips Are Power Efficient Long battery life is another best thing that M1 chips are offering to the world. With the most power-efficient chip model, the M1 chips can provide the most power-efficient laptops and you can experience a lot of efficiency in battery life. The Price of M1 Based Macs are Reasonable Apple has always provided its products in the perfect price range. $1000 for M1 based Macs is also a good example of it. With a lot more efficient and powerful chip apple is still selling these products at $1000. I haven’t seen any other $1000 laptop that provides more or even equal qualities like M1 based Apple Laptops. In fact, not The ThinkPad with Ryzen processors are behind M1 (technology-wise) and much expensive. It Has Neural Network Processing Technology The Apple M1 chips are built-in Neural Network Processing Technology (Machine Learning). Something that no other company is providing, not Intel nor AMD. For those folks who have no idea what NNP(Neural Network Processing) is, It’s a CPU-based machine learning technology that uses AI to observe and operates on human behavior. And Apple is going a step further with it using this phenomenal technology in the M1.
https://medium.com/macoclock/the-pros-and-cons-of-apple-m1-chip-5581f502a29
['Ghani Mengal']
2020-12-09 04:58:43.689000+00:00
['Apple', 'Mac', 'Apple M1 Chip', 'Tech', 'Technology']
1,944
How to Ensure Security for Your Remote Workers
How to Ensure Security for Your Remote Workers Remote work is more critical than ever: Five infosec leaders share five timely lessons. The way we work changed in response to COVID-19. The “new normal” is here to stay. Today, ensuring security for remote workers is more critical than ever. Organizations across the globe must rely on the IT department for technology solutions, especially as they relate to remote work. During a recent virtual forum hosted by SIM San Diego, four CISOs and one Director of Cybersecurity discussed how their organizations are ensuring security for remote workers during the COVID-19 pandemic. The group revealed trends in cybersecurity best practices and compliance among technology teams. You can use their feedback to gauge how your team is handling its response to business continuity concerns in this new-normal world of virtual collaboration. 1. Trust your communication and self-help documentation Source: Gary Hayslip, CISO, Softbank Gary Hayslip explained that Softbank was 100% cloud when COVID-19 hit. Aside from partnering with HR, their IT processes remained intact — business as usual. Most employees worked off a laptop and used Saas tools in the office, so preparing for the work-from-home environment wasn’t too tricky. As a cloud-based company, initial action included using desktop apps for IT support. Employees with a laptop needed nothing more than login credentials. This feature connects remote workers with live help and immediate information to combat potential security or device issues. However, Gary’s tech team also deploys self-help documents, which they regularly push out to staff. The dual-action approach decreases support tickets and reduces fatigue among the IT team, allowing them to focus on more important projects. A cybersecurity leader for the company, Gary continuously educates the company on critical security threats. And today’s biggest threat is highly-targeted phishing attempts: malicious parties hide behind COVID-19 emails, a fascinating subject in anyone’s inbox. So Gary’s team engages staff members to demonstrate what to look for when spotting risky emails — it’s a game of education and constant communication. Gary believes that communication is the most essential factor when ensuring security from remote workers. Open and easy-to-use tools empower people to seek help in potentially harmful situations. Tools such as Slack and video conferencing are a must for any security team.
https://medium.com/swlh/how-to-ensure-security-for-your-remote-workers-445659dcc62
['The Carrera Agency Designing North']
2020-07-30 10:31:00.861000+00:00
['Cybersecurity', 'Remote Working', 'Information Technology', 'Covid 19', 'Leadership']
1,945
A Beginners Guide to Scikit-Learn
Estimators The Scitkit-learn library provides a very large variety of pre-built algorithms to perform both supervised and unsupervised machine learning. They are generally referred to as estimators. The estimator you choose for your project will depend on the data set you have and the problem that you are trying to solve. The Scikit-learn documentation helpfully provides this diagram, shown below, to help you to determine which algorithm is right for your task. What makes Scikit-learn so straight forward to use is that regardless of the model or algorithm you are using, the code structure for model training and prediction is the same. To illustrate this let’s run through an example. Let’s say you are working on a regression problem and want to train a linear regression algorithm and use the resulting model to make predictions. The first step, with Scikit-learn, is to call the logistic regression estimator and save it as an object. The example below calls the algorithm and saves it as an object called lr. The next step is to fit the model to some training data. This is performed using the fit() method. We call lr.fit() on the features and target data and save the resulting model as an object called model. In the example below I am also using the train_test_split() method to split the dataset into test and train data. Next, we use the model and the predict() method to predict on previously unseen data. If we were to now use Scitkit-learn to perform a different task, say for example, we wanted to train a random forest classifier. The code would look very similar and have the same number of steps. This consistent code structure makes developing machine learning models extremely straight forward and also produces code that is highly readable and reproducible.
https://towardsdatascience.com/a-beginners-guide-to-scikit-learn-14b7e51d71a4
['Rebecca Vickery']
2020-09-25 14:36:15.782000+00:00
['Machine Learning', 'Data Science', 'Technology', 'Education', 'Getting Started']
1,946
XBRL: New Structure and Price of Information Focus of Day 1
SANTA CLARA, Calif. — The XBRL Pacific Rim Technology Workshop and Summit kicked off Tuesday afternoon with experts agreeing about the quality of the first 42 mandatory U.S. Securities and Exchange Commission filings and demonstrating palpable ambition for the future of free financial information. Well, mostly free. Free (as in libre) accounting information in new SEC filings was Topic A, but a new fee to view new interpretations and guidance created by the Financial Accounting Standards Board in its massive project to codify U.S. GAAP was soon Topic B. FASB’s “basic view” of its codification is free (as in gratis) to visitors to its Web site, but the price to enjoy “full functionality and advanced navigation” on FASB’s Web site is $850. FASB’s codification has added more references to literature than are contained in early versions of the XBRL U.S. GAAP taxonomies, but those references should be quickly incorporated into taxonomy updates. Workshop participants didn’t say so, but the $850 fee to access those references may be a non-trivial barrier to entry to at least some software developers looking to use FASB’s Web site to help them develop innovative ways to improve financial reporting. Whether $850 is fair to retail investors seeking to understand financial statements of companies in which they invest — or software designers who might create tools to help investors — is up to the SEC, which oversees FASB. The free to users (as in gratis) XBRL-tagged GAAP information itself, flowing to investors via the SEC and company Web sites, may be more useful to investors than $850 worth of accounting theory. (Academics get FASB’s professional view for free.) Campbell Pryde, XBRL US Chief Standards Officer, said a “codification extension” to align U.S. GAAP data tags with FASB’s new GAAP codification would be released in early August. Software providers who want to support the codification will be able to do so either by referencing the codification extension taxonomy on the XBRL US Web site or by incorporating the extension taxonomy into their own software, Pryde said. New data tags to deal with new accounting policies are on their way, he said, including one to report fair market value with respect to “transactions that are not orderly.” (Who knew that XBRL’s order would be used to bring transparency to — er — disorder?) Within a few years, taxonomy development will make the need for companies to create custom SEC filing tags “very, very rare,” Pryde said. XBRL US is also developing tools to validate notes disclosure in anticipation of the requirement for detailed notes tagging next year and data tags to automate the global reporting of at least 55 corporate actions and automatically translate those actions into international standards, helping put capital to its highest and best uses around the world. XBRL US plans to finish public review of the corporate actions taxonomy in May 2010 and publish the taxonomy the following month. Pryde also showed a taxonomy for mortgage backed securities to allow loan information now posted as unstructured ASCII and HTML on the SEC’s disclosure site to be structured in XBRL format analogous to GAAP filings. Also on Tuesday, Makoto Koizumi, Senior Consultant to Fujitsu Research Institute, gave an overview of Japan’s adoption of the XBRL standard for capital market, tax, and financial reporting. Fujitsu maintains Japan’s XBRL taxonomies under contract to the Japan’s Financial Services Agency. A group of XBRL US interns from Japan and Korea presented projects including the development of a tool to evaluate the use of custom data tags by SEC filers and to identify tags that are unnecessary because they aren’t being used by filers. XBRL US intends to make the tool freely available, Pryde said. Another intern project on corporate actions noted gaps in data standards and problems caused by the reporting of corporate actions in unstructured format.
https://medium.com/pjwilk/xbrl-new-structure-and-price-of-information-focus-of-day-1-311a6c865089
['Paul Wilkinson']
2017-06-01 01:10:02.166000+00:00
['Journalism', 'Business', 'Regulation', 'Finance', 'Technology']
1,947
CTF Forensic challenge
CTF Forensic challenge Photo by Mika Baumeister on Unsplash Last week a CTF event organized by the Spanish Guardia Civil was organized, the II NATIONAL CYBERLEAGUE GC. This challenge is oriented to students, due to that reason I could not participate. But I have I friend who participate, He knows I love forensic challenges so He sent me one of the challenges that were part of the competition. Their team did not manage to solve this challenge so let’s see what was about and how to solve it. Evidence There are two files: container snap.vmem If you have played other CTF challenges this seems a little obvious but let it break into parts. The container seems to be an encrypted container and snap.vmem it is a RAM acquisition. Secrets in live memory have been always a problem. Having a RAM acquisition can give us a lot of information in a digital forensics investigation. We can discover processes running, dump files, secrets, connections and a lot of useful information. The most popular tool for memory analysis is Volatility. Volatility is an Open Source project with a great and active community behind it, there are alternatives like Rekall but I personally prefer Volatility. Right now Volatility has a 3.0 version with a lot of improvements but it is under beta. Because of that, I used the latest stable release, Volatility 2.6. Volatility Analysis The first thing we need to do is to identify the operative system in order to properly analyzed the live memory adquistion. Every operative system handles memory in a different way. Running image info will give us the suggested operative systems profiles. ./volatility_2.6 -f evidencias/snap.vmem imageinfo The suggested profiles are Windows XP related, we can use one of them WinXPSP2x86 or WinXPSP3x86. Knowing the operative system we can start to extract useful information. I always start with pstree. This will let us know whats processes were running in the system. ./volatility_2.6 -f evidencias/snap.vmem --profile WinXPSP2x86 pstree The most interesting process to lookup is TrueCrypt. TrueCrypt was a program that allows us to created encrypted containers and partitions. Right now it is discontinued and has been replaced by Veracrypt. Some people thought that Truecrypt had hidden vulnerabilities but long history short, nothing was found. Either way, Volatility has some commands centred in analysing Truecrypt processed: truecryptsummary truecryptmaster truecryptpassphrase truecryptsummary can give us information about the TrueCrypt process.
https://infosecwriteups.com/ii-national-cyberleague-ctf-forensic-challenge-a0febba1e9f
['Carlos Cilleruelo']
2020-11-17 18:36:08.727000+00:00
['Security', 'Encryption', 'Ctf', 'Cybersecurity', 'Technology']
1,948
Silly Ideas 1 v0.1
Silly Ideas 1 v0.1 Every now and then i come up with silly ideas in my mind which just passes through. But recently i am thinking of at least making some note of them. From the above picture you might not know what i have made because it's just a rough sketch. So, i will try to explain as much as possible. Micro-bots: So, they are the small robots that can run in the trails after being triggered by the user. They will fill the trails making a tight bond. Stock: Stock will be a slim bag that will hold the micro-bots in it with slots system and wireless charging capabilities to charge the micro-bots Trails: Trails will be more of a flexible but yet strong trail where the bots can travel and make a strong bond. Note: The trail is a strong yet flexible part of the system. Micro bots are the ones which joins to make a strong bond.
https://medium.com/@rameshdura/silly-ideas-1-v0-1-4959e14c6599
['Ramesh Dura']
2020-12-16 02:04:16.896000+00:00
['Technology', 'Robotics', 'Apple']
1,949
When your blockchain needs to roll the dice
Many of the applications we use every day depend on randomness, from to lotteries over scientific application to leader election or private key generation. Today we talk about public randomness, which more precisely refers to values that are unpredictable until revealed but then immutable and available to everyone. Photo by Guillermo Velarde on Unsplash On blockchains randomess is not easy to create due to their deterministic nature. Every participant needs to agree on the same random value. Typical workarounds like block generation times or block hashes are easy to manipulate for a block creator. Common signature schemes like Secp256k1 or Ed25519 don’t produce unique signatures and can be bended at will. You also don’t want to trust a single party for publishing random values as this party might bias the values in a way that favours them or simply stop producing values. Even in more sophisticated commit-reveal schemes, there is at least one party that has the last mover advantage and can either manipulate the result or at least deny publishing its part such that the application halts. There have been attempts to include random beacon functionality into Tendermint directly, but I haven’t seen any production-ready solution yet. Now there is a project called drand, which aims to provide a random number generator for the internet. I’ll not go too much into detail of how it works as the Randomness Summit 2020 is basically a one-day marketing campaign explaining everything about it. But the deal is: instead of trusting a single party, you trust a group of oganizations (the League of Entropy) to produce an unpredictable and unbiased value. The so called random beacons are created every 30 seconds, are ordered by a round number, are signed and can be verified in a client. The CosmWasm drand client In drand, random beacons are distributed via HTTP, Gossipsub, Tor or Twitter. Such network sources cannot be accessed by a blockchain directly. However, we can create a CosmWasm smart contract which allows storing random beacons on chain. Using cross-contract queries, other contracts can then read those random values and use them in their logic. The random beacon of round 72785 Beacon verification Any beacon is self verifiable, i.e. you can check the validity of a single beacon without having to verify the whole chain. The signature is a BLS12–381 signature of the input previous_signature || round . The 32 byte randomness is simply a SHA-256 hash of the signature. In order to avoid accepting malicious beacons, the contract needs to be able to verify the signatures. Drand signature verification is available in the Rust crate drand-verify, which compiles nicely to Wasm and can be used in contracts. There is one caveat though: verifiying a single signature in Wasm takes about 200 milliseconds and costs 500 million CosmWasm VM gas (about 5 million Cosmos SDK gas). Once the beacon is verified, there is no need to store all of its data on chain. We can simply derive the 32 bytes randomness after verification and only store a map from round to randomness. Incentivization model Adding beacons to the randomness contract costs gas and once the randomess was added, it becomes a common good that is freely accessible by everyone. Since beacons become part of the redundant, public, immutable blockchain state, there is no way to restrict access to it without changing fundamentals of the technology. So the question is, how to incentivise an off-chain party to collect beacons and add them to chain? One approach is a bounty system. An application that depends on randomness at round n in the future can put a bounty on round n . Once the beacon becomes available and the bounty is higher than the cost to add it to chain, some actor will add it. If there is not yet enough bounty on n , whoever needs this beacon either needs to increase the bounty or add the beacon for a higher cost. This brings in some interesting ecomonic mechanisms. If the bounty is high enough, many actors will compete on claiming it such that the beacon is available very fast. If the bounty is too low the beacon might be added slowly, having to wait for a cheap block proposer. Multiple applications can pool to provide a bounty for future rounds. This can become very efficient when many apps agree to only use every e.g. 10th or 60th round (i.e. one beacon every 5min or 30 min). Further development Everything we discussed above is implemented and working. Right now, the 657 KiB large contract cannot be uploaded to x/wasm chains, because of a hard size limit. CosmWasm 0.12 will solve that. It might also be necessary to adjust transaction and block size limits as well as gas limits to allow contract upload and execution. Once everything is working, we are considering to add APIs provided by the VM for heavy cryptographic operations. This should lower the gas cost significantly. However, the contract does not depend on such optimizations. Last but not least the economic model might lead to situations where no participant is willing to pay a bounty and waits until others do. This might require better economics. Kudos None of this work would have been possible without the team behind drand, the Electric Coin Co. team for implementing BLS12–381 in Rust and the Filecoin team for implementing Hash to Curve in Rust.
https://medium.com/@simonwarta/when-your-blockchain-needs-to-roll-the-dice-ed9da121f590
['Simon Warta']
2020-11-05 16:41:53.372000+00:00
['Randomness', 'Blockchain Technology', 'Smart Contracts', 'Cosmos Network', 'Cosmwasm']
1,950
Why Interoperability will make 2020 the breakthrough year for blockchain
Gilbert Verdian, Founder and CEO of Quant Network, explains why recent developments in interoperability are likely to change attitudes to blockchain. Ever since the development of the first cryptocurrency (Bitcoin) in 2008, blockchain technology has been hailed by many as a potentially transformative force; a technology which could invest industry and commerce with a new dimension of secure and cost-efficient connectedness. The possibilities seemed limitless. An ‘Internet of Trust’ was on its way. Today, over a decade on, it is still to arrive — at least on a mass scale. While few would deny that the potential of blockchain as a force for good seems greater than ever, it’s equally hard to deny that, for most of the past decade, progress — in terms of practical application — has been slower than expected. It’s true that, after the inevitable early adopters, there have been some, even many, notable uses of the technology, but there’s also little doubt that its adoption has been less explosive than its reputation would seem to merit. But is all that about to change? Could 2020 be, at last, the breakthrough year for blockchain: the year in which the tipping point is reached and blockchain truly starts to become a standard — perhaps even default — approach to connectivity in many sectors? Here at Quant Network, we believe that this could be the case. Why? Because we have seen, over the past year — and particularly in the last quarter of 2019 — a palpable change in the market attitude towards blockchain. This change, an increased level of recognition of the technology’s power and maturity, has manifested itself in a significant increase in demand for our own (blockchain enabling) technology. We feel confident that this trend will grow to the point where 2020 will be remembered as the year blockchain achieved mass implementation. And there’s a significant body of evidence that this feeling is justified; evidence, in other words, that we are genuinely seeing the dawn of an era of mass blockchain adoption. Take, for example, the Deloitte 2019 survey Blockchain gets Down to Business1. This research polled a sample of 1,386 senior executives in a dozen countries, concerning the future of blockchain technology, and found that, for 53% of respondents, blockchain was a critical priority for their organisation in 2019 — a 10% increase over the previous year. The report goes on to say that blockchain is ‘gaining traction and acceptance in more industries’ as a result of “…executives increasingly expressing confidence in blockchains importance…”. It concludes that, today, the question for executives is no longer, “will blockchain work?”, but “will blockchain work for us?”. The findings of the Deloitte report(1) are echoed by the 2019 Gartner CIO Survey(2), which found that 60% of CIOs expect some kind of blockchain deployment in the next three years, with financial services leading the way over the next 12 months, while ComputerWorld(3) notes that industry commentators expect 2020 to be the year that distributed ledger technology (DLT) matures. Further, Finextra(4), the leading independent newswire and information source for the worldwide fintech community, reported, in December 2019, that “increased maturity of blockchain technology will certainly trigger adoption in the coming year(s)”, and the cryptocurrency news site, Bitrates, labels 2020 as the year that we will see ‘blockchain revolutionise the data processes of many industries and governments.” We could go on: many more informed sources voice similar opinions, based on experience and/or market research. But to continue quoting these sources would be to labour the point — which is that blockchain seems to be at the threshold of widespread acceptance and adoption. So let us assume that this is indeed the case, and turn to a different, but critical, issue: why? What has happened to trigger this relatively sudden change of attitude; this crystallisation of exciting potential into imminent reality? The answer to this question lies in understanding what’s been holding blockchain back — the concerns, in other words, which have prevented organisations at all levels from committing to the technology on a mass scale. There have, in fact, been several such concerns in recent years, all of them legitimate. One of these is the need for standards and regulation. Historically, blockchains have been developed independently, without the need for, or regard, to standards. This has led to siloed ecosystems that discourage harmonisation of distributed ledger technology. This was one of my concerns in 2015, when I put forward to establish Blockchain ISO Standards to harmonise the efforts to develop distributed ledger technology in collaboration rather than tribal competition, leading to the creation of ISO/TC307. Furthermore, there has been significant recent progress on this front, with input from the ISO and various standards organisations and working groups who are collaborating to progress the technology. Today, there are over 57 countries and international organisations who are actively participating in the development of Blockchain Standards with the first set of standards starting to be published in 2020. Scalability is another issue which has contributed to slow blockchain implementation. Again, however, the past year or two has seen major advances in this regard, and there is an increasing number of use cases which demonstrate that, today, blockchain solutions can deliver mission-critical scalability required by enterprise and regulated environments. These issues, and others, have all been legitimate points of concern. But by far the most widely recognised problematic issue is that of interoperability. Or, more accurately, the lack of it. It is this concern which, more than any other, has prevented blockchain from transforming the way the world connects. And this is understandable. After all, interoperability is the ability of blockchains to share, see and access information — and it’s easy to see why this is not just desirable, but critical, in a world where enterprises depend on ever-greater levels of collaboration and interaction. Without interoperability, organisations are concerned that, in adopting a blockchain solution, they could become isolated; effectively shut off from forming constructive relationships with complementary businesses. Other organisations fear that they could find themselves victim to vendor ‘lock-in’ — being unable to migrate to a different blockchain solution if (as could easily happen, given the nascent nature of blockchain) the technology changes. This is a non-trivial problem, as it is embedded in the very roots of blockchain development. It is a fundamental lack of communication ability which means that not only are different blockchains (Corda to Ethereum or Hyperledger, etc.) unable to easily communicate with each other, but neither can blockchains of the same type (e.g. Corda to Corda). To illustrate the impact of this limitation, consider, for example, a network of 10 banks, running on a single blockchain. This network will not be able to transact and communicate with a separate network of 1000 banks, even though they are running the same blockchain technology and version. The problem is further compounded by different networks and banks running completely different governance rules, blockchain technology, versions and regulatory controls. Clearly, a solution which delivers interoperability is an essential precondition for mass adoption of blockchain to be achieved. So important, in fact, is the need for interoperability, that, in its 2019 survey(2), Gartner names (the lack of) it as one of the reasons that blockchain “…remains immature for enterprise deployments”, and Finextra(4) claim that, in 2020, ‘Interoperability will move centre stage.” Very recently (January 2020), the World Economic Forum described blockchain technology as being “balkanised in silos.”(6) It is precisely this recognition of the importance of, and need for, interoperability that is at the root of our confidence that 2020 will be a breakthrough year for blockchain. Such reasoning may seem counterintuitive, unless you know that, at Quant Network, our core technology is focused on interoperability. So we’re well-placed to predict when a viable, practical, ready-to-go solution might arrive. When might this be? Well, the answer is that it’s here now. And it’s called Overledger. So, what is Overledger? First, it’s important to note that it’s not, itself, a blockchain. Rather, it is a protocol that — as its name implies — is an overlay on top of other blockchains. It is, in other words, an operating system (OS). In fact, it is the World’s first and only enterprise-grade blockchain operating system that allows interoperability of any current or future blockchain, legacy networks and enterprise applications and platforms, without adding an overhead of complexity or another blockchain layer. This allows the creation of multi-application chains (MApps) that can leverage the functionalities of different blockchains. For example, developers can pick a particular feature of one blockchain, such as high throughput, and couple it with a feature derived from another blockchain, such as decentralisation. This optimises the agnostics (the ability of one blockchain to interact with another) of the application. Overledger is also designed to be future proof, which it accomplishes by isolating the transaction layers between blockchains. With Overledger, developers can start building game-changing applications with no additional infrastructure, using existing resources, with only three lines of code. Conclusion While interoperability has been, and is, widely recognised as a key issue in the mass adoption of blockchain, it is no longer an obstacle. At Quant, we are confident that the removal of this obstacle, ecosystem collaboration and technology evolution signals the beginning of a new age of global connectivity: the Age of Blockchain. More information: Overledger — Overview Overledger White Paper Overledger — the Possibilities References 1. Deloitte Touche Tohmatsu Ltd, 2019. Deloitte’s 2019 Global Blockchain Survey. https://www2.deloitte.com/content/dam/Deloitte/se/Documents/risk/DI_2019-global-blockchain-survey.pdf 2. Gartner, Inc., 2019. Top 10 Strategic Technology Trends for 2020. https://www.gartner.com/smarterwithgartner/gartner-top-10-strategic-technology-trends-for-2020/ 3. Computerworld, 2019. Top 3 enterprise tech trends to watch in 2020. https://www.computerworld.com/article/3512109/top-3-enterprise-tech-trends-to-watch-in-2020.html 4. Finextra, Dec 2019. What may we expect for blockchain and the crypto markets in 2020? https://www.finextra.com/blogposting/18285/what-may-we-expect-for-blockchain-and-the-crypto-markets-in-2020 5. Bitrates, 2019. These 4 Trends Will Shape The Blockchain Industry In The 2020s. https://www.bitrates.com/news/p/these-4-trends-will-shape-the-blockchain-industry-in-the-2020s 6. World Economic Forum: These are the challenges blockchain faces in 2020. https://www.weforum.org/agenda/2020/01/blockchain-in-2020-epic-changes-and-monumental-challenges/
https://medium.com/@quantnetwork/why-interoperability-will-make-2020-the-breakthrough-year-for-blockchain-91d0185762aa
[]
2020-03-18 11:41:21.174000+00:00
['Distributed Ledgers', 'Blockchain', 'Enterprise Technology', 'Decentralization', 'Interoperability']
1,951
A Successful Medieval Siege Involved Good Psychology and a Dose of Chivalry
A Successful Medieval Siege Involved Good Psychology and a Dose of Chivalry Brute force was never top of the list “If a conflict breaks out in England one or other of the rivals is master in 10 days or less. It is a custom in England that the victors in battle kill nobody, especially none of the ordinary soldiers because everyone wants to please them.” — Memoirs of Philllpe de Commynes, French writer and diplomat (1447–1571) Can this be right? What about knights — and lances — and melees? Medieval men loved a good fight — didn’t they? Well, sure they did sometimes. However, in a landscape festooned with castles, there was a better way to get what you wanted — the siege. We’ve already looked at what life was like for castle defenders here, but what was it like for the attackers? If you think besieging a castle meant a lot of sitting around twiddling your thumbs, think again. There was much to be done and, as always, the rules of chivalry need to be followed. Well, they were supposed to be anyway. After all, a gentleman should be civil, and in a siege that civility had to start right from the beginning. In fact, it started even before that. Feeling angry? Let’s talk about it It was vitally important for a siege to have an obvious starting point. However, before a siege could officially begin, the defenders needed to be presented with the demands of the attacking lord. If the two sides came up with an agreement (i.e. the defenders gave their adversaries everything they wanted), the lives and property of the people in the castle would be spared. Should they refuse, they were guilty of treason toward the besieging lord. As traitors, they were fair game for all those nasty punishments like rape, seizure of goods, death, and whatever else an angry attacker might think up. Now, if you were the captain of the besieged castle, this put you in a pretty precarious position. On one hand, you have a duty to guard your lord’s property. He’s not going to be pleased if you casually give his fortress away to his enemy. On the other hand, should the castle fall, the new lord might chop you up like liver for his kitty’s dinner. Either way, your stuck — someone is going to want your head. So what do you do? Well, if you were smart, you would have this all sorted out beforehand. What you needed were your lord's wishes written out and signed in triplicate. The document would specify exactly how long you were expected to hold out, and the conditions on which you could surrender without fear of losing your honor. Most of the time, part of these conditions included holding a castle until reinforcements had the chance to arrive. So assuming you weren’t immediately giving in to the besieger's demands, things would proceed to the next step. On your mark, get set, go Usually, there was an event to signify the siege was officially starting. It could be something as simple as throwing a spear or javelin, or as grand as a shot being fired from a siege engine. After this, the fun began. Nobody wanted to just sit around twiddling their thumbs waiting for a castle to fall. The clock was ticking, and they wanted to get the thing over and done with ASAP. Attackers, therefore, had a menu of strategies they could pick from in order to help move things along. According to historian Jim Bradbury, there were 6 S’s to siege warfare. Attackers had their choice of Suborning or subverting key defenders, Scaring the garrisons with “propaganda,” Sapping the walls, Starving the population, Storming the defenses, and Shelling the besieged. Don’t worry, the defenders weren’t left out, they had their S’s too. These included Stripping the countryside to deprive the besiegers of the opportunity to forage, Sallying forth in sorties to attack enemy positions and assets, Sapping the attacker’s mines and machines, Suborning and subverting elements of the besieging forces, and Shelling enemy positions. How many of the S’s were used, and in what order was entirely up to the commander and the specifics of each siege. Hurry up and starve Starving the population was an easy one and really didn’t require much effort by the besiegers. All they had to do was wait for supplies inside the castle to run out. This came with its own risks, especially if the besieged were well stocked. Even the attackers risked hunger and diseases like dysentery, so the longer they sat, the more likely something would go wrong. On top of that, no one wanted to be parked outside a castle when the weather started to turn cold. However, there was an even more pressing problem. Imagine you’re in the middle of a siege and suddenly half your army walks out because their terms of service are up. In England, forty days was the usual agreement. After that, they grabbed their swords and their superman lunchboxes, and home they went. Poor Simon de Montfort had this very problem when he fought the Albigensian Crusade (1209–1229) in Southern France. It’s said he lost up to 40 strongholds because of men walking away from the fight. Gonna rock this place Shelling the castle walls was a good way to keep your troops busy and wear out the nerves of the besieged. Trebuchets, which are similar to catapults, were a popular option. Large stones could be hurled at the castle's walls, but other things like burning tar or sand were employed. Engineers plotted and tinkered to make sure the weapon would have the greatest effect possible. Considering these things could shoot up to 200 times a day, I’d say they did a pretty good job. The largest trebuchet ever built was called the War Wolf. Its height was between 300 and 400 feet, and it’s estimated it could hurl stones at speeds up to 120 miles an hour. Undermining the local authority Not everyone was fortunate enough to have a trebuchet as powerful as the War Wolf. Since castle walls ranged between 7 and 20 feet thick, hurling stones wasn’t always going to bring about a quick conclusion. So while some men were busy bombarding the walls, the sappers started their work. Sapping, or undermining, was basically digging a tunnel under the castle wall. Wooden supports were inserted, then burnt away in hopes the collapsing ground would take the castle walls with it. Although shields and tunnels offered protection for the diggers, the process wasn’t foolproof. Defenders could build counter-mines that intercepted those of the enemy to drive them out. The most famous use of undermining is probably the attack on Rochester Castle in England in 1215. King John had his men tunnel under the southeast corner turret of the tower. After propping it up with wood, he ordered the fat from 40 slaughtered pigs, “of the kind not fit for eating”, be placed in the tunnel to really get things fired up. It worked, and John's men eventually took the castle. We control the horizontal and the vertical Chivalry was all fine and dandy, but it was certainly something that could be set aside when it no longer suited one's needs. A psychological attack was one of those times. All types of trickery and deception could be used to wear down the morale of the besieged. The enemy might forge a document to the besieged commander from his sovereign ordering him to surrender. Soldiers might disguise themselves and slip into the castle, or a knight thought to be friendly might have secretly switched allegiance. Sometimes feelings of hatred grew so high, commanders were killed right in the middle of discussing terms of peace. Poisoning the water supply and catapulting dead animals or dung over the walls to spread disease were other methods attackers used that on paper were just not cricket. Then again, some days a guy was just in the mood for a little fun. Just take the siege of Rennes in Brittany in 1356–57. According to the story, an English esquire named John Bolton got bored while sieging the Breton castle and went hunting. It was a wise decision too because he came back with three braces of partridges. Pleased with himself, he rode up to the castle wall and offered to sell his catch to feed the ladies trapped inside. Well, this hit the pride of one of the Frenchmen, Olivier de Mannay. How about you come over here and I’ll win those birds for the ladies by a duel, he suggested. Well, not surprisingly, the governor refused to open the gates to let Bolton in. However, De Mannay was determined to fight. He swam across the moat, won the duel, and swam back to the castle — with the birds in hand. All right, that’s a wrap Possibly after all this, the attacking army broke through the walls. Maybe their siege towers were put to use, and the men overran the castle. Then again, maybe they didn’t, not every siege was successful. Anyway, a full-blown assault really wasn’t beneficial to anyone. More often than not when a fortification did fall it was due to human weakness or because relieving forces failed to come to a garrison’s aid. Either way, when it came down to it, everyone involved would much rather see the thing resolved by negotiation rather than warfare. For one, the attackers would much rather take important people prisoner than have them killed. These guys were an opportunity for financial gain in the form of ransom, and good insurance for the day when the present winner was on the losing side. For the besieged, as long as you made some effort you were usually ok in the eyes of your peers. And if it was hopeless, and you gave in at the start, well usually people understood that too. The good thing is, surrender didn’t have to end in total tragedy. David of Scotland offered the besieged at Wark in 1138 not only freedom if they surrendered, but fresh horses. (They had eaten theirs when the food ran out.) Then there was Conrad III, who wanted to imprison all the defenders except the women. He agreed to let them go with whatever they could carry. They carried out their men. Notes and sources: As I mentioned in the last post, in the early days, towns were held under siege more than castles. However, the process was the same either way. Just because people wanted a siege to end quickly, doesn’t mean it always happened. The sieges of Chateau-Gaillard (1203–4) and Rouen (1418–19) each lasted nearly six months; Calais (1346–47) took eleven months. Don’t know what I mean by “we control the horizontal and the vertical?” Oh, you youngins. It’s from a TV show, The Outer Limits. You know, back in the day when T.V.’s had tubes and analog dials. Medieval Warfare: A History, Maurice Keen (Oxford University Press) https://www.english-heritage.org.uk/visit/places/kenilworth-castle/history-and-stories/siege-kenilworth-castle/ https://stpaulscastles.weebly.com/square-towers-or-round-towers.html https://www.encyclopedia.com/women/encyclopedias-almanacs-transcripts-and-maps/siege-warfare-and-women-8th-c-17th-c https://deremilitari.org/2013/06/the-myths-of-medieval-warfare/ https://deremilitari.org/2013/11/medieval-siege-warfare-a-reconnaissance/ http://www.castlesandmanorhouses.com/architecture_03_walls.htm http://www.medieval-life-and-times.info/medieval-weapons/under-siege.htm https://www.english-heritage.org.uk/visit/inspire-me/10-things-sieges/ https://deremilitari.org/2013/06/the-myths-of-medieval-warfare/ https://www.ancient.eu/article/1230/siege-warfare-in-medieval-europe/
https://medium.com/lessons-from-history/a-successful-medieval-siege-involved-good-psychology-and-a-dose-of-chivalry-1a55f882cd8f
['Nicol Valentin']
2020-11-21 00:02:03.030000+00:00
['War', 'Culture', 'History', 'Technology', 'Psychology']
1,952
Momentum — The Mentor Driven Acceleration Program
Momentum Start-Up Accelerator is a 3-month startup acceleration program designed for ventures formed by Students & Alumni of the Interdisciplinary Center Herzliya (IDC). In 2001, IDC faculty decided to develop and launch the IDC Entrepreneurship Club (IEC), thus initiating what would become Israel’s largest Student Entrepreneurship Club, serving over 1,000 students annually. Momentum Start-Up Accelerator serves as the flagship program of the IDC Entrepreneurship Club — students and alumni of IDC, who have already begun working on their venture and are past the ideation stage, can apply during one of the two enrollment periods which occur at the beginning of the winter and spring semesters, accordingly. As the enrollment process starts teams are handpicked from a long list of applicants by various criteria as Idea Assessment, Pitching, Business Model and Scalability Potential. During the program, in which each batch period is one full academic semester, participants go through a planned syllabus structure designed to propell their Start-Up from Pre-Seed stage to competing at a higher level, whether by raising capital or by competing as a structured business over a market niche. In addition to the structure of the program, which covers all relevant aspects for an emerging Start-Up (whether it be Storytelling, Market fit & Validation, Legal & Fiscal Advisory, or the Demo Day — a highlight event in which participants pitch their Start-Up to VC & Angel Investors), participats also earn a Mentor to guide and advise them throughout the program. Mentors at Momentum serve as the key player of the program, as they are sharing with the participants invalueable lessons learned from their proven experience in the private sector. There are various great Acceleration programs out there, many focusing on different stages that Start-Ups are at, or even at specific emerging industries. In addition, most Acceleration programs are obligated to their investors due to the fact that such programs have to make sense economically. However, as a non-profit, full time volunteering based organization, what makes Momentum unique is that the participating teams are the client, and not the product — Momentum reversed the order of the model.
https://medium.com/@domhur/momentum-the-mentor-driven-acceleration-program-ea99db900fe
['Dominik M. Hurmann']
2020-12-19 23:53:10.375000+00:00
['Accelerator', 'Entrepreneurship', 'Startup', 'Israel', 'Technology']
1,953
Manager OKRs, Maker OKRs: How Early Stage Startups Should Think About Goal-Setting
Manager OKRs, Maker OKRs: How Early Stage Startups Should Think About Goal-Setting Google’s internal management approach has sustained and scaled pretty impressively over the years. Quantitative goal-setting, setting stretch targets — these principles are as evident in the 2020s as they were when I arrived in 2003. Underpinning it all are OKRs — Objectives and Key Results — the framework by which individuals, teams and the entire company is managed. Xoogler Don Dodge did a comprehensive job of documenting OKRs in an earlier blog post, but the basics were always this: Each quarter individuals and teams document their objectives for the next 90 days and grade the goals they set 90 days earlier. In Q4 teams also set metagoals for the next year. Now we have software startups who have basically built their business around OKR-style planning! Or you could, of course, use this OKR template from Homebrew portfolio company Coda. Photo by Jeffrey F Lin on Unsplash My Nine Years at Google meant 36 Quarterly OKRs and the correlating number of annual planning exercises. My role at YouTube had me often working through our OKRs with Larry and Sergey (one of those stressful exercises that in hindsight was amazing). I believe OKRs were originally recommended to L&S by John Doerr of KPCB, and since that time, OKRs have spread through tech companies, sometimes carried by Google alumnus themselves. OKRs are sensible, straight forward and on a planning cycle managers understand. And that’s the problem. In 2009 Y Combinator founder Paul Graham wrote an essential essay called Maker’s Schedule, Manager’s Schedule. The post discusses how engineers need long periods of dedicated time to build and managers (or people whose work generally involves lots of meetings) can honor this by not scheduling interruptions in the middle of these periods. It’s great — you should read it. Manager time vs Maker time gave me a lens to not just evaluate day to day schedules, but the general cadence of how we plan and build at Google. We consider ourselves a company founded and driven by Makers (our engineers), but somehow we settled into a Manager planning rhythm, one which mimicked accounting cycles rather than how things actually get built. “Quarterly goals?” Why are three months the right duration for building features, why not two months or four months? And there was the amusing “last week of quarter” push to try and ship all the features you’d committed to ~90 days earlier. Even more confusing were annual goals. By Q4, it’s pretty clear whether you’re going to hit the annual goals, the high level targets meant to inspire a year of work, but because you haven’t started next year’s planning cycle, the team has no documented targets for what the next 12 months look like. (Obviously the best managers start with an evergreen vision and then break into planning cycles — this isn’t about roadmapping within teams — but the Quarterly + Annual segmenting is still derived from financial planning, not hacking). What would I recommend for the Maker-side of early startup companies instead of Quarterly + Calendar Year Annual? These three: One Month — “What are we building this month” is the key question. Team leads get together the morning of the Monday prior to month’s end and document the next month’s feature releases. This is a bottom up process which includes items shipped completely intra-month and component work of projects which are greater than 30 days long (if you can’t break a complex project into at least 30 days goals, then it’s too big). Four weeks, a few weekends. Enough time to get a lot done. You don’t need to micromanage — for example, if the team spends two days per month bug fixing, just hold that time aside in your calculations, don’t document the bugs you intend to fix. “What are we building this month” is the key question. Team leads get together the morning of the Monday prior to month’s end and document the next month’s feature releases. This is a bottom up process which includes items shipped completely intra-month and component work of projects which are greater than 30 days long (if you can’t break a complex project into at least 30 days goals, then it’s too big). Four weeks, a few weekends. Enough time to get a lot done. You don’t need to micromanage — for example, if the team spends two days per month bug fixing, just hold that time aside in your calculations, don’t document the bugs you intend to fix. “N+12 Months” — “What will our product and business look like a year from now?” I like the idea of a rolling “one year out” vision, processing new learnings and opportunities. At any given time the entire organization can have a true north for where we want to be a year from now. It evolves, it learns, it doesn’t tick down to zero but rather always looks out over the horizon. “What will our product and business look like a year from now?” I like the idea of a rolling “one year out” vision, processing new learnings and opportunities. At any given time the entire organization can have a true north for where we want to be a year from now. It evolves, it learns, it doesn’t tick down to zero but rather always looks out over the horizon. Minimal Quarterly/Annual KPIs — Recognizing that quarterly and annual goals are important for financial reporting and goaling, you should keep a very narrow grasp on what you actually want to measure — just key drivers of business — and set quarterly targets. There can be a reality check — do these quarterly targets get achieved given what we’re building? For me, Monthly Goals combined with N+12 Goals create the right short-term Maker cadence with longer term vision. I never got the chance to try it at Google, but hope to find companies using this sort of planning cycle to see how it works for them. [This post updates a version I wrote in 2013. Wanted to refresh and re-share because it’s planning season!]
https://medium.com/@hunterwalk/manager-okrs-maker-okrs-how-startups-should-think-about-goal-setting-7be87195fadc
['Hunter Walk']
2020-12-21 14:40:35+00:00
['Goal Setting', 'Goals', 'Startup', 'Okr', 'Technology']
1,954
WICV2020 | German Automobile Associations VDA Strengthens Collaboration with China Ecosystem
Good morning everyone! it is a great honor to participate in our theme forum today. First of all, I want to introduce to you the conclusions of our current research, and I will discuss today’s topic again. Speaking of the VDA organization, I would like to give you a brief introduction. Just now we heard Mr. Li’s speech mentioned that we need to strengthen cooperation. We very much agree with this view. For example, we have seen that German automobile companies have always cooperated very closely with related parties of Chinese automobile companies. For another example, Cartel Germany also has related branches, which once again proves the importance of cooperation. This leads me to my speech today. Why do I want to give you a speech today? Let me give you a brief introduction first. The headquarters of the VDA German Automobile Industry Association is located in Berlin, the capital of Germany. We were founded in 1901. We had our headquarters in Frankfurt before 2010 and later moved to Berlin, the capital of Germany. We are an institution that can influence the German government in terms of automobile policy, but for us, it is not enough to focus on the internal situation in Germany. We need to go out, so we have also set up offices overseas. In addition to Beijing, We also set up an office in Brussels. China is a very important automotive market. Therefore, we came to China six or seven years ago and set up an overseas office in Beijing. So far, we have a total of more than 600 member companies. In addition, these member companies include not only automakers, but also others, such as suppliers and component suppliers. We can basically cover the entire automobile production process. square. Recently, our organization has undergone a certain reorganization in October. So far, we now have 1 chairman and 4 Managing Directors, because we hope to fully represent the situation of the entire industry and we also hope to handle our overseas offices well. Therefore, we have set up 4 managing directors. Especially under the premise that automation has become a new trend, there are extremely many working groups. Therefore, we hope that the interests of various suppliers can be well represented in the entire supply chain. We have also organized many top events in the automotive industry, including next year we will organize a series of important events such as car exhibitions, including many important customers, as well as suppliers, dealers and car companies will participate. Let me introduce to you VDA’s work. The main content of our work has been in the past many years. We have fully represented the German automotive industry in Germany and international standards development organizations. The automotive industry is very important. Moreover, the automotive industry involves many sub-fields, so we are responsible for this part of the work, covering various issues in the automotive field. Including standardized work, formulation, tool development, etc. In addition, the standardization department of VDA is equivalent to that of the German Automotive Technical Standards Committee. We all know that standards and regulations are very critical to the automotive industry. In the entire automotive industry, there are many different stakeholders. On the one hand, they are producers, automakers, and car companies. But for automakers, they will face many difficulties and challenges, such as pollution, noise, etc., so we need to introduce laws and regulations to carry out relevant regulations and carry out relevant supervision. At the other end of the industry, we see that there is a product side, and the product side requires the participation of suppliers. We will see a lot of competition, so we need to continuously optimize various processes, such as production processes, we need to set standards for related optimization work. These standards actually have similarities and differences, and hope to cover all aspects of the entire automotive industry. We hope to find a good balance between standards and specifications. Let’s take a look at the differences between standards and regulations. This is mainly an introduction from a German perspective. For example, let’s look at laws and regulations first. Laws and regulations are more concerned with personal safety and environmental safety, these social benefits. Because when we are driving, not only do we drive ourselves, but you will have an impact on many surrounding environments and individuals. We can also represent the most advanced technology in a standard way. This is the difference between standards and regulations. Products have different prices. Therefore, certain standards must be adopted to control the quality and cost of products. This is also an aspect of the standards. All in all, the standard is to support different manufacturers and car companies to fulfill their related obligations and urge them to complete their obligations themselves. For example, the two pictures shown for you are car seats. Children’s chairs need standards to regulate some of the details. For example, there are many standard details about how to set the seat belt on the car seat. Of course, the final consumer decides whether to buy such a product, but at least such a product should be provided to consumers to make such a choice, which involves standards. Regulations must be followed, but standards are selective. I would like to introduce you to the German national future travel platform. Today we are discussing intelligent connected vehicles. For intelligent connected vehicles, including standard specifications, Is a very critical element. Especially in Germany and other countries, such as China, we see that car leasing, online car-hailing, etc. are all very new trends. These require a unified integration platform from a technical perspective or from other perspectives. , We hope to integrate these elements into research and products. We need to introduce digital technology, we also need to introduce a lot of experts and human resources, so from a social perspective, how do we introduce these technical experts and professionals to decision makers, and only then can we find the best Solutions to solve various complex problems faced by the automotive industry. This is the significance of our establishment of such a platform, which can integrate the opinions and discussions of all parties to solve many technical, legal and social problems and challenges. This is the structure of the travel platform we just mentioned. Here is a simple chart for everyone, which is very intuitive. The entire NPM is highly systemic and consists of different working groups. Everyone knows that digitization and standardization are very important elements, so you can see that we have set up a special working group, the third working group and the third working group. The six working groups are respectively responsible for digitization and standardization. We hope to cover the most critical topics in the development of the entire automotive field by setting up different working groups. Now let me introduce these two working groups specifically for you. The first is the third working group that I just created, the digital working group. The work of the Digitization Working Group covers all aspects of digitization. On the one hand, it ensures emission reduction and sustainable development. At the same time, it hopes to meet personal travel needs and to continuously improve travel efficiency and safety. In addition, suggestions will be made to government decision-making departments on these issues. The following is the standardization specification and certification working group. I have just introduced to you what are the characteristics of specifications and standards. This is the work content of the working group. I hope to carry out relevant coordination work, including certification work. When it comes to intelligent connected vehicles, autonomous driving is a very important topic. In fact, ICV is a very new topic. Whether it is in Germany or other countries, ICV is just getting started. For intelligent connected vehicles, autonomous driving The testing and development of the car is very critical. You see that this is the time required for Level 3 and Level 4 autonomous driving. Now we are looking forward to the next five years from 2020–2025. We need to intensively carry out related fields. Standardization work, especially autonomous driving. What we introduce to you is the core functional standards that are underway or in the plan. These standards are consistent with our work philosophy. I just heard the previous speakers mentioned similar topics. China has also done a lot of work on functional standards. We are also making continuous efforts in this area. We hope to eventually raise it to the level of national regulatory policies. I think we can also carry out more international cooperation and discussions in this regard. China’s approach in terms of functional standards is similar, and we are very much looking forward to having more dialogues and cooperation with China on this issue in the future. In fact, for intelligent networked vehicles, the issue of functional standards is a very difficult issue that needs to be resolved urgently. We need to increase investment in this area. When it comes to standards, standards are not only standards for products, standards for vehicles, but also standards for testing. This is also the core content of our work, especially for intelligent networked vehicles. We hope to establish a unified test. process. Including how to set up and arrange the test roads, how to deal with special situations, and how to deal with tests under different scenarios. If we do not make a unified definition, there will be a lot of trouble. For intelligent networked vehicles, a large number of road tests will need to be carried out in the future. We need unified testing and standardized testing. We even say that we need to complete unified standards at the digital level and at the virtual level, including The same is true at the national level, one is simulation, another is verification, and the other is testing. These three tasks need to be strengthened at the national level. We also hope that the unified standard can be promoted to the world to form a global unified testing and release process. Let’s take a look at the current automobile network composition of different regions and various parties around the world, especially the relevant standards involved in enterprises. We need a top-down approach to accomplish this work. Some companies have very strong innovation capabilities, such as Bosch. Bosch has started relevant work a long time ago. You can see that the earliest time can be traced back to the 1980s. We need to learn from such professional technology and promote it to the entire industry. As you can see, we now have more data. We need to transmit the data to the Ethernet for data analysis and calculation. In this regard, we also need to promote the formulation of relevant standards to truly realize automated standards. This must involve the expansion of vehicle standards, because now vehicles are all networked, and more stakeholders are involved. This is especially true for intelligent networked vehicles. The number of companies involved is very large. Because there are huge numbers of upstream and downstream companies in the entire industry chain, how do we ensure that these upstream and downstream companies can communicate appropriately and efficiently. Another is how to ensure that each of the different components can meet the relevant standards, all need to invest in standardization work. For example, how to complete remote applications and remote diagnosis, support, etc., these issues need to achieve the development of extended vehicle standards. Finally, let me make a brief summary. Standardization is very critical for the entire industry, especially for improving the degree of innovation, and for enhancing competitiveness. We also hope to achieve a unified test and complete industry-friendly legislation. This means that we need to set up a globally unified timeline. Only in this way can we truly help customers and the industry complete this work and task, and ultimately achieve efficient global travel. Thank you everyone, the above is all the content of my speech.
https://medium.com/@chinapotion/wicv2020-german-automobile-associations-vda-strengthens-collaboration-with-china-ecosystem-42e3b7acf6b5
['China Potion']
2020-11-13 00:21:43.378000+00:00
['Automotive', 'Autonomous Cars', 'Technology', 'Germany', 'China']
1,955
Sam Tenorio Provides an Overview of How SDN Connects Multiple Sites to its Data
Networks today are fundamental to the success of most organizations, allowing them to transmit data accurately and efficiently. As a valuable digital resource, many businesses have implemented SDN models in an effort to increase productivity and enhance daily operations. As the Founder and CEO of Seamless Advanced Solutions (“SAS”), Sam Tenorio III knows how important software-defined networking (“SDN”) is in enabling the management of complex systems less complex. As an expert in the field, he is here to outline everything you need to know about SDN. Understanding SDN Networking technology has undergone a transformation over the past decade and many software-defined networks (SDN) were developed in response to the challenges associated with fragmented network architecture. The SDN origin can be traced back to a research collaboration between Stanford University and the University of California at Berkeley, yielding the OpenFlow protocol in 2008 which enables a dynamic, programmable, efficient network configuration. Since its original inception, SDN has evolved into a reliable networking technology offered by industry leading vendors such as Cisco, VMare, Juniper, Pluribus and Big Switch. Unlike traditional networks, which are heavily dependent on hardware to manage physical connections, SDN separates the network control plane allowing network control to be centrally managed. Sam Tenorio explains that “one of the key elements in SDN is the concept of separating the control plane from the data plane. Separating these two elements is important because it enables network functions to be automated, making the architecture of SDN programmable, centrally controllable, and agile”. The Benefits of SDN An SDN provides a wide range of cost-related savings, including the reduction in hours spent on manual tasks. Sam Tenorio explains that “company’s now have the option of improved connectivity and enhanced internal communication. Ultimately, an SDN allows businesses access to software and hardware from multiple vendors resulting in a customized network services and infrastructure”. Enhanced Security SDN also enables a variety of security benefits. For example, a customer can split up a network connection between an end user and the data center and have different security settings for the various types of network traffic. It also allows businesses to selectively block malicious traffic while still allowing normal traffic flow. It can also make it easier to collect network usage information, which can support improved algorithms designed to detect network attacks. The Bottom Line SDN enables companies to automate processes that previously required an IT staff member to manually execute those processes. Automation is faster, more reliable, and cost effective. Moving forward, Sam Tenorio explains that there are multiple developments to watch for. “One is the ability to automate the provisioning of data center services, making it easier to horizontally extend access to data. Second is the ability to more easily allow customers to work across multiple domains to monitor and track what is going on across the infrastructure. Unsurprisingly, according to a recent IDC study, a premier global market intelligence firm, the worldwide data center SDN market will be worth more than $12 billion by 2022”.
https://medium.com/@samtenorio/sam-tenorio-provides-an-overview-of-how-sdn-connects-multiple-sites-to-its-data-6a212fa8516a
['Sam Tenorio Iii']
2020-12-10 18:45:32.290000+00:00
['Data Science', 'Tech', 'Technews', 'Technology', 'Data']
1,956
5 Useful Golang Modules Developers Should Know in 2021.
5 Useful Golang Modules Developers Should Know in 2021. These Go modules will save you a lot of time Go is an incredibly powerful and versatile language for multiple purposes. It’s reliable, fast, and easy to use. Golang has fantastic documentation. Go also is very easy to scale. Today, we look into the best Go modules to make working with Go even easier. So here’s a list of ten useful Go modules!
https://medium.com/better-programming/5-useful-golang-modules-developers-should-know-in-2021-340ed21177e5
['Bryan Dijkhuizen']
2020-12-21 23:18:32.355000+00:00
['Technology', 'Golang', 'Software Development', 'Programming', 'Go']
1,957
The Tale of Unidirectional Dataflow in Angular
The Lowdown on Unidirectional Dataflow Unidirectional dataflow is not a word we often see, nor is it something that organically comes up in conversations. However, it’s a neat little data management pattern that can simplify how your application responds to change. For most apps, everything eventually boils down to data. The flow of data between the different layers, such as views, components, and services, can determine the effectiveness of an application’s modularity as it grows. So what exactly is unidirectional dataflow? Unidirectional dataflow is a programming pattern that deals with how data is updated. There are two main patterns currently in use — unidirectional and bidirectional. The uni part of unidirectional refers to the idea that data can only flow in one direction. For the front end, once a view has been rendered, an action is required to change the data and re-render the entire or portion of a view from scratch. React does this by default and there is no explicit binding available to allow the view to update without direct action from the layers beneath it. Updates happen in one direction and one direction only. The data is not expected to change and is therefore considered immutable. When something happens and a change occurs, that state no longer exists and it is considered a new entity entirely by the view. Change is caused by an external action. When this happens, the view is wiped clean and re-rendered with the new data. This is unidirectional in a nutshell: a one way dataflow where data can only move in one direction and cannot backtrace in the direction it just came from. With unidirectional dataflow, there’s no need to track states because a change in data will result in a complete change in the view. Hypothetical unidirectional dataflow To understand this in an Angular context, we need to travel back in time — way back to the original iteration of Angular.js
https://medium.com/madhash/the-tale-of-unidirectional-dataflow-in-angular-5d0b451cc7aa
['Aphinya Dechalert']
2019-10-04 10:32:11.213000+00:00
['Angular', 'Web Development', 'Software Development', 'Technology', 'Education']
1,958
Amazon’s First Non-Employee Customer and What He Bought
Amazon’s First Non-Employee Customer and What He Bought The story of John Wainwright and his journey to the first purchase at Amazon. Author purchased rights via adobe stock photos Vincent Van Gogh once wrote that, “Great things are not done by impulse, but by a series of small things brought together.” In many ways, the beginnings of a business is a canvas, an idea that’s initiated with a scary, committal brushstroke, evolving into a beautiful final product, or a botched and forgotten disaster. It’s particularly interesting looking at the humble origins of ubiquitous, global companies. Their beginnings are almost always inspirationally small, the first half of their founder’s proof of perseverance. Jeff founded Amazon in 1994, walking away from a well-paying finance job in New York City. His supervisor implored him not to leave, namely because Bezos was a good employee and worth keeping, but also because he thought Jeff was on a fool’s errand. He’d seen so many great minds fall into the abyss of failed startups. Most of you know the rest of this story. But let’s go back, and see the part most don’t know: the story of their first sale.
https://medium.com/publishous/amazons-first-non-employee-customer-and-what-he-bought-fb1a07d42ced
['Sean Kernan']
2020-12-08 00:19:19.547000+00:00
['Life', 'History', 'Technology', 'Sean Kernan', 'Artificial Intelligence']
1,959
10 things I learned from Covid-19 with D3.js
10 things I learned from Covid-19 with D3.js A Guidebook to making an OpenSource project Global Timeline of Covid19 by Stacy S. Cho Overview Global Covid19 Timeline has earned a lot of 💌 👏 😍from many people. I really appreciate for everyone’s support As of now April 16, 2020, Global Covid19 Timeline has earned - 9K+ User Traffic - 58 Likes & 15 Comments from LinkedIn Post - 100+ Visits & 25+ Clone from Github repository Photo by Priscilla Du Preez on Unsplash Douglas Noel Adams, an author of The Hitchhiker’s Guide to the Galaxy once said, Programming gets things done I was actively looking for software engineer jobs in NYC before the Corona Virus swept New York & New Jersey. I had many interviews scheduled and I was very hopeful and motivated to start my career as a software engineer. The virus swept the U.S., especially New York and I received many emails and messages about interview process being pended and rejected due to the the virus. New Yorkers were asked to Stay at Home and my job hunting period stopped until further notice. For the first few days of my status being changed from Job seeker to the unemployed, John Hopkins’s Covid Website & Google’s Coronavirus (COVID-19) map became my Netflix. I was checking the graphs and the world map every hours hoping that the trend goes down and everything comes back to the normal, but the trend goes an opposite of what I hope so.
https://medium.com/javascript-in-plain-english/combining-what-i-learned-from-covid-19-with-d3-js-2c03d9df521e
['Stacy Cho']
2020-04-17 19:15:28.940000+00:00
['Technology', 'JavaScript', 'Programming', 'D3js', 'Web Development']
1,960
Python: Best Technology For Data-Driven Business Solutions
Digital transformation is fast transforming a significant number of businesses. For businesses primarily driven by data, it is crucial to becoming a remarkable digital entity, with optimized costs and efficiency. Success today is a complex process involving addressing ever-changing business and users’ needs, applying the right tech solutions, and building secure and stable applications. Technology is off-course at the frontiers and controlled by programming. However, it is crucial to choose a technology that meets your data-driven business goals and challenges to impact your growth and success significantly. This is why it is important to choose the latest and proven development technology. On the grounds of programming, there are multiple languages. However, Python technology stands out from the crowd. Python has been used for over 30 years and is still one of the most dynamically growing languages today. Regardless of your business size, Python technology is the solution to your complex business needs because it is known for its versatility. It presently powers a great list of the world’s most prominent platforms such as Microsoft, Facebook, Google, Spotify, and Uber. However, you may wonder — what makes Python different, especially for data-driven solutions? As a python development company, we’ve over time witnessed plenty of insights on the usefulness of Python. Therefore, this article talks about its many benefits and a lot more. 1. Supports Businesses of all sizes Python is an excellent choice, whether you own a small business or a conglomerate. It also drives efficiency regardless of the time pressure and helps corporations develop complex applications with the highest security standards. 2. Seamless integration Python can be easily integrated with many other frameworks, programming languages, external services, and existing infrastructure elements. Developers will also enjoy the enormous power to implement diverse internet protocols, scripting language, and data formats. In the end, you achieve a program that runs on diverse operating systems using the same byte. 3. Simplified Data Structures Python technology comes with a built-in list and advanced structures, making it incredibly easy to harness a high data typing level. That means it ultimately reduces the need for more extended code, which significantly reduces time. What’s more, data structure combinations in Python are adequate for further implementations in richer structures. The result is that you can harness Python as a powerful tool for data science. Whether for analytics, artificial intelligence, or machine learning, Python offers incredibly better data manipulation and repetitive tasks automation than other technologies. 4. Enriching ready-to-use solutions Python is so easy to read, write, learn, and debug. What’s more, python technology has a vibrant and enormous community that has created thousands of open source libraries. Therefore, you have a greater advantage of building flawless products efficiently and often without beginning from scratch. 5. Stable, secure, and battle-tested Python technology is an excellent choice for creating diverse financial applications dealing with sensitive data. It is a stable and trusted technology, well-founded on a high level of security capabilities. What’s more, Python fits so well with other web and application interfaces, thereby efficiently reducing your task. 6. Versatile, multiple usabilities, and efficient Python technology is incredibly known for its astonishingly great usage across diverse software development technologies. There is a vast range of use from operating systems to language development, prototyping, enterprise, and business applications, operating systems to applications. It is also incredibly suitable for making scientific applications and game development. Therefore, the product and functionalities you have in mind, Python, make it easier to quickly achieve your goals. Overall, Python is a high-level dynamic programming language, offering a remarkable versatility for rapid and robust development. You can harness Python’s power to meet industry practices, security needs, business goals and handle projects of practically any size. It’s no wonder companies like Microsoft, Facebook, and Google use Python for a vast range of solutions, including task automation, deep learning, machine learning, AI, data science, and lots more. We work with organizations seeking growth, and agility to build the next generation of services. Get In Touch to talk with our experts or mail to salesteam@monocept.com Source: Python Technology
https://medium.com/@monocept/python-best-technology-for-data-driven-business-solutions-f7f027502cb0
['Gangadhar Heralgi']
2020-12-26 16:45:33.843000+00:00
['Business Solutions', 'Monocept', 'Digital Transformation', 'Python Technology']
1,961
Some people might take interest. Some may support you in your quest. But at the end of the day, nobody cares, or will ever care about your dream as much
Some people might take interest. Some may support you in your quest. But at the end of the day, nobody cares, or will ever care about your dream as much Toledo vs Eastern Michigan Live Tv Nov 19, 2020·5 min read Life is a journey of twists and turns, peaks and valleys, mountains to climb and oceans to explore. Good times and bad times. Happy times and sad times. But always, life is a movement forward. No matter where you are on the journey, in some way, you are continuing on — and that’s what makes it so magnificent. One day, you’re questioning what on earth will ever make you feel happy and fulfilled. And the next, you’re perfectly in flow, writing the most important book of your entire career. https://gitlab.com/gitlab-org/gitlab/-/issues/285027 https://gitlab.com/gitlab-org/gitlab/-/issues/285028 https://gitlab.com/gitlab-org/gitlab/-/issues/285029 https://gitlab.com/gitlab-org/gitlab/-/issues/285030 https://gitlab.com/gitlab-org/gitlab/-/issues/285031 What nobody ever tells you, though, when you are a wide-eyed child, are all the little things that come along with “growing up.” 1. Most people are scared of using their imagination. They’ve disconnected with their inner child. They don’t feel they are “creative.” They like things “just the way they are.” 2. Your dream doesn’t really matter to anyone else. Some people might take interest. Some may support you in your quest. But at the end of the day, nobody cares, or will ever care about your dream as much as you. 3. Friends are relative to where you are in your life. Most friends only stay for a period of time — usually in reference to your current interest. But when you move on, or your priorities change, so too do the majority of your friends. 4. Your potential increases with age. As people get older, they tend to think that they can do less and less — when in reality, they should be able to do more and more, because they have had time to soak up more knowledge. Being great at something is a daily habit. You aren’t just “born” that way. 5. Spontaneity is the sister of creativity. If all you do is follow the exact same routine every day, you will never leave yourself open to moments of sudden discovery. Do you remember how spontaneous you were as a child? Anything could happen, at any moment! 6. You forget the value of “touch” later on. When was the last time you played in the rain? When was the last time you sat on a sidewalk and looked closely at the cracks, the rocks, the dirt, the one weed growing between the concrete and the grass nearby. Do that again. You will feel so connected to the playfulness of life. 7. Most people don’t do what they love. It’s true. The “masses” are not the ones who live the lives they dreamed of living. And the reason is because they didn’t fight hard enough. They didn’t make it happen for themselves. And the older you get, and the more you look around, the easier it becomes to believe that you’ll end up the same. Don’t fall for the trap. 8. Many stop reading after college. Ask anyone you know the last good book they read, and I’ll bet most of them respond with, “Wow, I haven’t read a book in a long time.” 9. People talk more than they listen. There is nothing more ridiculous to me than hearing two people talk “at” each other, neither one listening, but waiting for the other person to stop talking so they can start up again. 10. Creativity takes practice. It’s funny how much we as a society praise and value creativity, and yet seem to do as much as we can to prohibit and control creative expression unless it is in some way profitable. If you want to keep your creative muscle pumped and active, you have to practice it on your own. 11. “Success” is a relative term. As kids, we’re taught to “reach for success.” What does that really mean? Success to one person could mean the opposite for someone else. Define your own Success. 12. You can’t change your parents. A sad and difficult truth to face as you get older: You can’t change your parents. They are who they are. Whether they approve of what you do or not, at some point, no longer matters. Love them for bringing you into this world, and leave the rest at the door. 13. The only person you have to face in the morning is yourself. When you’re younger, it feels like you have to please the entire world. You don’t. Do what makes you happy, and create the life you want to live for yourself. You’ll see someone you truly love staring back at you every morning if you can do that. 14. Nothing feels as good as something you do from the heart. No amount of money or achievement or external validation will ever take the place of what you do out of pure love. Follow your heart, and the rest will follow. 15. Your potential is directly correlated to how well you know yourself. Those who know themselves and maximize their strengths are the ones who go where they want to go. Those who don’t know themselves, and avoid the hard work of looking inward, live life by default. They lack the ability to create for themselves their own future. 16. Everyone who doubts you will always come back around. That kid who used to bully you will come asking for a job. The girl who didn’t want to date you will call you back once she sees where you’re headed. It always happens that way. Just focus on you, stay true to what you believe in, and all the doubters will eventually come asking for help. 17. You are a reflection of the 5 people you spend the most time with. Nobody creates themselves, by themselves. We are all mirror images, sculpted through the reflections we see in other people. This isn’t a game you play by yourself. Work to be surrounded by those you wish to be like, and in time, you too will carry the very things you admire in them. 18. Beliefs are relative to what you pursue. Wherever you are in life, and based on who is around you, and based on your current aspirations, those are the things that shape your beliefs. Nobody explains, though, that “beliefs” then are not “fixed.” There is no “right and wrong.” It is all relative. Find what works for you. 19. Anything can be a vice. Be wary. Again, there is no “right” and “wrong” as you get older. A coping mechanism to one could be a way to relax on a Sunday to another. Just remain aware of your habits and how you spend your time, and what habits start to increase in frequency — and then question where they are coming from in you and why you feel compelled to repeat them. Never mistakes, always lessons. As I said, know yourself. 20. Your purpose is to be YOU. What is the meaning of life? To be you, all of you, always, in everything you do — whatever that means to you. You are your own creator. You are your own evolving masterpiece. Growing up is the realization that you are both the sculpture and the sculptor, the painter and the portrait. Paint yourself however you wish.
https://medium.com/@toledovseasternmichiganliveonn/life-is-a-journey-of-twists-and-turns-peaks-and-valleys-mountains-to-climb-and-oceans-to-explore-d90b2de3b2b7
['Toledo Vs Eastern Michigan Live Tv']
2020-11-19 00:02:19.799000+00:00
['Technology', 'Sports', 'Social Media', 'News', 'Live Streaming']
1,962
The Journey Day 10…
Reflection Eternal… Today marked the end of my first two weeks in my role as Intern for GameDevHQ. With all that’s has been learned in this last sprint I took today to really hone in on some higher level coding that I had omitted over the course of the week. This week I took a challenge set by myself. Given what I learned, could I code my way to making this game function for the next two sections on the merits of my own code? Would I be able to make my backgrounds pop and start to put my own fingerprint on the project that I was charged to make as divergent as I wanted to? The API resolve came back and the answer was a resounding “Yes!”. I had managed to make two fully functioning power-ups, the triple shot as well as a speed boost that turned up the tempo by 170%. I was able to demolish and rebuild a fully functioning and beautiful background that was actively animated and 3D. I was able to add my first sounds giving my newly created game it’s first voice. It was a neat experience to click play to run a test of my game and have music pour out of my speakers for the first time. My little engine that could began to take shape for the first time this week that didn’t feel so out of the box. Not only was I able to accomplish this, but I was able to watch quite a few Unity tutorials on Youtube and spend a concerted effort getting used to the new software tool kit that will reside in my shortcuts bar right next to it’s cousins Ableton Live 10 as well as Logic Pro X. Being so close in purpose and function, Unity is like a long lost cousin returning home. Similar DNA but I needed some time to familiarize with the particular attributes of this family member. The finalized background scene for Titan Escape With game elements in place I then went back and watch all of the lesson material to check my code solutions. I was really curious to see how clean and dry they were. There were major changes to our iteration model that introduced classes and ID’s. this involved much less code and was far more clean and concise than the ways that I had found. Even still, those were the only main differences. My logic use was sounds and based on what I did learn I was able to produce the same game results. What I did learn were tricks that good developers pick up along the way to have the most dry and clean code. Upon changing out my code to reflect the new learnings, it was time to tackle the last power up Included in the project list. The creation on a shield booster. The setup was easy and applying what I learn from the most recent lessons I was able to have it up and running within 10 mins. With working shields I was now able to focus on making my first character animation on pickup of on the power-up objects. The final piece before starting the UI. Home stretch at last. So close yet so far. but the journey is just beginning. After two weeks of crash course C# and Unity instruction I have found a very comfortable home as a game development intern. That being said, we best not forget that I am also a web development student and just finished my last unit of Javascript. With applied Javascript complete, I can now focus on learning React.js fully. What a week. With heavy coding projects every day, a Javascript final, pair coding, helping cohorts sort out the blockers while I tweaked out my own project, I was able to find the time to do some really good work and take on quite a few new learnings. I can’t wait until week three. A hui hou!
https://medium.com/@anthony-8789/the-journey-day-10-ef4a4f5c8896
['Tony Miller']
2020-12-12 01:54:19.066000+00:00
['Technology', 'Game Development', 'Unity', 'Coding', 'Web Development']
1,963
How to Set Up Your Own PaaS Within Hours
How to Set Up Your Own PaaS Within Hours Five simple steps to quickly set up your very own private PaaS environment using available free open-source technologies Photo by Fotis Fotopoulos on Unsplash There are many useful and free open-source software on the Internet, we just need to know where to look. The PaaS setup that I’m going to recommend works well for private/on-premise setup as well. There’s no coding involved, just some CLI configurations. Having your own PaaS is useful if you’re running a team of engineers requiring flexibility, privacy, and data ownership. I mainly use it for rapid prototyping purposes and hosting my own suite of web applications with minimal traffic. I would not recommend this setup for production purposes with high load unless you really know what you’re doing. I’ll keep the article short and list down the high-level steps as the instructions on the websites are very easy to follow. I’ve done a couple of setups with minimal issues and they are usually done within a few hours. Assuming you’re starting from scratch with the intent to deploy in the cloud, here are the five simple steps: Setup a cloud account and rent a VM (e.g. AWS/EC2) Register a domain from a provider (e.g. AWS Route53) Set up the PaaS by following simple steps (CapRover) Create and deploy open-source applications out of the box (Wordpress, Jenkins, GitLab, and many more) Configure backups just in case Please note that I’m not paid in any way by any of the companies listed in this article. I’m recommending them soley based on positive experiences. #1: Setup a cloud account and rent a VM You’ll need a virtual machine to host the applications. There are a variety of cloud providers that you can choose from. Here are some of the popular ones with 1-year free-tier option(s)/credits: You can click on any of the above links to set up a new free account. After creating the account, provision a virtual machine—recommended server setup to be Ubuntu 18.04 with at least 1GB RAM. #2: Register a domain name Applications will be deployed under a sub-domain, e.g. appone.apps.domain.com, apptwo.apps.domain.com, etc. so it’s essential to have your own domain. Here are some websites which I’ve used for my domains: If you’re using a cloud provider, e.g. AWS/Azure, it may be more convenient to register your domain with them to have everything managed centrally. #3: Set up the PaaS This section forms the main bulk of the setup. Although there are a number of available open-source PaaS available, e.g. Dokku, CapRover, Flynn, etc. I’ll be using CapRover as an example. I’ll further break down this portion into four steps: The steps above should get your PaaS up and running. #4: Create and deploy applications Here comes the fun part. After logging into the CapRover dashboard, navigate to the Apps screen via “Apps” at the left sidebar. Click on the “One-Click Apps/Databases” button.
https://medium.com/the-internal-startup/how-to-set-up-your-own-paas-within-hours-83356523413d
['Jimmy Soh']
2020-06-24 00:58:38.258000+00:00
['Programming', 'Technology', 'Software Development', 'Startup', 'Software Engineering']
1,964
Blockchain Introduction and Investment Themes
Can you believe someone can raise $32 billion in 60 seconds with just a Blockchain white paper and without a product out in market? Yes it can happen! Blockchain is disrupting the way in how we transact with each other. This smart, trusted, decentralised and encrypted network of computing blocks is poised to be the next disruptive wave in digital business. Blockchain provides new infrastructure to build next wave innovative applications beyond Cryptocurrencies which can have meaningful impact on people and society as whole. In partnership with Target, Sunnyvale, Women in Product hosted an event on blockchain introduction and its investment trends on May 10th. Audience got an overview of blockchain, existing/future applications, and some recent investment trends. Additionally, the barriers in adoption to blockchain technology were also examined. Our speakers consisted of product leaders in the tech industry including: TT Ramgopal, Global Head of Android Partnership Engineering, Google; Senia Maymin, Founder and CEO, Silicon Valley Change Executive Coaching; Aarthi Srinivasan, Director of Product Management, Personalisation / Machine Learning / Blockchain, Target; Satyajit Gupte, Lead Data Scientist, Target. The evening started with a lightning talk from TT Ramgopal on innovation and learnings from 11 years at Google in different roles. He shared nuggets of wisdom and experiences from Google and his day to day work experience at Google. Following his talk, Senia Maymin conducted a small workshop and discussion on ‘How to be more brave’. These two lightning talks were followed by a session on Blockchain. Here are my key takeaways from the session. Lightning talk from TT Ramgopal Be a firestarter not a firefighter Think 10x not 10% e.g Google innovation of Loons, renewable energy Default to open- share it to all- Android open source platform and it led to exponential growth of platform because developers were creating applications which increased the overall platform value Google’s approach to innovation- We didn’t invent all the ideas e.g. Google’s OKR approach was taken up from Intel, who invented it Fail fast and move ahead after doing vigorous post mortem and taking learnings forward Innovation can’t be ordained. However, you can create an environment where it will evolve organically. Focus on the user and other things will fall in place Lightning talk from Senia Maymin- How to be more brave? Start with small things and show bravery in those. Schedule meeting with yourself at 3 pm everyday and ask yourself can I be more brave now? What is the the one thing you are most afraid of? List down that one thing and you can conquer that with practice. Practice and practice till you are no longer afraid of it. Senia Maymin, PhD, MAPP, MBA, leads Silicon Valley Change Executive Coaching, a global executive coaching network for premiere companies focused on their highest performers. She is the co-author of Profit from the Positive: Proven Leadership Strategies to Boost Productivity and Transform Your Business. Blockchain Introduction, Type and platforms What is Blockchain? A Blockchain is a growing list of digital records or blocks that are secured and linked. It’s a digital ledger in which transactions are validated by a decentralised network of nodes(computers), chain linked and recorded chronologically. Each block contains: Link to previous block Timestamp Data Blockchain was first coined by Satoshi Nakamoto(Bitcoin Paper). It’s revolutionary because it creates a decentralised system and eliminates the intermediaries Types of Blockchain There are two types of blockchain- Permissioned: Restricted actors can validate block. Various methods of consensus are used e.g. Byzantine fault tolerance Permissionless: Anyone can participate or validate block. Common validation method proof of work Example: Blockchain platforms Bitcoin Ethereum Hyperledger R3 CORDA Barriers in Adoption of Blockchain Cryptocurrency is muddying waters of the blockchain technology Overvalued- Too many ICOs without real product Lack of platform standardisation and changing protocols using forks Lack of regulation & unproven value Learning curve/ Migration efforts A lot of VC money is being pumped into the Blockchain space. $4 Billion funding has already happened and it is expected to grow by $20 Billion by 2024. Applications of Blockchain Unified User Identity Regulatory compliance Decentralised compute and storage Lending platforms and cryptocurrencies AI & iOT applications Supply Chain What is Bitcoin ming and what are steps in Bitcoin Protocol? Bitcoin mining is the process by which transactions are verified and added to the public ledger, known as the blockchain. Bitcoin mining is also the means through which new bitcoin are released. Steps involved: Start, broadcast new transaction Verification: User Signature & Funds Proof of work: Prevent double spending Mining: Earn bitcoin rewards Recheck transactions & start new block Interesting facts about Bitcoin:
https://medium.com/women-in-product-blogs/blockchain-introduction-and-investment-themes-ca44a91af93f
[]
2018-07-31 22:50:52.883000+00:00
['Innovation', 'Blockchain', 'Cryptocurrency', 'Future Technology', 'Bitcoin']
1,965
Building Instacart Pickup
Building Instacart Pickup By: Kevin Henrikson I’m driving with my family from San Mateo to Juniper Lake for a camping trip. I need to stock up my cooler on the way, but I want to make sure we get to our campsite before it gets dark — it’s a long drive. I don’t have the time to shop at my regular grocery store, so I open up my Instacart app to see if any store along my route is offering curbside pickup with the new Instacart Pickup product. I pick my store of choice along Interstate 80, fill up my cart, and follow the in-app prompts to my designated pickup parking spot. The experience is powered by a ton of new features we packed into the app including: “Pick Your Pickup” UX that lets customers easily view and select the pickup location most convenient to their route. that lets customers easily view and select the pickup location most convenient to their route. Smart Storefronts that adapt to location-specific store inventories. that adapt to location-specific store inventories. Opt-in, location-based alerts that allow customers to let the shopper know when they’re on the way and getting close. that allow customers to let the shopper know when they’re on the way and getting close. Customized navigation and prompts which sends customers to the mapping app of their choice, automatically directing drivers from their current location to the store. As a customer nears a store, we send them prompts in real-time to guide them to the handoff location. which sends customers to the mapping app of their choice, automatically directing drivers from their current location to the store. As a customer nears a store, we send them prompts in real-time to guide them to the handoff location. Send a Friend or Family Member which lets customers share their order details with friends and family to delegate pickup to another driver. Place the order and send a friend to pick it up! While Pickup and Delivery products share a lot of the same technological DNA, Pickup flips the script of the typical delivery service model, folding the customer into the fulfillment chain’s last mile. When you involve the customer in the fulfillment flow, it presents a bunch of really interesting technical challenges for engineering and product teams. Building a dynamic storefront In our Delivery product, we identify a store location to deliver from based on the items a customer puts in their cart. With Pickup, customers choose their exact store — not just the actual retailer but the exact physical location of the store they want to pick up from. This prompted us to build a storefront that changes from location-to-location to reflect store-specific inventories. When a pickup customer selects a store along their route, they’ll see location-specific inventories as they fill up their cart. Our Item Availability model underpins this dynamic storefront. It relies on a combination of historical found/not-found rates, inventory data, and other inputs to predict whether the customer’s preferred items are in stock at any given store location along the customer’s route. Sticking the landing In contrast to the passive, one- and two-hour delivery windows available in our Delivery Product, with Pickup, customers participate in a choreographed curbside handoff. Even the smallest hitches and delays are magnified if you’re on the road with a kid in the car seat behind you, trying to get to a campsite before dark. We rely on a host of machine learning models, geolocation technologies, and in-app prompts to nail the second-by-second hand-off. In-app prompts guide customers to designated curbside pickup areas. Once a customer selects their store of choice and completes their order, our Drive Time model looks at historical transit times to estimate how long it will take a customer to arrive at the store. We’ve traced rings of geofences around each pickup-eligible store. If a customer with location-based alerts enabled crosses the fence, we’ll adjust the minute-to-minute arrival time. We rely on historical fulfillment data to calculate just how long it can take someone to get in and out of store parking lots — we like to call this the “Parking Model”. This is especially useful for pickup customers in crowded urban centers vs. customers in suburban or rural stores with wide-open parking lots. From the data, we can see it takes longer to get through a parking lot on a crowded Sunday in downtown Los Angeles than it does in suburban Orange County, CA. Even though Pickup and Delivery are two different products and experiences, the orders are being fulfilled by the same set of shoppers, and we’ve built an Order Prioritization model to choreograph the hand-off. The model organizes incoming orders in a shopper’s queue based on historical fulfillment data. If a customer crosses a geofence, we will bump the pickup order to the top of a shopper’s queue, alerting the shopper that it’s time to make the handoff.
https://tech.instacart.com/building-instacart-pickup-705a406b6350
[]
2020-01-14 18:41:34.207000+00:00
['On Demand', 'Mobile App Development', 'Technology News', 'Machine Learning', 'Fulfillment']
1,966
TikTok Is Putting Children at Risk
TikTok Is Putting Children at Risk With 18 million users under the age of 14, age restrictions need to be raised immediately. Photo by Aaron Weiss on Unsplash By the end of 2021, TikTok expects to surpass 50 million users in the United States alone. While the popular platform reigns supreme among the American youth, nearly 30% of users are under the age of 18, despite TikTok’s meager attempts to set an age limit. Without any effort on the platform’s behalf to verify the ages of its creators, many young and impressionable children find themselves amongst an abundance of video content, most of which is geared towards an older audience. Though TikTok’s Community Guidelines clearly state that no inappropriate material, such as sexual language or scenarios will be tolerated, one only has to spend a matter of time on the app before sexually suggestive content reveals itself. The worst part? Not only is this content completely accessible to minors, but some of it is also even created by minors themselves. If TikTok doesn’t instill stronger guidelines to protect minors and quickly, there’s no telling how detrimental the results may be. The Ever-Growing Presence of Youth on Social Media The user base for social media is getting younger and younger each year. According to an article published on BBC, around 50% of all children aged ten own or have access to a personal smartphone. As many social media platforms only require a birthdate in order to sign up, accessing content under the age of thirteen couldn’t be easier. In order to combat inappropriate usage by minors, TikTok released parental controls that allow grownups to filter content and monitor their children’s accounts. These controls include both a screentime limit and a restricted mode, which supposedly blocks mature content. However, TikTok has proven to be busier blocking content that follows community guidelines as opposed to removing inappropriate material, making their so-called “restricted mode” virtually useless. Additionally, the majority of inappropriate content on TikTok isn’t obvious to the technology’s filters. This often manifests in dancing trends not appropriate for minors or vulgar and sexual lyrics to popular sounds. Therefore, even when monitoring your child’s usage of the platform, it’s still very likely that they will be exposed to content geared towards older audiences. The Problem With Influencers Each and every social media platform comes with its own influencers — popular creators that have significant followings and often end up creating a career out of their social media presence. The most striking example is that of YouTubers, who once reigned supreme in the world of internet content. While YouTube creators were typically in their early twenties to thirties, TikTok influencers are earning their fame at as young as 13. Most notably, Charli D’Amelio recently skyrocketed to just under 100 million followers at the age of 15. Other young creators include Addison Rae, Loren Gray, and Zoe Laverne, all of whom are under the age of twenty. As social media influencers get younger, so do their fanbases. A 28-year old fashion vlogger typically attracts a different crowd than a 15-year old dancer on TikTok. However, there is a clear difference in maturity levels between these two. While the 28-year old influencer may have a better idea as to how to present herself as a role model, the 15-year old is still a child herself and cannot be expected to set a good example for her fans. The problem with influencers is a topic in and of itself, yet the danger to minors on TikTok goes hand-in-hand with the rise of younger stars. TikTok stars have typically risen the ranks by participating in dance trends, often created by older users. Many of these dance trends are sexually suggestive and performed to explicit songs, such as “WAP” by Cardi B and “3 Musketeers” by ppcocaine. While these dance trends could be considered harmless fun, performing them for an impressionable following of young users poses two problems: the sexualization of the minor performing the dance, and the audience’s desire to participate in the trend itself. Although not every dance on TikTok is inherently sexual, there are many involving suggestive gestures that should not be recreated by children. Pedophilia on TikTok Not only is TikTok exposing minors to inappropriate content, but it’s also a hotbed for pedophiles due to the massive amount of younger users. TikTok does have privacy settings; users can set their accounts to private and manually approve followers. However, with the rise of young TikTok influencers like D’Amelio, many users want to give themselves the best possible chance to be discovered. A public account may attract a bigger audience, but this isn’t always desirable. In fact, there is no real security keeping convicted sex offenders and pedophiles off the platform. With the ever-growing presence of children on the app, more and more opportunities arise for these offenders to get their hands on inappropriate content concerning minors. This argument is not new to the conversation surrounding TikTok. Plenty of articles have been posted calling out the app’s allowance of minors and pedophiles, but no real action has been taken on the side of the company. Instead, the responsibility of protecting the minor is left up to the parent, who in many situations has no knowledge of their child’s usage of the app. The Solution Age restriction on social media platforms is difficult to enforce, but it is incredibly important to have. Due to the nature of the content posted on TikTok, the app should consider raising its age limit from 13 to 16. Though younger minors could technically still get around this restriction, it may stop young influencers from finding success and therefore discourage children from using the app in the first place. Secondly, parents should be spending more time monitoring their children’s social media presence. The conversation surrounding the type of content posted on TikTok needs to be louder, as many parents are unfamiliar with the app and therefore don’t see a problem in allowing their children to use it without supervision. This very lack of conversation could be putting our children in danger, without us realizing it. Social media, while highly regarded as a way to stay connected, is one of the most dangerous places for children to be. TikTok is no exception. Otherwise, we may find ourselves creating an online world that will ultimately be our downfall.
https://medium.com/swlh/tiktok-is-putting-children-at-risk-dc624a4a3ecd
['Natalie Vinh']
2020-12-06 22:40:11.894000+00:00
['Influencers', 'Children', 'Technology', 'Tiktok App', 'Social Media']
1,967
20 Crowdfunding gadgets that simply blew our minds
Remember the groundbreaking Pebble ePaper watch on Kickstarter? It blew up because of its cool concept and product quality. Since then, crowdfunding campaigns have delivered so many new product concepts that we’ve never seen before. So we’re presenting 20 amazing crowdfunding gadgets that wowed us. Over the years, we have come across many crowdfunding campaigns with outstanding product concepts. Recently, things have become even better-and we’ve seen some of the most amazing crowdfunding gadgets that you’ve never seen before. From personal holographic displays to couch consoles, these crowdfunding gadgets are undoubtedly cool and useful. Related: 10 Futuristic tech gadgets-rolling TVs, VR treadmills, and more ​So today’s daily digest highlights 20 amazing crowdfunding gadgets you can easily add to your life. You’re bound to like these for sure. Bird Buddy Smart Bird Feeder The Bird Buddy Smart Bird Feeder actually notifies you when birds arrive. It then snaps a photo and organizes them into a beautiful collection. You’ve never been this close to your feathered friends with any of the other crowdfunding gadgets out there. Lumos Ultra LED Bike Helmet The Lumos Ultra LED Smart Helmet protects you during your twilight and nighttime rides. It comes with integrated LED lighting, turn signals, and other smart features that keep you visible and predictable to other drivers. It’s one of the most practical crowdfunding gadgets we’ve seen. Vaonis Vespera Lightweight Telescope With the Vaonis Vespera Lightweight Telescope, you can stargaze in your own backyard or during your outdoor adventures. Use it to capture photos of galaxies, nebulae, the moon, and more. Looking Glass Portrait It sounds incredible, but the Looking Glass Portrait is a personal holographic display. Just capture a photo with any camera, and this device lets you transform it into a realistic hologram. It can even move and talk in tandem with you. ROLI LUMI Portable Beginner’s Keyboard Want to learn how to play the piano? The ROLI LUMI Portable Beginner’s Keyboard lets you play your favorite songs faster since each key lights up with the exact note you need to play. Artiphon Orba Handheld Musical Instrument If you’re a musician, you’re going to love the Artiphon Orba Handheld Musical Instrument. It works intuitively as a synth, loop, and controller. Best of all, it fits in the palm of your hand so you can take it anywhere. Knocki Make Any Surface Smart Device The Knocki Make Any Surface Smart Device makes the surfaces around you smarter. Just attach it to a surface and watch it transform into a touch interface. You have to see it to believe it. Rotrics DexArm Robot Desk Arm The Rotrics DexArm Robot Desk Arm will turn your desktop into a workshop. Its capabilities include drawing, laser engraving, 3D printing, and more. It’s one of our favorite crowdfunding gadgets for creative professionals. Lofelt Basslet Watch-Size Subwoofer Why listen to your music when you can feel it? The Lofelt Basslet Watch-Size Subwoofer gives you powerful bass right on your body for an immersive music experience. Square Off NEO & SWAP Automated AI Board Games For a cool table game experience, check out the Square Off NEO & SWAP Automated AI Board Games. The pieces move on their own thanks to AI, and the games adapt to your skill level. It’s one of the most fun crowdfunding gadgets out there. Reevo Hubless eBike Hands down, the Reevo Hubless eBike is one cool bike. Its wheels are spokeless, so it looks pretty futuristic. With three drive modes-pedal, adaptive, and throttle-it takes you up to 37 miles in style. Ebite Inc. Couch Console The Ebite Inc. Couch Console is the perfect accompaniment to your gaming nights and Netflix binges. This modular couch tray has compartments for your remote, snack, and cups. It also has a charging outlet. BeYou Transforming Chair The BeYou Transforming Chair is one of our must-have crowdfunding gadgets for your home office. It lets you focus and relax in ten different positions so that you can work comfortably, no matter how you like to sit. Mendi Brain Training Headband Stretch your brain power with the Mendi Brain Training Headband. This wellness gadget gives you access to fun, stimulating brain exercises that help you focus and relax. It uses the same technique used by pro athletes, executives, and astronauts. OmniCharge Omni Off-Grid Portable Camping Power Station When you’re off-grid, you still need power. And the OmniCharge Omni Off-Grid Portable Camping Power Station can help. It comes with 16 different power ports and boasts a 1,500 Wh battery. That should be plenty for your weekend camping trip. VAVA Portable SSD Touch Secure External Storage Want an incredibly secure SSD? The VAVA Portable SSD Touch Secure External Storage is the crowdfunding gadget you’ve been looking for. It has fingerprint and AES 256-bit encryption, so nobody’s getting to your files but you. WooBloo SMASH Portable Smart Projector Create a theater-like experience for your movie nights with the WooBloo SMASH Portable Smart Projector. Set it up in your living room, bedroom, or backyard for an incredible project. You can command it via Alexa, too, for convenience. HeimVision Assure B1 2K Ultra HD Security Camera Get all the home security features you need with the HeimVision Assure B1 2K Ultra HD Security Camera. This sleek little camera is weatherproof, has two-way audio, boasts a 130-degree field of vision, offers voice assistance, and more. Recharge FlexBEAM Therapy Device Post-workout sore muscles are painful. But you can quicken your body’s recovery process with the Recharge FlexBEAM Therapy Device. It targets infrared light on overworked areas of the body to turbocharge the healing process. ORA GQ Graphene Headphones And the ORA GQ Graphene Headphones deliver quite an audio experience. They use graphene, which is a Nobel prize-winning material that disperses sound throughout the driver for premium quality. It also gives you a 70% increase in battery life. These crowdfunding gadgets are truly impressive. When companies and their customers work together to create a new product, the result is nothing short of amazing. Which of these gadgets do you want to get? Let us know in the comment section. Want more tech news, reviews, and guides from Gadget Flow? Follow us on Google News, Feedly, and Flipboard. If you use Flipboard, you should definitely check out our Curated Stories. We publish three new stories every day, so make sure to follow us to stay updated! The Gadget Flow Daily Digest highlights and explores the latest in tech trends to keep you informed. Want it straight to your inbox? Subscribe ➜
https://medium.com/the-gadget-flow/20-crowdfunding-gadgets-that-simply-blew-our-minds-e869441a92e8
['Gadget Flow']
2021-02-15 17:35:26.043000+00:00
['Crowdfunding', 'Kickstarter', 'Indiegogo', 'Gadgets', 'Technology']
1,968
The Most Popular Car in the World:
The Future Car There are many car companies which are sold in millions. But, we see especially that which car is the most popular and the most sold in the World. That is……………….. Read more on The Most Popular Car in the World:
https://medium.com/@alimurtaza110/the-most-popular-car-in-the-world-7b29f5e006de
['Ali Murtaza']
2020-12-24 11:03:14.532000+00:00
['Technology', 'News', 'Innovation', 'Cars', 'Future']
1,969
Independent Component Analysis (ICA) In Python
Suppose that you’re at a house party and you’re talking to some cute girl. As you listen, your ears are being bombarded by the sound coming from the conversations going on between different groups of people through out the house and from the music that’s playing rather loudly in the background. Yet, none of this prevents you from focusing in on what the girl is saying since human beings possess the innate ability to differentiate between sounds. If, however, this were taking place as part of scene in a movie, the microphone which we’d use to record the conversation would lack the necessary capacity to differentiate between all the sounds going on in the room. This is where Independent Component Analysis, or ICA for short, comes in to play. ICA is a computational method for separating a multivariate signal into its underlying components. Using ICA, we can extract the desired component (i.e. conversation between you and the girl) from the amalgamation of multiple signals. Independent Component Analysis (ICA) Algorithm At a high level, ICA can be broken down into the following steps. Center x by subtracting the mean Whiten x Choose a random initial value for the de-mixing matrix w Calculate the new value for w Normalize w Check whether algorithm has converged and if it hasn’t, return to step 4 Take the dot product of w and x to get the independent source signals Whitening Before applying the ICA algorithm, we must first “whiten” our signal. To “whiten” a given signal means that we transform it in such a way that potential correlations between its components are removed (covariance equal to 0) and the variance of each component is equal to 1. Another way of looking at it is that the covariance matrix of the whitened signal will be equal to identity matrix. Identity Matrix Covariance Matrix The actual way we set about whitening a signal involves the eigen-value decomposition of its covariance matrix. The corresponding mathematical equation can be described as follows. where D is a diagonal matrix of eigenvalues (every lambda is an eigenvalue of the covariance matrix) and E is an orthogonal matrix of eigenvectors Once we’ve finished preprocessing the signal, for each component, we update the values of the de-mixing matrix w until the algorithm has converged or the maximum number of iterations has been reached. Convergence is considered attained when the dot product of w and its transpose is roughly equal to 1. where Python Code Let’s see how we can go about implementing ICA from scratch in Python using Numpy. To start, we import the following libraries. import numpy as np np.random.seed(0) from scipy import signal from scipy.io import wavfile from matplotlib import pyplot as plt import seaborn as sns sns.set(rc={'figure.figsize':(11.7,8.27)}) Next, we define g and g’ which we’ll use to determine the new value for w. def g(x): return np.tanh(x) def g_der(x): return 1 - g(x) * g(x) We create a function to center the signal by subtracting the mean. def center(X): X = np.array(X) mean = X.mean(axis=1, keepdims=True) return X- mean We define a function to whiten the signal using the method described above. def whitening(X): cov = np.cov(X) d, E = np.linalg.eigh(cov) D = np.diag(d) D_inv = np.sqrt(np.linalg.inv(D)) X_whiten = np.dot(E, np.dot(D_inv, np.dot(E.T, X))) return X_whiten We define a function to update the de-mixing matrix w. def calculate_new_w(w, X): w_new = (X * g(np.dot(w.T, X))).mean(axis=1) - g_der(np.dot(w.T, X)).mean() * w w_new /= np.sqrt((w_new ** 2).sum()) return w_new Finally, we define the main method which calls the preprocessing functions, initializes w to some random set of values and iteratively updates w. Again, convergence can be judged by the fact that an ideal w would be orthogonal, and hence w multiplied by its transpose would be approximately equal to 1. After computing the optimal value of w for each component, we take the dot product of the resulting matrix and the signal x to get the sources. def ica(X, iterations, tolerance=1e-5): X = center(X) X = whitening(X) components_nr = X.shape[0] W = np.zeros((components_nr, components_nr), dtype=X.dtype) for i in range(components_nr): w = np.random.rand(components_nr) for j in range(iterations): w_new = calculate_new_w(w, X) if i >= 1: w_new -= np.dot(np.dot(w_new, W[:i].T), W[:i]) distance = np.abs(np.abs((w * w_new).sum()) - 1) w = w_new if distance < tolerance: break W[i, :] = w S = np.dot(W, X) return S We define a function to plot and compare the original, mixed and predicted signals. def plot_mixture_sources_predictions(X, original_sources, S): fig = plt.figure() plt.subplot(3, 1, 1) for x in X: plt.plot(x) plt.title("mixtures") plt.subplot(3, 1, 2) for s in original_sources: plt.plot(s) plt.title("real sources") plt.subplot(3,1,3) for s in S: plt.plot(s) plt.title("predicted sources") fig.tight_layout() plt.show() For the sake of the proceeding example, we create a method to artificially mix different source signals. def mix_sources(mixtures, apply_noise=False): for i in range(len(mixtures)): max_val = np.max(mixtures[i]) if max_val > 1 or np.min(mixtures[i]) < 1: mixtures[i] = mixtures[i] / (max_val / 2) - 0.5 X = np.c_[[mix for mix in mixtures]] if apply_noise: X += 0.02 * np.random.normal(size=X.shape) return X Then, we create 3 signals, each with its own distinct pattern. n_samples = 2000 time = np.linspace(0, 8, n_samples) s1 = np.sin(2 * time) # sinusoidal s2 = np.sign(np.sin(3 * time)) # square signal s3 = signal.sawtooth(2 * np.pi * time) # saw tooth signal In the proceeding example, we compute the dot product of the matrix A and the signals to obtain a combination of all three. We then use Independent Component Analysis to separate the mixed signal into the original source signals. X = np.c_[s1, s2, s3] A = np.array(([[1, 1, 1], [0.5, 2, 1.0], [1.5, 1.0, 2.0]])) X = np.dot(X, A.T) X = X.T S = ica(X, iterations=1000) plot_mixture_sources_predictions(X, [s1, s2, s3], S) Next, we use ICA to decompose a mixture of actual audio tracks and plot the result. If you’d like to try it yourself, you can get the audio samples here. I encourage you to actually try listening to the different audio tracks. sampling_rate, mix1 = wavfile.read('mix1.wav') sampling_rate, mix2 = wavfile.read('mix2.wav') sampling_rate, source1 = wavfile.read('source1.wav') sampling_rate, source2 = wavfile.read('source2.wav') X = mix_sources([mix1, mix2]) S = ica(X, iterations=1000) plot_mixture_sources_predictions(X, [source1, source2], S) wavfile.write('out1.wav', sampling_rate, S[0]) wavfile.write('out2.wav', sampling_rate, S[1]) Sklearn Finally, we take a look at how we could go about achieving the same result using the scikit-learn implementation of ICA. from sklearn.decomposition import FastICA np.random.seed(0) n_samples = 2000 time = np.linspace(0, 8, n_samples) s1 = np.sin(2 * time) s2 = np.sign(np.sin(3 * time)) s3 = signal.sawtooth(2 * np.pi * time) S = np.c_[s1, s2, s3] S += 0.2 * np.random.normal(size=S.shape) S /= S.std(axis=0) A = np.array([[1, 1, 1], [0.5, 2, 1.0], [1.5, 1.0, 2.0]]) X = np.dot(S, A.T) ica = FastICA(n_components=3) S_ = ica.fit_transform(X) fig = plt.figure() models = [X, S, S_] names = ['mixtures', 'real sources', 'predicted sources'] colors = ['red', 'blue', 'orange'] for i, (name, model) in enumerate(zip(names, models)): plt.subplot(4, 1, i+1) plt.title(name) for sig, color in zip (model.T, colors): plt.plot(sig, color=color) fig.tight_layout() plt.show() The accompanying Jupyter Notebook can be found here.
https://towardsdatascience.com/independent-component-analysis-ica-in-python-a0ef0db0955e
['Cory Maklin']
2019-08-22 14:32:53.200000+00:00
['Machine Learning', 'Technology', 'Artificial Intelligence', 'Data Science', 'Programming']
1,970
Meet Stacey, Operations Coordinator at Arkafort
What is your role at Arkafort? At Arkafort, I am the Operations Coordinator which basically means that I am in charge of liaising between clients and the various divisions at Arkafort. I usually deal with sales and tenders, invoicing, scheduling, project planning and management, HR and coordinating schedules to cater for our 24/7 services. In a nutshell, I make sure that all operations are running smoothly and that our clients are ultimately happy and satisfied with the service provided. How would you describe your experience at Arkafort so far? I describe the past 6 years as exciting, dynamic and unpredictable. It is really hard to put this experience into words. I first started working with SmartICT (before the company became Arkafort in 2017) as an administration clerk. I remember the excitement of those first days at work, especially when I used to be involved in meetings. I used to be so anxious! I am definitely past that phase now, and meetings have become a regular (and most interesting!) in my daily routine. 😊 I must admit that Arkafort allowed me to grow and develop as a person. And not just on a professional level, but also on a personal level. It changed my perspective on life. The various encounters, experiences and challenges we faced have deeply affected the way I look at things now. I now know how to handle things better and learned how to cope with the difficult situations in life. For me, Arkafort is not just a workplace, it’s my second home. Arkafort Cloud Launch 2020 How would you describe the values and priorities of Arkafort as a company? I would say that our main priority is our clients. Arkafort focuses on solutions in order to make things easier for the clients and address, manage and simplify any IT-related aspect of their work. Loyalty and commitment are important values for the company and for the team. I really like the fact that we have always managed to keep our feet firmly on the ground as a company. And that we work as one professional team. How do you ensure customer satisfaction? As I mentioned before, customer satisfaction is our top priority. We like to ensure that we build a relationship with our clients and work hard on keeping a good, constant and open channel of communication. At Arkafort we like to move away from transactional services and focus on what the company can do for your business and on how best to support and the individual needs and specifications of each and every company we work with. We therefore cater for our clients’ specific needs and adapt our solutions accordingly. Our role is also to make sure that we have the best deals with our suppliers, so as to provide our customers with high-quality products at reasonable prices. Is there a particular memory at Arkafort which you cherish? Every day brings about new challenges, connections and projects and I treasure every moment at Arkafort. One particular experience which I will always cherish, is the participation at ICE 2019 in London. It was my very first business trip with Arkafort and I truly enjoyed representing the company there. Gordon (left), Stacey and Justin at ICE London 2020 Which Arkafort projects would you describe as most exciting and innovative in terms of operations? Definitely the coordination of the 24/7 operation. It is really challenging but also very exciting. Here is where I try to balance resources efficiently and effectively to support clients in their respective time-zones and maintain a well-balanced operational system. I am also quite excited about the new and innovative cloud platform, as this will give a new dimension to the company and to the services we offer. It will make things easier for both clients and company and will enable us to be even more efficient and straightforward in terms of addressing clients’ needs. What’s next? At this point I would like to see the company grow even further. I look forward to being part of more engaging and innovative new projects!
https://medium.com/arkafort/meet-stacey-operations-coordinator-at-arkafort-f31b3216b853
['Elaine Falzon']
2020-12-03 18:34:25.071000+00:00
['Women In Tech', 'Technology', 'Team']
1,971
A Show of Force, and Humanity, on the Front Lines of Last-Mile Delivery
A little over a month ago, I walked into the office of Eclipse Ventures, ready to embark on the next chapter of my investing career. The email I received from the team said I’d be kicking off my tenure with an on-boarding process. I anticipated the usual transitionary period filled with HR chores to familiarize myself with the inner workings of the fund, and the likelihood of a few early projects to cut my teeth on. The “quiet before the storm” indeed. What transpired over the next four weeks was nothing short of a Category 5 health and economic hurricane — and the most powerful on-boarding I’ve ever experienced. If you ever want to measure the true character of a leader or organization, pay attention during moments of crisis. Over the last month, I’ve watched our founders hustle and adapt to unprecedented turbulence. I’ve witnessed the Eclipse community come together to support people they’ve never met, and I’ve seen everyone on my team — each one of us — step up to help the portfolio. Now, one month in, I have more conviction than ever that industry verticals like logistics, manufacturing and supply chain need to be digitally transformed. This isn’t an academic hypothesis. For many of us, the struggle to obtain household essentials is a daily endeavor. Hospitals can’t get the supplies they need, and the most vulnerable members of our society can’t access basic necessities because our existing infrastructure was built for an analog era. The entire logistics industry, from global freight shippers to regional delivery companies, is being put to the test by the pandemic’s massive disruption to supply chains and the deluge of online orders it has unleashed on retailers for everyday items. It’s a test we’re largely failing. Established leaders like UPS and FedEx are struggling to keep up, as are the newer app-based services like Instacart and Good Eggs. And despite Amazon’s continued dominance, things aren’t looking so prime time right now. That said, delivery companies with a technological core that can scale up rapidly, agile operations, and the best people in place are meeting this unprecedented moment — if not in perfect stride — then certainly with intense hustle and conviction. That’s what’s happening at Bay Area-based AxleHire, which has seen its same- and next-day deliveries grow by double-digit percentages with each passing week. It’s currently delivering more than 50,000 time-sensitive packages a week to people’s homes: groceries, medications, meal kits and other vital goods. News video of AxleHire delivering free meals to hospital workers, which the company has committed to doing weekly for now. (Credit: KTVU) High-profile brands using AxleHire include meal-kit makers HelloFresh and Blue Apron and retailers as large as Ikea. The 5-year-old company’s bread-and-butter clients are grocers, pharmacies and e-commerce businesses in major cities up and down the West Coast, as well as in New York City — the regions hit hardest by coronavirus thus far. “Right now, we’re one of the most flexible logistics companies in terms of handling this,” says Daniel Sokolovsky, Founder and CEO of AxleHire. “Our clients are praising us. As we speak, UPS and FedEx are refusing to deliver their packages.” Indeed, logistics companies — and especially last-mile services like AxleHire — are playing an increasingly central part in sustaining economies that have been crippled by the pandemic. And just like Amazon and Walmart, both of which are collectively hiring hundreds of thousands of workers in the current crisis, AxleHire is also giving the rapidly growing numbers of unemployed the chance to generate much-needed income. Over the past month, the company has brought thousands of new drivers and warehouse workers onto its platform. AxleHire expects to add 1,000 workers here in the Bay Area alone, Daniel told a local TV station. As Daniel describes it, AxleHire is a company that is one-half software and one-half operations. People sign up to be drivers through its app, which also is how retailers request deliveries to their customers. The software takes care of the rest, from paying drivers and tracking mistakes to automating backend processes like sorting packages and sequencing routes. The technological core allows for rapid scale up, and people are brought onto the platform or into headquarters to fill in any gaps not covered by the software. And right now, there’s plenty of physical work to go around. But the current crisis is also a time to innovate. One of the newest features AxleHire has launched is a no-touch delivery option, which allows both the driver and customer to minimize human contact by verifying receipt and identity via the app. Customers can now give drivers instructions such as where to leave a package or what access code to use to enter a gated residence. It’s not that the pandemic caused the company to shift its business strategy. It has had the opposite effect: accelerating initiatives that were already planned. And this has only strengthened Daniel’s belief in AxleHire’s importance to the economy and the community. There was no greater proof of that than when the Bay Area’s first shelter-in-place orders were announced last month. This prompted Daniel to pull up the ordinance online to read it for himself. Actually, he took a screenshot — because there, at the top of the order, read a phrase that described AxleHire to a T. Among the entities deemed essential to people’s public health, the first line read: “Businesses that ship or deliver groceries, food, goods or services to residences.” Not only has this crisis underscored the importance of logistics in modern life, it is catalyzing innovation across essential industries. It has also intensified our desire to work alongside ambitious founders at a time when we truly need their skills, passion and tenacity. Let’s all roll up our sleeves and transform the verticals that need it the most.
https://medium.com/eclipse-ventures/a-show-of-force-and-humanity-on-the-front-lines-of-last-mile-delivery-1af51127de4
['Matt Mulvey']
2020-04-14 18:57:28.259000+00:00
['Technology', 'Grocery Delivery', 'Logistics', 'Startup', 'Coronavirus']
1,972
How to Get an Arduino and Your Smartphone to Talk to Each Other
How to Get an Arduino and Your Smartphone to Talk to Each Other A comparison of various Arduino communication methods Photo by the author. Created with Canva. Imagine being able to communicate with any device from anywhere in the world. You can control your TV, lights, washing machine, dishwasher, toaster, coffee machine — anything you want — just for a few bucks. You can do it with Arduino. I have a lot of these lying around in my room, ready to automate another device of mine. In one of my previous articles, I covered how you can use Arduino as a switch to turn any device on or off. Today, I will discuss the pros and cons of various wireless methods used to create communication between Arduino and smartphones. This would not only allow you to use your phone as a switch, but it would also enable it to send specific commands for operations and get the state from Arduino.
https://medium.com/better-programming/how-to-get-arduino-and-smartphone-talk-to-each-other-9abaf39d2ff6
['Shubh Patni']
2020-11-02 15:58:25.164000+00:00
['IoT', 'Programming', 'Technology', 'Arduino', 'Software Development']
1,973
Discover How Mediflic is Making the World a Healthier Place
By Angie Gallagher In this series, we feature the stories of the companies who are using the BurstIQ platform to further their mission of providing new and innovative health-tech solutions to patients around the world. Today we had the pleasure of talking with Mediflic’s CEO Diptiman Chattopadhyay and CTO Sayak Dhar. Introducing Mediflic, a health-tech startup based in India. Mediflic from Winsome IT Solutions India Pvt. Ltd is an online healthcare service providing a portal that acts as a bridge between the patients and the doctors. It is centered around giving medical services at doorsteps to individuals everywhere throughout the world. Digital transformation has been making a tremendous effect on the medical industry during the late decades. The growing popularity of the beneficial patient treatment delivery system prompted a few innovations in the medical sector. These days, the pace of utilizing the most recent technologies in the healthcare industries are expanding step by step. Mediflic is the result of digital transformation that has been making a progressive change in the Indian medicinal services industry with the usage of different technologies. How will your company improve health access & equity while making the world a healthier place? Diptiman Chattopadhyay, CEO of Mediflic In this one-stop medical destination, patients from all over the world can look out for their necessary medical treatment programs in India. The patient can locate the best clinic or hospital for his/her prerequisite, check the reviews and specialties of the clinics/ hospitals, and can experience the detailed treatment strategies virtually. Through this portal, patients can choose their favored specialist, speak with him/her, do the important. Diagnosis and send him/ her the reports. The specialists will check the reports and guide the patient about further treatments. All these can be done virtually on this portal. Patients don’t have to visit the doctor from time to time to get treated. From finding the best specialists to getting restored, all these will be possible online on this site. How does the BurstIQ (BIQ) platform help your application? Sayak Dhar, CTO of Mediflic With organizations like BurstIQ bringing blockchain advancements to the quickly developing well-being industry, we are certain that blockchain will make a critical, conceivably transformative effect in this industry. Presently, there is no other industry that needs transformation more than medicinal services. Computerized medical care organizations can get to new information sources to upgrade AI-empowered items and showcase those items to focused and engaged people. Apart from this, and healthcare systems can improve clinical results and lessen cost through AI-empowered analytics, customized post-acute administrations, better care coordination, etc. What are your final thoughts on an aligned mission to help the world? Mediflic is here to make healthy living feasible for everybody. We want to make living healthy possible for people around the world. We bring world-class healthcare into our patients’ homes and intend to make medical care progressively accessible, yet additionally moderate and accountable to our patients’ needs. We have the vision to make all aspects of medicinal services available from anyplace, with improved access from your own cell phone/ tablet/ PC or laptop. Mediflic aims to become a people care centric stage, wherein everybody quickly gets to know about the vast opportunities of Indian healthcare services for their and their family’s physical, psychological well-being. We are on a mission to improve the medical care experience for all humanity. We are an organization that is guided by qualities and driven by our plan to perform incredibly. We want to make a sound living of propensity for individuals around the world.
https://medium.com/@burstiq/discover-how-mediflic-is-making-the-world-a-healthier-place-a2efa833d9bd
[]
2020-10-27 15:35:22.105000+00:00
['Health Technology', 'Blockchain', 'India', 'Blockchain Startup', 'Healthcare']
1,974
How to Secure a Spring Boot Application with TLS
Creating a Spring Boot Application In this section, we will create a Spring boot application and expose the following endpoints: GET v1/books/ : List all books POST v1/books/: Create a new book GET v1/books/{book_id}: Get a book resource DELETE v1/books/{book_id}: Remove a book Step 1: Creating a Spring Boot Project Browse to your favorite IDE and create a Spring boot project with web, h2, data-jpa and Lombok dependencies. Following is the pom.xml file: pom.xml Step 2: Configuring h2 database In this application, we will use the h2 in-memory database as our backing database. Add the following configuration in the application.properties file to configure h2 database: application.properties Step 3: Creating the domain object In this application, we will manage book information. Users of this application/API can perform CRUD operations. We have created the following book entity: Book entity Step 4: Creating the REST endpoints To facilitate the REST endpoints, we have created the following REST controller. It serves the aforementioned endpoints: BookController.java We also have created the following repository: BookRepository Step 5: Insert data at application startup To help with the testing, let's insert a few book details in our database. To do so, create a new SQL file named data.sql in src/main/resources directory. Spring boot automatically executes this file at startup. data.sql Step 6: Testing the API Let us now start the application and test a few endpoints to ensure the application is working as expected. We are using HTTPPie to access the endpoint and we are receiving data from the server as shown below:
https://medium.com/swlh/how-to-secure-a-spring-boot-application-with-tls-176062895559
['Somnath Musib']
2019-12-03 05:01:02.046000+00:00
['Technology', 'Software Engineering', 'Coding', 'Software Development', 'Programming']
1,975
Autonomous Vehicle Technology Report
Fascinating read — 2020 Autonomous Vehicle Technology Report from Wevolver — this guide to understanding the state of the art in hardware & software for #selfdrivingvehicles #innovation#technology#imagination#imagine#ai#artificialintelligence#autonomousvehicles
https://medium.com/@chasethisnow/autonomous-vehicle-technology-report-a6e2c50237f1
['Michael Chase']
2020-12-14 20:01:34.275000+00:00
['Technology', 'Autonomous Vehicles', 'Report', 'Wevolver', 'AI']
1,976
Securing REST APIs with HTTP BASIC authentication
Securing REST APIs with HTTP BASIC authentication Security is no longer an afterthought, it's a must in any application these days. In the modern era, security is an essential feature in any application, your REST services are no exception to it. In this short tutorial, we’ll discuss how to secure your REST APIs with BASIC authentication. We’ll use Spring Security to implement the security of the API Application Endpoints We’ll use a fictitious application that lets us manage our favorite web URLs. Let us call this application Pocket. It has the following endpoints: GET /pockets: Returns all available pockets GET /pockets/{id}: Returns the available pocket for the supplied id. Return an error for invalid ids POST /pockets: Create a new pocket PUT /pockets: Update an existing pocket DELETE /pockets/{id}: Delete a pocket by the supplied id Application Development Let us now generate a Spring Boot project with the web, lombok, data-jpa, and H2 dependencies. Let us now the Pocket model object: Pocket Model POJO Let us now create the PocketRespository that lets us manage the pockets in the H2 database: Now that the repository is ready, let us create the PocketController class that provides the HTTP endpoints: We’ve used a custom exception that is thrown if there is an invalid pocket id is supplied: At this point, the API is ready to manage the pockets. However, the application is not at all secure and all pocket endpoints are accessible to everyone. Let us add the security aspect to the application. To begin with, let us first add the spring-boot-starter-security dependency in the pom.xml. This dependency brings-in the Spring Security infrastructure to the application. We now want to implement the HTTP Basic authentication. This Security configuration class lets you customize the security settings of your application. Let us explain the changes: You have extended the WebSecurityConfigurerAdapter class that lets you customize the Spring Security default configuration You have overridden the configure method and ensured that all request to the application needs to be authenticated and the authentication mode is HTTP BASIC You have provided a Bean definition for UserDetailsService. This interface is used to configure application users. We’ve used the in-memory one and hardcoded two user details Lastly, you have provided a Bean definition for PasswordEncoder. A PasswordEncoder is used to encode the plain-text password to a scrambled text to protect it from unauthorized access. For testing purposes, we’ve provided the NoOpsPasswordEncoder which does not provide any encoding at all. In a secure application, we should use better PasswordEncoders such as BCrypt We’ve created a few sample pockets in the main class: Testing the Application If you access the endpoints without BASIC authentication, you’ll be returned an HTTP 401 status code indicating you are Unauthorized: If you provide the BASIC authentication details, you can access the endpoints: Source Code The source code for this article is available in the GitHub repository at https://github.com/musibs/medium-articles/tree/master/spring-rest-basic-auth
https://medium.com/thecodefountain/securing-rest-apis-with-basic-authentication-fe9c2ffbbeb0
['Somnath Musib']
2021-01-03 15:24:45.642000+00:00
['Security', 'Spring Boot', 'Programming', 'Development', 'Technology']
1,977
Is quantum computing relevant to cyber security? — Arya College
Quantum computing is emerging on the horizon. This possible arrival presents a big step for technology and our society especially candidates of B Tech Colleges. One area where it will have an important impact is in the realms of cryptography and Cyber Security. The new techniques associated with quantum computing have the potential to turn the cryptographic world upside down. It considers severe implications for information security as well as the world at large. Significant effects on our online security What is quantum computing? In general, quantum computing holds the properties of quantum mechanics to perform computations. This contrasts with us every day or classical computers, which links to the properties of classical physics. Quantum computers rely on the units of information known as qubits. These can exist in states of zero and one, as well as super-positions of both zero and one. In comparison, classical computers just use zeros and ones in order to store information for the students of Engineering Colleges in Jaipur. What can quantum computers do? Quantum computer work under completely different principles than the computers that students of Best B Tech College in Jaipur use in their daily lives. Also, they have different capabilities. Many experts expect that they will be able to compute things and solve mathematical problems that classical computers simply are not able to do. Potential applications of quantum computing Modeling complex chemical reactions, which may lead to innovation and advances in chemistry. High-level financial modeling. Predicting weather and climate fluctuations with greater accuracy. Running more complex AI programs. Advanced computations in physics. Breaking currently secure cryptographic algorithms, as well as introducing new cryptosystems. Why are quantum computers a threat to cyber-security? The unique properties of quantum computers allow students of Computer Science Engineering at B Tech College in Jaipur to perform computations. They are currently impossible with classical computers. It creates an important impact on the cyber-security landscape. The most important parts of digital security rely on cryptographic calculations. They are easy to perform in one direction, but almost impossible to perform in reverse. Quantum computing represents a new wave of technology that comes with a host of different techniques. However, some of them have the potential to break various cryptosystems that make everyday communications safe for the students of Top Engineering Colleges in Rajasthan. What challenges lie in the way of quantum computing? Quantum computers are extremely temporary machines. It makes them immensely difficult to build and operate. They need to be isolated from the outside environment and kept almost at absolute zero (-273ºC) in order to be usable for the professionals of Top Engineering Colleges in Jaipur. If not, they produce quantum decoherence, which is essentially the loss of information to the environment. Thanks for Read our blog, you can check out full blog on official Page Arya College, Arya College is one of the Best Engineering College In Jaipur Rajasthan. In This College Many Branches for Engineering you can make great future with us. Arya College Provides Computer Engineering, Electrical Engineering & Electronics Engineering’s Branch for our Engineering students with top companies placements in campus.
https://medium.com/@aryacollegejaipur/is-quantum-computing-relevant-to-cyber-security-arya-college-f97f7f953659
['Arya College']
2020-12-01 05:44:38.841000+00:00
['Cyber Security', 'Engineering', 'Cryptography', 'Quantum Computing', 'Technology']
1,978
What has Evernote Become Today?
Are you a long time user of Evernote? How do you feel about its current state and the near future? From the early ’80s, well into the new millennium, we only drove Mercedes-Benz. No, we weren’t rich; we just discovered how to buy and sell, so they were the best value. We bought three-year-old cars from a trusted dealer who bought cars coming off lease. They were serviced to like-new condition. We kept them for a few years and repeated the process. They were great cars, and we thought we would drive them forever. Then, in 1998, Daimler bought Chrysler. Maybe Daimler improved Chryslers, but I can guarantee you, Chrysler ruined Mercedes, at least for a while. They weren’t the cars they once were. Plastic replaced steel. Things broke and wore out that shouldn’t have. We went back to Toyota. That’s how I feel about Evernote today. In 2000, Stepan Pachikov created Evernote. It was a simple concept well executed. I began using it casually around 2004 but didn’t get serious with it until 2010. At that point, I went all in, paid for a premium subscription, storing everything in it, admittedly, too much stuff. It was a fantastic product. Between 2015 and 2018, the company became fairly rocky, with many people coming and going starting with Phil Libin, one of the co-founders. The company took off in many different directions, most of which did not resonate with its user base. It was then that I started evaluating the competition. I had lost faith in the company and needed an exit strategy in case it failed. Most of the same competitors were around then, but I encountered the same problems as now. Nobody had the same feature set. There were things that Evernote did better than anyone else and a couple of things that no one else did. So, I still kept one foot in a few of these other products, just in case. Notion, Nimbus, and OneNote led that pack, although OneNote was a distant third. Microsoft is an even bigger elephant than Evernote. But I stayed (somewhat) loyal to Evernote. Then in 2018, Ian Small came on board. He began communicating openly with the user base. He told us he knew the company had screwed up and promised he was going to fix this. And those promises kept coming for two years with a lot of back-end stuff apparently going on that we, as users, didn’t really see the results of. Sometime in mid-2020, the new platform went into beta and open beta in August. There were a lot of complaints and suggestions made during this period. In October, they released version 10 to the public, ignoring most of the beta users. It was not well received, to put it mildly. Buried deep in the announcement below all the hype and fanfare was the warning that should have set off alarm bells. “But please be aware that there are a handful of features — such as AppleScript, import folders, and the ability to edit note creation dates — that are not yet available. If your workflow depends on these features, you can continue to use our legacy apps.” Opinions on the new version varied from those that thought it was great to those that thought it was pure garbage. The ones that thought it was great admittedly used a reduced feature set than most of us. Many sat in the middle of those extremes and felt that the product was still in beta testing or that it was “a work in progress.” But isn’t all software a work in progress? When it’s released to the public, it should be ready for public consumption. You’re always working on improving it and rolling out the next version, but the release is ready and working correctly. And based on the fact that installing the new version wiped out the old version, tells me that Ian Small and the rest of the Evernote team thought it was ready for public consumption. But they were wrong. Or lying. Or both. I am also a beta tester of Adobe Lightroom. As with Evernote, I signed an NDA, so I can’t discuss specifics, but I can tell you that their rollout is totally different, although there are many hiccups. Each .x release comes with the caveat that it is not ready for your production work. You are told to test it alongside your production version, and they make it very easy to do so. Once they have ironed out the kinks, they release the final version in that series, telling you that you can replace your old version with this one. They don’t force you to. They don’t delete the old one. It’s just a suggestion. Then the cycle starts over with another incremental version. That’s not what Evernote did. Despite the caveat they buried in the press release, they said, Hey, save your work, we will delete your existing version and replace it with this. That tells me that it is good enough to use on an everyday basis. It wasn’t. One of the things that Ian Small said was his biggest challenge was all of the different versions that existed for each platform. In other words, the Apple version was different from the PC version, and so on. But isn’t that true of most software? When I watch tutorials for Photoshop or Lightroom, they are frequently shown on a Mac. I have to translate that to my PC. Things are different. Things are in different places. One platform has things the other doesn’t due to differences in the OS. It’s been that way since I moved from an Apple IIe to an IBM PC. So, they had to level the playing field. They built it on a platform that would work across all operating systems. Sort of. Not really. To do that, they had to strip the program down to a set of least common denominators, least being the operative word. They took away most of the stuff that made Evernote unique. That made it the best. When I tested those other apps, I returned to Evernote because they didn’t have this thing or that feature. Now Evernote didn’t have them either. So, after the Internet blew up with complaints, Evernote finally came back with, yeah, but you can use the ‘legacy’ app (read: the one that works) alongside the new one as you test it, and we bring it back up to speed. But here’s the thing. Most of the users didn’t sign on to be beta testers. And those of us that did were told testing was over. Thanks for your suggestions. We are going to ignore them and release the product as is. Go away now; we don’t need you anymore. Because now, every user is a beta tester. I don’t mind being a beta tester at Adobe because, first, they aren’t removing features. They haven’t come to us and said, “we’re having trouble making this exposure adjustment work, so we’re going to take it out.” And second, there may be glitches, but every release is ‘good enough.’ I can still get my work done. It has many more moving parts than Evernote and works across at least as many platforms. And I spend a lot more time in Lightroom than Evernote. So, I know it can be done. But I need Evernote to do what it does. And right now, it doesn’t. And I really don’t get this mentality of running two versions side by side. Why? How? I’m in an airport, and I need to bring up a document, so I run v10. Whoops, I can’t find it. Hang on a minute and let me crank up the legacy version (one that works). Evernote is heralded as ‘a note-taking app,’ but I have always taken exception at that. Notepad is a note-taking app. Evernote is an archival document storage and retrieval database. Pretty simple, really. But both the storage and retrieval has been better than anyone else for years. Its web clipper is unparalleled. Its search ability renders the whole tag vs. notebook debate moot. The input folder, a feature unique to Evernote, allows me to add documents from my computer or scanner quickly. The ability to search text in images and PDF files is also unique. So, let’s cripple all that so the platforms are all the same. Why? Nobody outside of Evernote cares. I know from reading forums that the Mac version didn’t have the input folder capability. I never tossed and turned at night worrying about them. Hey, sucks to be you. I didn’t one day rise up and cast off the shackles of tyranny and oppression and shout, “If they can’t have it, I don’t want it either!” And neither did anyone else. No one uses a PC and Mac. Nobody has an Android and an iPhone. (I know I’m going to hear from the six of you that do but work with me here.) We have what we have, and we want to keep working with it the way we have been. And the fact of the matter is if they want a platform that works across all devices, that already exists. It’s been around for about thirty years. It’s called the World Wide Web. But here’s another thing Evernote should have taken note of. Many people didn’t use the web version unless it was the only one that worked in a given situation. Why? Because it lacked all the features that the native apps had. It was the least common denominator. And that’s what version 10 was. As a matter of fact, it looked and felt a great deal like the web version most had been avoiding. Here’s another problem I see with this new common platform. People are using it on all of the same devices they did before, Web, PC, Mac, iPhone, Android. But each of these platforms seems to have different problems. People have different complaints. If you look at the forums, you can see that what each user is complaining about seems dependent on the device on which they use it. So, how is that common platform working out? Not very well, it seems. But, you say, you are not forced to use the new version. You can stick with the legacy app. (The one that works). But how about this. Why didn’t Evernote stick with the legacy version? Until the new one worked. Until it was out of beta. Or, do like Adobe. Here’s the latest version. I hope you like it and please give us your feedback. Meanwhile, we will leave your old version in place because, you know, IT’S THE ONE THAT WORKS! Now, it’s a couple of months down the road. Evernote has released a lot of updates; I’ll give them that. Some have solved some problems and put some things back. They seem to be trying really hard. We are up to 10.4.4 something as of today. I don’t know how many individual releases there have been because I’ve only looked at a couple of them. I got tired of them wiping out my version and making me reinstall it, setting my preferences all over again. They have since fixed that problem, which is another admission that the new version isn’t ready for prime time. A few days ago, Ian Small released this in his blog. You should read it as it is the best one of these he has done IMO. But the TL;DR version is, we screwed up, but we are going to fix it. That has renewed my faith and hope in the future of the product. But we should remember one thing. It’s pretty much the same thing he said two years ago. I’m still testing and using the other guys. But I’m also still relying on Evernote for my day to day work. Of course, that means I am using the legacy version. (You know). I’m keeping a close eye on the forums and hoping they can right this ship. They will need to replace the features I rely on, such as an import folder, before coming back to v10. There are also many annoying glitches or features, I’m not sure which, that I hope they iron out. The larger challenge for them will be retaining their user base. And by that, I mean us premium users. The ones that are footing the bill for all of this. A lot have left if you believe the forum posts. At least we should be footing the bills. As we know, most of their working capital has come from VCs. There have been several vast infusions of money over the last ten years or so. But those guys aren’t going to keep ponying up if the user base dwindles. Too big to fail is bullshit. I want to say I’m excited over Evernote’s future, but that’s the product’s problem. It’s not exciting. It’s a utility. You put stuff in, and you get stuff out. Neither of those things is as good as they were six months ago. If they can fix that, I’ll stick around. I may even renew my premium sub. But they have to show me something first. The ball’s in your court, Ian.
https://dbvirago.medium.com/what-has-evernote-become-today-7aac9f49b1a
['Darryl Brooks']
2020-12-13 14:50:08.141000+00:00
['Technology', 'Software', 'Productivity', 'Evernote', 'Business']
1,979
What You Absolutely Need to Know About Elon Musk’s Starlink.
Have you ever been frustrated waiting so long for a broadband internet service at your remote location? If yes, then SpaceX’s Starlink has some good news for you. Starlink is Elon Musk’s ambitious project, which aims to deliver high-speed broadband connectivity no matter where you live on the planet. Starlink public beta service wowed its users and test results show that Starlink satellite internet can deliver blazing-fast download speeds. Without further ado, let’s dive deep into Starlink Satellite Internet. What is Starlink? Starlink is a satellite-based broadband service designed to deliver high-speed internet connectivity to anywhere else on the planet. Starlink comprises a network of satellites operated in Low Earth Orbit (LEO) at a distance of 550 km above the Earth’s surface. Starlink satellite features a compact design and weighs just 260 Kg (573 pounds). The satellite is on the leading edge of technology, which autonomously de-orbits the spacecraft once it reaches its life-cycle. SpaceX equipped Starlink satellite with four phased array antennas, which enable data transfer in a short duration. Thanks to its built-in navigation sensors, which provide accurate altitude information for precision data transfer. The first SpaceX’s Starlink mission launched with Falcon 9 rocket, which deployed 60 satellites into space. To date, with fifteen consecutive missions, SpaceX has successfully launched 900 satellites on Low Earth Orbit (LEO). Image Courtesy : SpaceX on Unsplash How Does Starlink Internet Works? The Starlink ground station transmits internet data signals to its satellites orbiting on LEO. One of its satellites receives the signals and communicates it with other satellites in its constellation using lasers. When the signals reach the satellite which covers the receiver area, then it transmits the signals to the ground user terminal. Since Starlink satellites operated in LEO, the up and downlink latency is comparatively low during the data transmission. Because the satellites are much closer to the ground receivers, hence, it can cover only a limited area. To provide extensive coverage, the number of satellites is vital in their network constellation. The Federal Communications Commission (FCC) of the United States has approved SpaceX to deploy 12,000 satellites on LEO. The company requested permission for furthermore 30,000 satellites to provide worldwide coverage. SpaceX plans to deploy 120 satellites per month. This could lead Starlink to cover extensive areas and so they can offer sophisticated service in rural areas and remote regions. A Starlink Satellite (dotted straight line) Spoted in the Night Sky. Credits : Andy Holmes on Unsplash Starlink — The Next-Gen Internet Starlink internet service is now undergoing public beta named “Better Than Nothing Beta” in parts of the Northern US and Canada. Starlink aims to expand its service rapidly to the rest of the world by 2021 says the company’s website. Starlink is ambitious about delivering high speed, low latency, broadband internet connectivity to remote locations where internet is expensive, unreliable or completely unavailable. The public beta users reported on various online portals that their internet speed ranges from 50 to 150 Mbps, with intermittent outages and latency differs between 20 to 40 ms (Millisecond). The Starlink user terminal kit priced at $499 and the internet subscription fee priced at $99 per month with no data caps. Final thoughts The Federal Communications Commission (FCC) announced the bid winners of the Rural Digital Opportunity Fund Phase 1 auction, recently. Elon Musk founded SpaceX company has won around nine hundred million dollars in federal subsidies. This fund will bring internet connectivity in rural and hard-to-reach areas in the United States through SpaceX’s Starlink project. Deploying the Starlink satellites on Low Earth Orbit is exorbitant for SpaceX. The company estimated that it cost around 10 billion US dollars to build and launch the first phase of 4,400 satellites. However, Business Insider report shows that the company can generate revenue of $50 billion a year from the Starlink project, once it serves the global market. Elon Musk deems that this revenue would fund his visionary dream project of colonizing the red planet, Mars.
https://medium.com/@pravinan/what-you-absolutely-need-to-know-about-elon-musks-starlink-afc190dab2b6
[]
2021-01-10 03:11:02.081000+00:00
['Starlink Internet', 'Internet', 'Spacex', 'Technology', 'Elon Musk']
1,980
Why we need to act now to save Europe’s tech ecosystem
Photo by NASA on Unsplash Europe’s tech ecosystem is broken. Progress on diversity has stalled, untapped talent is being squandered and discrimination is still rife. Undoubtedly, diversity and inclusion efforts have been steadily creeping up the sector’s list of priorities. 90% of people agree that diverse teams lead to higher performance in business¹. In fact, diverse teams outperform their homogenous counterparts by up to 48%². If we look at the world of venture capital, investment teams with diverse educational or ethnic backgrounds are up to 22% more likely to realise an exit on their investment portfolio³. Despite the clear upside of diverse teams, progress has halted. A whopping $91 out of every $100 invested in 2020 went to all-male founding teams, whilst 84.5% of founders were university educated and only 0.9% of investment went to Black founders⁴. So, what’s driving this asymmetry between intention and results? Maybe it’s because companies still see diversity as a tick-box exercise. Perhaps the tech sector is still too difficult to break into. Or because once a diverse candidate is hired, there are no measures to effectively support them in the workplace. As Suranga Chandratillake, General Partner at Balderton Capital⁵ says, “Step one is hiring someone who is different, step two is setting them up for success, step three is standing back to let them show you how they see the world.” It’s been shown that inclusive workplaces attract and retain the brightest talent, with an increase in individuals’ feelings of inclusion strengthening team performance by up to 17%⁶. Yet many efforts aimed at diversifying the ecosystem often prioritise siloed, short-term initiatives focused on giving diverse hires a seat at the table, over long-term efforts to ensure they are given a voice once they sit down. In short — diversity without inclusion is ineffective. To change the face of the startup ecosystem, we need to firstly build a bridge between young, untapped talent and the sector, and secondly, create inclusive communities at scale. The problem is that conversations about inclusion are often reduced to the efforts of a single company. But how can a candidate feel included when their firm has only just started prioritising diverse hiring and they are the only person in the office who looks like them? In an industry where progress on diversity and inclusion still has a long way to go, we need to consider inclusion on a broader scale. At Unbox, we’re solving this very problem. Unbox makes the process of discovering and growing untapped potential simple. How? We offer startups on-demand access to diverse top talent. What does that look like in practice? #1 Unbox focuses on discovering intersectional talent We are on a mission to remove social barriers to the tech ecosystem as we believe global tech will only be able to fulfil its true potential if diversity and inclusion are at its core. To achieve this goal, we are pioneering an innovative application process to minimise bias, partnering with diverse organisations and publicly pledging our commitment to diversity and equity. #2 Unbox offers talent, on-demand We want to empower every young person to achieve anything, irrespective of their background. We know that short-term placements at startups across industries, roles and technologies offer the best environment to learn and apply real-world professional skills. We also know that startups’ talent needs are constantly in flux. Our platform is optimised specifically for students and startups. #3 Unbox is not just a platform, it’s a movement We don’t just focus on getting young talent in the door. By building an inclusive life-long community and pipeline of diverse talent, we will empower young people to take charge of their own futures and transform the sector from the inside out. By empowering the world-changing companies of tomorrow with the very best talent, we are on a mission to change the face of the tech ecosystem. If you’re keen to be on the frontline of powering a movement that is building a more inclusive future, sign up for exclusive access to our beta platform today. Break out of the box.
https://medium.com/@unboxtalent/why-we-need-to-act-now-to-save-europes-tech-ecosystem-290daa2cf6da
[]
2020-12-24 15:49:49.008000+00:00
['Startup', 'Technology', 'Talent', 'Diversity', 'Inclusion']
1,981
Why are so many coders musicians?
I am currently at the beginning of my career as a developer. I came in ‘through the back door’ so to speak. I’ve been a musician my whole life, and it was through my interest in building music tools that I got into software. Going from music to code was a gradual transition. It took me some time to make the decision because of my fear of alienation. I didn’t want to turn into a cubicule zombie, typing mindlessly into a computer all day long, detaching from life and art. I have had a couple of corporate jobs before and couldn’t take it. For the creative type, alienated work sucks the marrow out of life. Musicians need creative expression in their work and have low tolerance towards soul crushing jobs. Even though coding implies staring into a screen for long hours, to my surprise, I found it was not alienating. In it I found a new way of expressing creativity. Just like working in music, either producing, composing, or playing, it didn’t feel like actual work. Why was that? In addition to this, I noticed many developers were musicians. After a few months in the industry, I realized this was not a coincidence. Finding musicians made me realize my fears were unjustified and it also made me wonder why many coders were musicians. What are the commonalities between the two professions that make this relationship? In this three part series of posts, I’ll talk about the different qualities that relate these two professions. Long-term commitment There seems to be a quality of focus for musicians and for coders. That reserve and focus is needed for people to be able to concentrate and develop skills for the long term. Staying on track and persevering through continuous frustration is a personal trait I find in both disciplines. Developing musicianship requires long-term commitment and a continuous training of brain plasticity to incorporate fine hand movements in instrumentality, to train the ear to distinguish between notes, chords and timbres, to learn how to read scores, and to transform the theoretical abstraction of harmony, counterpoint and instrumentation into mental representations of sound. As a beginner coder I’ve found myself in a similar process. Learning the fundamentals and becoming comfortable and creative with them requires a maturity of the concepts that takes a long time. A lot of the concepts in programming are abstractions that you can’t relate to day-to-day experiences, hence, they require a long time to settle in. This being said, though the nature of the two disciplines requires a similar mindset, it doesn’t mean the skills are the same. I don’t think logic and algorithmic thinking translate directly into music, which requires knowing how to count and having a good ear and coordination. At the same time, I don’t see how these last skills would translate into coding. A lot of musicians can’t code and a lot of coders couldn’t be musicians no matter how hard they tried. Nevertheless, the process by which you gain the skills is similar and rewards the kind of personality that is able to engage in long-term practice and learning. Thanks for reading. Please continue with the second part of the series, with other relationships I have found between music and coding.
https://medium.com/hackernoon/why-are-so-many-coders-musicians-60389fb8b645
['Francisco Rafart']
2018-03-08 13:19:10.481000+00:00
['Software Development', 'Technology', 'Programming', 'Learning', 'Music']
1,982
The Next Big Things in Machine Learning Is Coming Sooner Thanks to Covid-19
Technology The Next Big Things in Machine Learning Is Coming Sooner Thanks to Covid-19 Amazon and Apple already have a huge head start. Photo by Clark Street Mercantile on Unsplash Covid-19 has caused massive disruption to industries around the world. Airline and tourism were among the first ones to be hit hard due to entry restriction by most countries in an effort to curb the pandemic’s spread. Giant startups like Airbnb and Uber, among others, have also laid off some of their workers in an effort to shed expenses. Famous consulting firms, Accenture and Deloitte, did the same thing a couple of months ago. The list goes on and on. Among these industries, there is one, in particular, that may usher in the next trend in machine learning and AI— the retail industry. The following headlines hint at why: These global fashion companies have decided to close down physical stores and move their focus to online sales. The main reason, aside from shopper’s reluctance to venture into public places, is staggering rents. Take Singapore: when the “circuit-breaker” provisions went into effect — the national stay-at-home order—retail stores were closed. However, these stores still need to pay rent even though it could not operate. Meanwhile, online retailers’ stock prices, such as Amazon and Sea Group’s Shopee, are rising sharply amidst the spread of Covid-19. Mostly because people who are staying at home can still do online shopping even during lockdowns. The shift from companies having thousands of offline stores into focusing mainly on online sales is paving the way for retail AI. What is retail AI? Retail AI is essentially AI-related technologies that help the retail industry, from enhancing customer experiences, demand forecasting, inventory management, recommendation engines, chatbot, and so on. AI for retail has been around for a while now, such as H&M using AI to manage inventory and Sephora’s Color IQ for personalized make up recommendations. However, with big companies shifting to online sales, new use cases for AI in retail will surely emerge. New startups that specialize in providing retail AI solutions have also emerged, such as vue.ai and syte.ai. However, there are two major players that have been preparing for this shift. Amazon and Apple. Amazon’s Outfit-VITON Example of Outfit-VITON (source) Amazon recently published a paper on CVPR 2020. CVPR stands for Computer Vision and Pattern Recognition, it is by far the most prestigious conference un the field of image processing and computer vision. Outfit-VITON (Virtual Try-ON), as they called it, will let them generate an image of how a combination of various outfits may look on a particular model. In other words, it would let users shop online and try on outfits virtually. The paper shows a promising result, producing better images than the previous state-of-the-art models. However, we have yet to try it since Amazon has not released such a feature on its website at the time of writing this. This technology may not be that accurate to help people choose their sizes for now, but it will help people to see how they would look like using a combination of clothes without having to go into physical stores. In the near future, it may be possible for a machine learning model to recommend the size of outfit that will match you the best, or even recommending outfit style for you. Oh, wait, Amazon already did that last one. Apple’s Augmented Reality Screenshot by Author from Apple (source) Did you remember when Apple first released the Measure app and everyone with an iPhone posts Instagram stories and tweets about them? Shortly after, the more critical people started questioning the accuracy of the app when they tried to measure the same object more than once and got different results. But that is about to change with the addition of this component in Apple’s latest products. The LiDAR. LiDAR stands for Light Detection And Ranging. It works by emitting laser to the environment and measuring the time it needs to bounce back into the sensor to calculate the distance of each point. The sensor will give extra information about the depth to the device, which would greatly increase the distance perception needed for AR. When Apple first announced iPad Pro with LiDAR sensors, the first thing that came to my mind is “Apple is really getting serious with AR.” With the release of iPhone 12 Pro with LiDAR, they are further confirming their intention to focus on AR in the coming years. Screenshot by Author from GANs for Good video (source) In the recent GANs for Good video, Ian Goodfellow explains how he and his team at Apple use GAN to enhance the user’s experience on AR. He explains that placing a shiny object that is supposed to have reflections on them is quite hard due to the fact that the camera could not see objects and environments outside of its point of view. Their solution? They use GAN, Ian’s creation back in 2014, to generate how the environment might look like and use it to help generate a more realistic reflection for the virtual object. If a company is not serious about developing AR, having a special team with Ian Goodfellow as the director is not something they would do. You can watch the part where Ian explains about how Apple use GAN to enhance their AR experience here. Retail AI x Augmented Reality So what will happen if we combine the power of retail AI and Augmented Reality? Advertisement of IKEA Place by IKEA (source) IKEA had developed IKEA Place using Apple’s ARKit to let its customers virtually place furniture inside their home through AR. This application is arguably easier than the one for online fashion retail, due to the fact that they do not have to take into account the dimension and body shape of people who would use the furniture. IKEA Place is just one of the examples of what retail AI and AR can do together. With the superior AR performance of LiDAR equipped iPhones, it will be far easier to achieve better results for retail AI that use image processing. For example, having the exact measurements of a person will produce better product size recommendation rather than having to predict the measurement of said person from a flat image and dealing with the prediction errors. Other use cases? The first thing that comes to mind is using AR to superimpose a virtual interior design of a building to help the users better visualize living inside the new design instead of just seeing it through rendered pictures. Not just furniture though, imagine something that can superimpose 3D design to the current building. An even more complicated AR use case is for beauty products and hairstyles. This would let users try on make-up, color their hair, and even switch their hairstyles using the help of machine learning. L’Oréal collaborated with Modiface to make an app for styling your hair, but it was not… refined. I tried downloading the application to see how it works but it was very glitchy and hard to make it work. Since then, Modiface has been acquired by L’Oréal. Last year, L’Oréal and Modiface has also teamed up with Amazon to let users virtually try make-up products while shopping through Amazon. Final thoughts The retail AI industry is growing rapidly, with syte.ai raising 22 million and vue.ai raising 17 million, both on their Series B funding. With big fashion companies such as Topshop, Zara, and H&M moving to online sales, the industry might be growing even more rapidly now. Augmented Reality is still in the phase where it is considered as something high tech and fun to play with, but not something that most people want to use yet. Apple might change that. They had a history of moving the market. Apple may not necessarily be the first-movers, but they are the one who actually make the masses want to use new technologies. Take the iPhone and the Apple Watch for example. Touchscreen phones have been around for a while before Apple introduced the iPhone, but a lot of people still used Blackberry phones before that. Similar with Apple Watch, the smartwatch industry had existed long before Apple released theirs. However, not a lot of people wore one back then. If history repeats itself, Apple could make Augmented Reality way more popular and mainstream when they release their AR Headsets and AR Glasses, which is rumored to be in the next couple of years. These advancements in AR coupled with Amazon’s recent research to let buyers try on outfits virtually and a lot of popular clothing brands moving to online sales might be the spark that AR needs to make it trending. We might see the adoption of retail AI and AR by companies in near future. Trying on outfits and other products online will greatly enhance the online shopping experience and might become a catalyst for fully online brands with no physical store. One last thing, if, or when, the retail AI and Augmented Reality finally takes off, Amazon and Apple might become two of the biggest players with the head start they have right now.
https://medium.com/swlh/the-next-big-things-in-machine-learning-is-coming-sooner-thanks-to-covid-19-df259d35aac5
['Rionaldi Chandraseta']
2020-10-29 17:32:26.568000+00:00
['Artificial Intelligence', 'Technology', 'Machine Learning', 'Retail', 'Augmented Reality']
1,983
ANDROID TRICKS YOU MUST KNOW THIS YEAR
All the computers in the world are on a network. They’re linked by our cuffs. But I’m a computer. Jack’s a computer — Akilah — PA Young — all the cyclones. We’re all computers. Do you know the great thing about computers? They can be hacked.” ― Beth Revis, The Body Electric Andriod phones have undoubtedly revolutionized technology and its regarded as one of the most important inventions in the history of mankind. Since more or less everyone is using an android phone but you don’t know how much wonders can you do with that small phone in your pocket. Today I am gonna give you certain hacks and tricks you can do with your phone. Let's start without wasting any much time. LOST YOUR PHONE? WANNA MAKE THE THIEVES PAY FOR THIS HEINOUS ACT? Lost your phone? wondering how to prevent the thieves from further using that? Don’t worry we have got you saved. All you need is to block your IMEI number written at the back of your phone box from the concerned authorities so that they won’t be able to connect to the internet, surely not in this world. but what if you don’t have your phone box in which it arrived? Just Dial this code on your phone *#06# It will show you a 16 digit number. Just screenshot it and save it on your google drive to make the thieves taste their own poison in case it gets stolen FORGOT YOUR PASSWORD? Need a quick fix? If you think that taking your phone to a mobile repair shop if you have forgotten its password is the only thing then you might reconsider your choice after reading this one. If you ever forgot your password and have your data saved somewhere else and want to fix this, all you need is to follow these simple steps Firstly Power off your phone. The next step is to our your android phone in recovery mode/bootloader mode. The technique to put your phone in recovery mode is different for different android versions. Usually, you can do that by pressing the power and volume up/down button at the same instant. Then your phone will load up like this The next step is to do factory data reset and boom… there you go!!! BUYING A NEW PHONE? PERFORM THIS QUICK TEST FIRST Most of us purchase the phone just by reading the system specs while some don’t even fret to do that. Sometimes the phone comes with certain built-in problems like audio, camera, or touch system issues. To avoid that you must perform this test. For that, you need to put your system in testing mode. If you are a Samsung user you can test your phone by just dialing this code *#0# and it works for most of the phones. NEED SECURITY ON INTERNET? HERE IS HOW YOU CAN DO THAT Have you ever wondered if you try searching for some product on the internet you will find most of the ads on Facebook related to it? Google is actually spying on you and has an image of you according to what you search on it. Do you want some privacy? Alright here is a hack for it and it's our cute duckduckgo browser. The only browser in the world that is free of all those annoying ads like that of Google and no data breaches like google. I would personally recommend you all to use this instead of google because none of you will want to have google poking nose in your private searches RUNNING OUT OF SPACE? NO NEED TO BUY AN SD CARD If you are buying sd in 2020 then my friend you are a boomer! Gone are the days when people used to keep 4 to 5 sd cards to store their data. It the era of cloud computing. you can save everything from 200kb picture to your 10Gb apps or any software. In this estimation, I would still recommend not to use google drive because not only its short of space but also lacks privacy. Here is the app you should go go Its MEGA.NZ an app that comes with a cloud storage space of 50GB!! Yes, you heard it right and you know what’s more exciting? This space can be upgraded to several TBs. So get rid of SD cards and upgrade to cloud storage. TIRED OF PEOPLE CHECKING YOUR PHONE AGAIN AND AGAIN? WANNA HIDE YOUR DATA? If you think using an app lock on your gallery folder and your apps are going to save you then you may be wrong!! People can easily see your app lock application and can delete that in a matter of seconds. Need something more powerful than this? don’t worry we have got you covered again. Here is how you can do that. All you need is to install the app named calculator+ yes, you heard it right! it opens with a calculator interface until you dial a specific code to access your saved media. It can also act as an app lock if you want to lock an app and the most legit thing is it will take the photos of intruders that will try to open that app again and again. Don’t worry you can’t uninstall it like any normal app lock. Let's keep that a secret.
https://medium.com/@22sebty22/android-tricks-you-must-know-this-year-89104b159bc0
['Aamir Iqbal']
2020-11-27 15:45:05.847000+00:00
['Advancement', 'Phone', 'Technology', 'Hacks', 'Corona']
1,984
A decentralized platform to build E commerce websites on the Blockchain
A decentralized platform to build E commerce websites on the Blockchain Global decentralize network to buy and sell goods and services securely & convert crypto into fiat with no conversion fee Whitepaper Buy Tokens How It Work What is Adoafo? Adoafo is a market place where people will pay no conversion cost for buying and selling of goods and services, We’re reinventing the global equity blockchain — that will be secure, smart and easy-to-use platform, where people can convert any crypto to their fiat currency very fast and without paying any conversion cost for the fiat. Adoafo Ecosystem Adoafo Ecommerce Platform Adoafo will be the first Ecommerce platform to combine two vital aspects that are necessary for the progression of the industry, integrating Blockchain to store all customer data securely and using technological advancements such as Artificial Intelligence and Virtual Reality to help differentiate Adoafo from competition in terms of user experience. Identification Tag The RFID-enabled Asset identity tag (Asset ID Tag) is tamper-proof and theft-proof on both a material design as well as a data level. Adoafo Wallet The wallets are designed to be the gateway to and from the Adoafo Ecommerce platform in terms of transferring assets. Users will be able to track their portfolio and transfer various tokens to the platform to shop for goods and services in merchants’ stores on the Adoafo platform, and to send assets from the platform back to the wallet for either short or long-term storage. Adoafo Blockchain The transition to a cashless society is a long-term objective of Blockchain advocates, and this includes the termination of the use of debit/credit cards and bank transfers for Ecommerce transactions, eliminating the usage of fiat currencies altogether. Profiler Platform The multifaceted Profiler platform, powered by Adoafo, aims to revolutionize the Blockchain industry by providing a new level of transparency. This will be a gigantic step in upgrading the credibility of the industry in the eyes of skeptical outsiders, which is necessary in order to achieve the ultimate goal of mass adoption. Growth Pool The growth pool is designed to promote engagement and interaction between the Adoafo team and the community. RoadMap 24TH FEBRUARY, 2020 Concept Was Born 15TH December, 2020 Early Supporters Sale 15TH December , 2020 The Concept in Development Early Supporters Sale 20TH JANUARY, 2021 The Adoafo Private Events begins Private Token sales Events 10TH FEBRUARY, 2021 Team Expansion 2ND MARCH, 2021 Launch of Adoafo Wallet Launch of Profiler platform Official Start of the membership program Launch of the Growth pool 30TH APRIL, 2021 Launch of the official Adoafo Platform The Adoafo Token Sale The Sale Details Please refer to the Adoafo website for the details and latest updates on the Token Sale details. Token Distribution Schedule Treasury and Reserve Fund 20% • Early supporters 25% • Private Sale 25% • Public Sale 25% • Founding team 5% Use of Funds We currently anticipate that the net proceeds collected in the Public Token Sale will be used as follows:
https://medium.com/@siwizzy/a-decentralized-platform-to-build-e-commerce-websites-on-the-blockchain-474963f7e263
['David Onwuekwu']
2020-12-12 06:18:36.315000+00:00
['New Tokens', 'Ecommerce', 'Technology', 'Blockchain', 'Token Sale']
1,985
An Interesting Video Taking Apart the AirPods Max
An Interesting Video Taking Apart the AirPods Max The Headphones Are Surprisingly Repairable The latest AirPods Max seems to be more repairable than expected from Apple. Tech YouTuber Snazzy Labs broke down the headphones in an unlisted video. He goes over getting in and where some parts are located. Interestingly, it seems that both H1 chips are housed in the left ear cup (presumably because of space). The right ear cup is a little bit more cluttered with the battery. The video also breaks down the different sensors in the cups and how the signal is transferred through the headband to the other ear cup. It’s a very interesting video and worth watching.
https://medium.com/drknode/youtuber-breaks-down-airpods-max-in-secret-video-106bc3c85cbd
['Henry Gruett']
2020-12-26 21:11:19.274000+00:00
['Technews', 'Tech', 'Technology', 'Airpods Max', 'Apple']
1,986
10 Best leading Universities for Quantum Computing Research
Learn more about the top universities that specialize in Quantum Computing Research Quantum Computing has the potential to transform computation by solving certain types of obstinate issues. Only a few companies and start-ups have access to quantum computers that are composed of many tens of thousands of qubits. Some of these computers can be accessed through the cloud. Quantum computers use quantum to process information in a completely different way from traditional computers. Quantum computers transmute information using qubits (quantum bits), which can be either one or zero, unlike traditional computers that work on binary bits. In just a few short years, quantum information theory was combined with quantum computing they have evolved from a partial subject into a full class with well-funded programs, which takes place at quantum centers and institutes at top universities. Google, Microsoft, and IBM continue to invest heavily in quantum computing because they believe it will transform our world by solving problems that traditional computers cannot. 10 Best leading Universities for Quantum Computing Research 1. The Institute for Quantum Computing- University of Waterloo- Canadian University, University of Waterloo started offering quantum computing classes such as cat adoption agencies and adoption applications. Mike Lazaridis (creator of BlackBerry), funded the institute. The quantum computing powerhouse employed 296 researchers. It has published over 1500 research papers since its inception. This institute’s strength is its ability to combine academic excellence with entrepreneurial innovation to bring about technology. 2. The University of Oxford- This university says they are focusing their research on quantum computing because of its huge potential. It has the potential for transforming different areas such as finance, security, and healthcare. David Deutsch, the university’s founder, first described universal quantum back in 1985. Oxford and the University of York were the first to demonstrate NMR quantum computers. The university remains one of the most prestigious universities in quantum computing. Also read: Top 10 Best Artificial Intelligence Software 3. Harvard University — Harvard Quantum Initiative- The mission of Harvard University is to help architects and researchers to investigate and recognize better approaches to change quantum hypothesis into a valuable framework and gadget. The people group of analysts in the university has a distinct fascination for propelling designing and study of quantum frameworks and applications. 4. MIT — Centre for Theoretical Physics- MIT is an exploration leviathan establishment. It has a high expanded arrive at that centers profoundly around quantum computing and quantum data. The strength of the university’s hypothetical material science is presently applied to quantum data and quantum computing. MIT specialists additionally investigate regions like quantum calculations, Quantum data hypothesis, estimation and control, applications, and associations. Also read: The Top 10 Digital Process Automation (DPA) Tools 5. The National University of Singapore and Nanyang Technological University — Centre for Quantum Technologies- The university community was established to connect together physicists, specialists, and PC researchers to explore on fundamentals of quantum physical science and to develop gadgets dependent on quantum marvels. The scientists and specialists of quantum innovation are inferring their revelations in detecting, correspondences, and computing. 6. The University of California Berkeley- The Berkeley Center for Quantum Computation and Information includes analysts from the schools of Engineering, Chemistry, and Physical Sciences. These analysts work on head and central issues in quantum gadgets, quantum cryptography, quantum data hypothesis, quantum calculations, quantum control, and the trial acknowledgment of quantum PCs. 7. University of Maryland — Joint Quantum Institute- The Joint Quantum Institute includes quantum researchers from the National Institute of Standards and Technology (NIST), the Department of Physics of the University of Maryland (UMD), and the Laboratory for Physical Sciences (LPS). Every one of these organizations brings major hypothetical and exploratory examination programs with a devotion to control and send the quantum framework. Also read: Top 6 Tips to Stay Focused on Your Financial Goals 8. University of Science and Technology of China (USTC) — Division of Quantum Physics and Quantum Information- The Division of Quantum Physics and Quantum Information centers around the field of quantum data and quantum optics. The division for the most part coordinates towards fiber-based quantum correspondence, quantum memory, and quantum repeater, free-space quantum correspondence, quantum establishment, superconducting quantum computing, optical quantum computing, quantum reenactment with ultracold molecules, quantum metrology, and related hypotheses. Progressed test stages are worked by analysts to lead state-of-the-art research. 9. The University of Chicago — Chicago Quantum Exchange (CQE)- The Chicago Quantum Exchange is a focal point of analysts showing distinct fascination for modern endeavors and propelling scholastics in the designing and study of quantum data and computing. The goal of the middle is to advance the ID and investigation of quantum data and computing advances and furthermore the improvement of new applications. The prevalent objective is to apply research advancements to foster drastically new sorts of gadgets, materials, and computing strategies. Also read: How to Calculate Your Body Temperature with an iPhone Using Smart Thermometer 10. University of Sydney — Australia- The University of Sydney quantum science bunch accentuates coordinating to the most difficult issues of quantum computing and applying these bits of knowledge to develop new advancements. The scientific searches of the university are locked in with profound modern and enterprising exercises. The exercises associated with the university range from quantum data and quantum computing, major physical science to innovation advancement. Originally published at https://www.the-next-tech.com on July 7, 2021.
https://medium.com/@thenexttech/10-best-leading-universities-for-quantum-computing-research-f9a7a1139d5b
['The Next Tech']
2021-07-08 04:20:19.757000+00:00
['Tech', 'Quantum Computing', 'Research', 'Technology News', 'Universities']
1,987
Oppo created panic in the world of mobile phones, launched this powerful phone, know what its specifications and battery backup.
Oppo created panic in the world of mobile phones, launched this powerful phone, know what its specifications and battery backup. Technoven Apr 23, 2020·2 min read HomeRedmi Oppo created panic in the world of mobile phones, launched this powerful phone, know what its specifications and battery backup. Xiaomi launched Redmi 7 smartphone for Rs 7500, apart from strong battery and processor, these are the features Diamond-cut design has been given on the back panel of Oppo A12 smartphone, which has been seen in many Realme smartphones before. Oppo A12 has a dual rear camera and entry-level specifications such as MediaTek Helio P35 processor and a 13-megapixel primary camera. Honor launches two powerful budget smartphones, know price and features Talking about the specifications of Oppo A12, this phone with nano dual sim works on Android Pie based ColorOS 6.1.2. This phone comes with 6.22 inch HD + (720 × 1520 pixels) display, 89 percent screen-to-body ratio, 1500: 1 contrast ratio, 450 nits brightness and Gorilla Glass 3 protection. In addition, the phone has 4 GB RAM with the Octa-core MediaTek Helio P35 processor. Nokia launched this budget smartphone in India; Learn price, features and offers here The Oppo A12 handset is dual rear camera cellus, whose primary camera is 13 megapixels and its aperture is F / 2.2. It has a secondary camera of 2 megapixels. It acts as a depth sensor. Samsung launched triple rear camera smartphone in the budget range For selfie, you will get a 5 megapixel camera, which has an aperture F / 2.4. This camera is located in a waterdrop-style notch. Not only this, the camera app of this phone comes with digital color mode, about which the company claims that it works to improve the picture with natural color. Nokia launched this cheap smartphone, learn all the features Talking about storage, in addition to 64 GB storage, there is support for microSD card up to 256 GB. The battery of the phone is 4320 mAh. For connectivity, OPPO A12 has 4G, Bluetooth 5.0, GPS. Samsung Galaxy A20s: Samsung’s budget smartphone launched with Dhansu features Talking about the price of Oppo A12, the price of its 4 GB RAM and 64 GB storage variants in Indonesia is IDR 2,499,000 (about 12,300 rupees). The phone will be available in two color options — Black and Blue. How long will the Oppo A12 launch in India, information is yet to be known.
https://medium.com/@technoven/oppo-created-panic-in-the-world-of-mobile-phones-launched-this-powerful-phone-know-what-its-ad34809ebb3
[]
2020-04-23 06:00:49.003000+00:00
['Technews', 'Technology News', 'Technology', 'Mobile Marketing', 'Tech']
1,988
Useful JavaScript Tips — Functions and Casting
Photo by Gyorgy Bakos on Unsplash Like any kind of apps, JavaScript apps also have to be written well. Otherwise, we run into all kinds of issues later on. In this article, we’ll look at some tips we should follow to write JavaScript code faster and better. Function Parameters A JavaScript function can accept zero or more parameters. For instance, we can write: const foo = () => { //... } or: const bar = (a) => { //... } or: const baz = (a, b) => { //... } Starting with ES6, we can set default values for parameters, For instance, we can write: const baz = (a = 1, b = 2) => { //... } Then if we don’t pass in anything when we call baz then a is 1 and b is 2. Since ES2018, trailing commas are allowed for function parameters. This helps reduce bugs because of missing commas around parameters. For instance, we can write: const baz = (a = 1, b = 2,) => { //... } We can also call a function with trailing commas: baz(3, 4,) Also, we can use the spread operator to spread array entries into arguments. For instance, we can write: const args = [2, 'foo']; baz(...args); We can also destructure object parameters into variables so that we don’t have to remember the order that arguments have to be passed in. We can write: const baz = ({ a = 1, b = 2 }) => { //... } We destructure the properties into variables. If the property isn’t set, then we set the default value given. Switch Statements An alternative to the if-else statements is to use the switch statement. For instance, we can write: switch(a) { case 1: //do something break; case 2: //do something break; case 3: //do something break; } We check the value of a and then do something given those values. We need a break statement at the end of each case so that the code below a matching case won’t run. Alternatively, we can use return : switch(a) { case 1: return 'foo'; case 2: return 'bar'; case 3: return 'baz'; } Also, we can provide a default case which runs if none of the values of a matches. For instance, we can write: switch(a) { case 1: //do something break; case 2: //do something break; case 3: //do something break; default: break; } if and else The if statement makes a program run something if a condition is true . For instance, the following always runs: if (true) { // do something } And this never runs: if (false) { // never runs } We can add an else block below an if block if we need to do something if a condition is false . For instance, we can write: if (foo) { // do something } else if (bar) { // do something else } else { // fallback case } delete Operator The delete operator lets us remove a property from an object. For instance, we can write: const dog = { breed: 'chow chow', color: 'brown' } Then we can remove the color property by writing: delete dog.color; We can also write: delete dog['color']; Converting to Strings We can call the toString method on primitive values and objects to get the string representation of it. Converting Number to String To convert a number to a string, we can use the String function or the toString method. For instance, we can write: String(100); which returns '100' , or: (100).toString(); which returns the same thing. Convert Boolean to String Likewise, we can use the String function or toString method to convert booleans to a string. If we write: String(true) or: true.toString() we get true . We can do the same with false . Convert Date to String The String function and toString method can also be used to convert Date instances to strings. For instance, we can write: String(new Date(2020, 0, 1)) Then we get: "Wed Jan 01 2020 00:00:00 GMT-0800 (Pacific Standard Time)" and we can do the same with toString : (new Date(2020, 0, 1)).toString() Converting Special Values to String The String function can also convert null , undefined , or NaN to a string. If we write: String(null) and we get 'null' . If we write String(undefined) and we get 'undefined' . String(NaN) returns ‘NaN' . Photo by AZGAN MjESHTRI on Unsplash Conclusion JavaScript functions can take zero or more function parameters. We can set default values for them. The delete operator can be used to remove a property. There are also many ways to convert data types with JavaScript.
https://medium.com/dataseries/useful-javascript-tips-functions-and-casting-8c05bcf9dc19
['John Au-Yeung']
2020-07-02 15:26:41.828000+00:00
['Technology', 'Programming', 'Software Development', 'Web Development', 'JavaScript']
1,989
SYNCHRONIZE: KEEPING IN TOUCH WITH FRIENDS MADE EASIER FOR YOU
Have you been feeling stifled sitting in your house all day? Found yourself dispirited by the monotony of quarantine? Have you been missing, being able to talk about everything and nothing with your friends? Recently felt like your friends have no time for you? If your answer to these questions is ‘yes’ then continue reading to find out about a new simple solution to all your scheduling conflicts. INTRODUCTION The virtual world is more important than ever in the pandemic and has quickly become vital to our new lifestyle. It makes things like education and business thrive even with people stuck in their homes. However, there is one aspect of life we find ourselves majorly missing out on, and, that is our friends. Between virtual job conferences, online classes, exams and assignments, we find ourselves having very little time left for socializing with friends. When we do have time, it appears they have none and any common free time is wasted in trying to schedule a virtual meeting. This is what inspired the idea of “Synchronize”, an easy-to-use, user-friendly plugin to help solve all your scheduling disputes. “Synchronize” will sync all your calendars into one making it easy to use. It will also allow users to form groups with family and friends and after comparing your calendars, it will notify all parties if they have common free time thereby, saving a lot of valuable time. You can even set up parameters such as setting some days as “personal” when you do not wish to be disturbed. “Synchronize” will ensure that no meetings are scheduled on such days. You can also set meeting time lengths with the default setting at 30 minutes. Another beneficial feature is the ability to track your meetings with various groups. MARKET DEMAND “Synchronize” benefits everybody, as it can be used for setting up meetings with anybody ranging from family and friends to business associates. It will also benefit all the people who wish to manage time more effectively and keep a track of their virtual meetings. RELEVANT PRODUCTS IN THE MARKET Although there are already a few calendars and meeting scheduler apps and plugins such as Calendly, Meetingbird, Doodle etc. none of them offer the integrated simple solution that “Synchronize” will provide. They also have different working principles and concepts. REQUIRED KNOWLEDGE OR EXPERTISE Knowledge on how to create plugins, how to sync calendars, marketing knowledge etc. are must haves. Necessary languages include but are not limited to: HTML, PHP, CSS, JavaScript. It is also imperative to conduct polls to receive valuable customer feedback and try and implement useful suggestions obtained. MY SOLUTION You can form various groups as per your choice and convenience for different friend circles, family members who live far, business associates, clients etc. and simply allow “Synchronize” to access your calendars and sync them into one. Then simply set meeting lengths, define personal days etc. and the product will offer you meeting times available based on your predefined criteria easily eradicating all scheduling from your part. It is a great way to save time and energy and get rid of the hassles of conflicting schedules. PRODUCT PRICING The product can be free with limited features such as limiting the number of groups that can be formed etc. while the full version will be made available to paying customers. Paying customers will enjoy benefits such as being able to make as many groups as required at a time, obtain any subsequent new features before the non- paying guests among others. For paying customers, the pricing should be reasonable with respect to the services provided as many of our customers may be students who can’t afford high prices. This is also intended to attract more paying customers and increase revenue. Furthermore, family packs etc. can be offered to interest people in buying the product for their family members. MARKETING PLAN The product’s marketing will be carried out in digital media, print media as well as by word of mouth. For digital marketing, digital advertisements such as advertisements on television, digital advertising boards, advertisements on various websites will be done. Advertisements in print media will include newspaper ads, ads in magazines, banners, hoardings, billboards, posters etc. Advertisement by method of word of mouth depends largely on the reputation of the product. Satisfied customers are highly likely to invite their family and friends to try out “Synchronize” and this will create a large customer pool for the product. Marketing is one of the major aspects of any company or product and it is therefore, imperative to have great marketing strategies. As Joe Chernov has said “Good marketing makes the company look smart. Great marketing makes the customer feel smart.” RISK ANALYSIS Improper marketing poses the biggest threat to the success of “Synchronize” as people cannot use that which they are unaware of. That is why it is very important to have proper marketing. Another source for errors is improper updating of calendars by the customers. Not everyone is in the habit of using calendars and keeping a proper track of time. It might be difficult for such people to be able to properly schedule meeting in the beginning. However, once you start to regularly use “Synchronize”, it will be easy to form the habit of using calendars and properly updating and using them. MILESTONES AND TIMELINE The first step would be to do intensive planning and research for the product and creating it. The next step would be testing and improvements followed by marketing and launching. The initial stages of expansion are estimated to be slow but once the product gains a firm foothold in the industry, it is expected to achieve exponential growth. The end goal is to go global with “Synchronize”. The planning and research stage would take approximately 5–8 weeks. The development of the product is estimated to take about 6–10 weeks followed by another few weeks of testing and improvements. The marketing is a simultaneous process which should go on from at least a few months leading up to the launch of the product and continue even beyond. Deliverables include: Ø Customer feedback and polls Ø Product code Ø Tests and test results Ø Documentation required Ø Presentations FUTURE PROSPECTS There are innumerable future prospects for this project as we can add new features to make the processes even simpler and more effective over time. Potential future features include keeping detailed track of meeting and analysing behaviour to recognize patterns. “Synchronize” may even be able to offer its own A.I assistant for customers to help manage time and virtual gatherings. CONCLUSION Though the post pandemic world is rather dissimilar from our lives before, the needs of people haven’t changed. In fact, in these difficult times, it now more important than ever for us to feel “normal” and that requires us to enjoy all parts of our lives in a balanced, healthy manner. However, our hectic lives and conflicting schedules make it a herculean task to even enjoy spending some time with our friends in a virtual setting. It appears that we are always busy with one assignment, one task, one test or another and when we do get time for our friends, it seems they don’t have any for us and whenever we find some common free time, it is wasted because we are unable to effectively schedule a meeting. This is what inspired the idea of “Synchronize”, a user-friendly, easy-to-use plugin to help you manage and eradicate all you scheduling conflicts. Staying in touch has never been easier!
https://medium.com/@anushkabanerjee/synchronize-keeping-in-touch-with-friends-made-easier-for-you-7c10cb0b9f54
['Anushka B.']
2020-12-20 11:35:44.395000+00:00
['Scheduling', 'Friends', 'Virtual Meetings', 'Business', 'Technology']
1,990
JSONField Models in Graphene Django
JSONField Models in Graphene Django Graphene + GraphQL In my humble opinion, Graphene is very stable and is probably the best library for creating GraphQL endpoints in Python. While working with Graphene, I came across an issue where JSONFields were always returned with quotation marks and backslashes — JSONString . Example — meta field is returned as a JSON string rather than a JSON object This article aims to address this issue using GenericScalar , which is currently an undocumented type in the current version of Graphene documentation. TL;DR This can be easily solved by overriding your meta field type in your DjangoObjectType to GenericScalar such as below. Example Detailed Example with Code In this section, I am going to run through a few snippets of codes for creating a simple mutation and query that we can work with. Example folder structure of a Django app Database Model Simple database model for demonstration. GraphQL Types Creating a DjangoObjectType for our Person object. Django Filters Let’s add some basic filters to our code while we are at it. GraphQL Mutations For simplicity’s sake, we will only need a single createPerson mutation for this example. We are going to pass in metadata as a dictionary/ JSON data input for our mutation. GraphQL Schema Finally, we will stitch everything together inside schema.py and we are ready to try things out! Results Trying Out The Mutation Let’s create a Person object using our newly created CreatePerson mutation at your /graphql endpoint using the built-in GraphiQL IDE or any API client of your choice (Postman, Insomnia REST Client, etc.) # Create a Person object mutation createPerson($input: CreatePersonInput!) { createPerson(input: $input) { person { id } } } { "input": { "firstName": "Albert", "lastName": "Joe", "metadata": { "is_admin": true, "email": ' } } } # Query Variable"input": {"firstName": "Albert","lastName": "Joe","metadata": {"is_admin": true,"email": ' albert@example.com Solution As you can see in the meta field below, it’s returned as a JSONString which is not what we wanted. Things could get uglier especially when you’re dealing with a large JSON object. The issue was apparently caused by the fact that JSONFields are by default treated as JSONString in Graphene. However, this can simply be fixed by applying GenericScalar to types.py such as below.
https://medium.com/open-graphql/jsonfield-models-in-graphene-django-308ae43d14ee
['Jerry Ng']
2021-01-23 08:55:55.117000+00:00
['Python', 'GraphQL', 'Graphene', 'Technology', 'Django']
1,991
75M People Voted for Donald Trump
What if I was to tell you we could eradicate racist biased behaviors in four generations? I know it’s crazy, but exciting concept huh? Half the country is considerate of their fellow man. Some Trump supporters voted for him because of his policies, economics, or because they feel whites should rule supremely. Chappelle was right! You use to love masks! Couldn’t get enough of them. What happened there? You and masks fell out? As I watched the vote count go up, Biden and Donald gaining neck and neck, I realized what she meant when Angela Davis said it wasn’t enough to not be racist, you have to be anti-racist. To be actively anti-racist means to breed out elitist traits in your children and your community. You have to defend someone being attacked by a Karen. You have to be willing to hear hard things, process them, and hold yourself accountable. Being anti-racist means bringing external consultants into your workplace. Don’t ask Blacks & minorities in your office to take on the additional work of diversity, inclusion, equality, and anti-racism on top of their daily tasks. Call on the professionals of The 11 Step Program brought to you by The Bias Rehab Center.
https://medium.com/justice-for-all-of-us/75m-people-voted-for-donald-trump-655f3c5a878b
['Brooke Sinclair']
2020-12-02 15:27:02.944000+00:00
['Startup', 'Entrepreneurship', 'Technology', 'Venture Capital', 'Silicon Valley']
1,992
Demystifying Tech Team: How Our Dev Department Works
Have you ever wondered about how your banking app works? For most consumers, everything tech and finance-related is a deep mystery. For service providers and software engineers in particular, it is a hard work. Dev team is the heart of every online business, but its contribution is often left behind the curtains. If you’re curious about how ‘the magic’ happens, we’re ready to demystify. Team Structure Every department could benefit from a well-planned structure, but tech teams cannot function adequately without it. Imagine dozens, if not hundreds, of developers working on multiple projects, receiving tasks from multiple departments, and urgently fixing bugs at 3 a.m. Once your business passes the startup stage, this model is no longer efficient. For this very reason, Mercuryo’s dev department is divided into units. Each unit is dedicated to developing a particular tool from our signature toolbox, be it a user-focused crypto wallet or something more complex like API as a service. It is still quite troublesome to ask a team to check on somebody else’s code related to a completely different product despite documentation maintenance. That’s why specialization matters. Mercuryo’s tech units consist of several teammates, including product manager, tech lead, testers, system analytic, DevOps engineer, and various back-end, front-end, and mobile devs. We’re trying to keep an optimal number of team members in each unit and not overload them. By the end of 2021, we’re hoping to staff ten complete units. A product manager monitors the unit’s work. This person has to be tech-savvy and always has access to the code, so they can check and comment on it anytime. Product managers are always well-aware about what their team is working at, and ready to assist or correct their work. Technical Lead is the team member with the most profound product knowledge. A tester, back-end developer, or anyone else from the unit can act as a tech lead. They guide and assist other developers and help them whenever an issue occurs. They are responsible for the uniqueness and performance of the stand. Each unit has its own stand to test out the tasks it tackles. Basically, a stand is a testing environment for the dev team dedicated to a certain unit. Every product’s feature should be tested on the stand before the team can release it. Most of the devs work remotely, although they can join the team in one of our offices if they choose to. To coordinate the workflow more efficiently, the teams have weekly meetings to discuss the plan for the week ahead. Most tasks come from support, business development, financial departments, and sometimes, from our B2B partners. If any issues arise, developers have to be ready to fix them asap. “People tend to get extra sensitive when it comes to money, so a quick reaction is a must. In a way, you can compare developers and doctors as they have to be ready to step up immediately in case of an emergency. Developers aren’t saving lives, but they deal with technological and, in our case, financial aspects of people’s daily routine, which sometimes can turn into a life and death situation.” Konstantin Stasenko, CTO at Mercuryo Infrastructure Specifics Initially, the project was employing the PHP monolith. It is pretty common for a startup to use PHP as an ideal temporary solution to launch faster. However, once the company starts scaling, it may turn into a setback as soon as the company starts growing, new people join the team, and the amount of code increases significantly. Later on the way, Mercuryo had to face the issue. Once we realized our team could not handle the number of tasks it gets, it was clear we had to make some serious changes and lean toward microservices. Microservices came out as a great alternative that fitted the needs of a rapidly growing fintech company simplifying the development process. Microservice architecture proved to be the best fit for Mercuryo as it is a lot easier to design a few microservices than support one large monolith structure. GoLang proved to be a much better fit for a company like ours, and the biggest struggle was to rewrite existing products in a new language. Until now, some bits and pieces of our architecture remain as a PHP monolith, but eventually, we hope to finalize the transition. Searching for Talents Speaking of challenges, highly-skilled developers are worth their weight in gold. A lack of talented software engineers on the market is one issue. Secondly, a common trend for a developer is to stick to one company and rarely change the employer unless something extraordinary happens. Finally, some developers cannot pass the interview process as their experience is not enough to keep up with the rest of the team. A typical interviewing process consists of two parts. First, the candidates have to undergo an hour-long tech evaluation, demonstrating their skills in the live mode. This interview with Technical Leads covers all the critical aspects of their future work to get complete information on the candidate, their hard skills, and expertise. Those who successfully passed the first stage will be offered the second interview meant to define their soft skills and the ability to work in a team. You would think that finding a rare specialist with an ideal set of hard skills and impeccable expertise is already good. However, we came across excellent professionals with serious commitment issues who failed every deadline and struggled to show up. Fortunately, although these cases are an exception, it is always better to filter out such people before joining the team. The Bottom Line The work of developers can be described as magic. Those are developers who bring all the wild ideas and solutions into life, creating simplicity by applying maximum effort, invisible for the end-user. Forming a strong and solid tech team is not easy, but we’re getting there at our own pace. Want to join our team or know an excellent developer looking for new opportunities? We’re still hiring!
https://medium.com/mercuryo/demystifying-tech-team-how-our-dev-department-works-46dc3777a8e
['Mercuryo Hare']
2021-09-07 10:59:58.016000+00:00
['Fintech', 'Product Development', 'Technology', 'Tech Development', 'It']
1,993
Kim K, virtual humans, and how weird becomes mainstream.
🧞 Now the virtual companions are coming So where next for virtual humans? Strap in, because this is where things get wild. A new breed of startups wants to make it easier to create these digital people. Take a look at Synthesia, or Alethea AI, which I featured in the last New Week Same Humans. Businesses will remain the key clients for a while to come. But technology wants to be democratised, and soon enough startups will empower us all to build virtual humans of our own. Why would anyone want to do that? To answer that question, we need to understand another deep shift reshaping the world around us in 2020. This one is a shift in our relationship with digital technology. Across the last few years, we’ve all become accustomed to the proliferation of a new kind of virtual entity: the AI assistant. Right now, our relationship with those entities is primarily functional. Alexa, order me some washing powder. But in 2020, that relationship is changing. It’s moving beyond mere functionality, and starting to encompass power higher-order human needs: status, creativity, and even companionship. People are becoming friends with AIs. For over three years I’ve been tracking one of the primary examples of this: the AI-fuelled chatbot app Replika, which bills itself as ‘your AI friend’. Chat to your Replika and over time it gets to know you. It asks about that job you applied for, or the argument you had with your girlfriend. It wants to keep you company, and make you reflect on your own behaviour. Total downloads are more than 3 million. The New York Times recently reported that numbers had spiked during the pandemic: over half a million people downloaded Replika in April alone. A virtual human inside the Replika app Once again we meet our old friend, the weird. It’s easy to laugh at the idea of a person being ‘friends’ with an AI. But if you’re inclined to do so, ask yourself: have you ever said anything to Siri or Alexa with any emotional content? Ever asked an AI assistant about your mood, or a feeling, or a friend? Apple have long known that millions do speak to Siri like that. Way back in 2017 it revealed that it was hiring psychologists into the Siri team, so that she can ‘have more serious conversations’ with users. The human needs for social connection, self-understanding, and companionship are deep, and eternal. When a technology unlocks new ways to serve these kinds of human needs, new behaviours, attitudes and lifestyles are often the result. Those behaviours may seem strange at first. But eventually, the power of the underlying human needs at work tends to overwhelm our collective resistance. That’s what happened 20 years ago, when the internet unlocked new ways to serve the human needs for social and romantic connection. And it’s happening again now, with AI. What does this have to do with virtual humans? Siri, Alexa and other AI assistants are only disembodied voices. But we humans are better able to form a connection with beings we can see and hear. So in the years ahead, these AI-fuelled entities will merge with virtual humans, to become a new kind of virtual companion. It’s already happening. The first iteration of Replika was a text-only chatbot. But now users who pay can create an avatar with a realistic face and voice, and talk to it as they would a friend on the phone. Meanwhile, tech giant Samsung are working on a new line of ‘humanoid friends’ called NEONs. ‘NEONs will be our friends, collaborators, and companions, continually learning, evolving, and forming memories from their interactions’ says Pranav Mistry, CEO of the Samsung innovation lab behind the project.
https://medium.com/new-world-same-humans/kim-k-virtual-humans-and-how-weird-becomes-mainstream-4d8a78e48299
['David Mattin']
2020-11-04 13:51:31.991000+00:00
['Future', 'Technology', 'Business', 'Trends']
1,994
10 Extraordinary GitHub Repos for All Developers
10 Extraordinary GitHub Repos for All Developers Interview resources, build your own X, a list of great public APIs, and more Photo by Vishnu R Nair on Unsplash GitHub is the number one platform for sharing all kinds of technologies, frameworks, libraries, and collections of all sorts. But with the sheer mass also comes the problem to find the most useful repositories. So I have decided to curate this list of ten fantastic repositories that provide great value for all software engineers. All of them have a lot of GitHub stars, underlining their relevance, popularity, and usefulness. Some of them will help you learn new things, some will help you build cool things, and all of them will help you to become better software engineers.
https://medium.com/better-programming/10-extraordinary-github-repos-for-all-developers-939cdeb28ad0
['Simon Holdorf']
2020-03-31 16:07:48.275000+00:00
['Creativity', 'JavaScript', 'Technology', 'Productivity', 'Programming']
1,995
Zero Inbox Wastes Time
Zero Inbox Wastes Time How cluttered email can actually be more efficient At this moment, I have 6809 messages in the inbox for my 9to5 job, and 4,202 of them are unread. The very idea makes my friend Heather shudder in horror. But I don’t miss important messages, I can generally find what I need, and I don’t waste time trying to clean up email. If having 6,000 emails in your inbox would give you palpitations, you probably have a working system in place. If, however, you lose things in your cluttered email or feel guilty for not living up to the Zero Inbox standard, I have a system that will help you feel more in control and allow you to let go of impossible perfectionism. Why Zero Inbox is Difficult (or Impossible to Achieve) Sheer volume. I sampled my office email from the last semester and found that I have about 200 email threads updated daily. That doesn’t include the messages coming into Messy Desk Consulting or to the Gmail account I use for shopping and “free” downloads. I must not be the only one. According to The Radicati Group, over 306 billion emails will be sent & received per day in 2020. If I set a goal to read, process, delete, store, forward, or reply to every one of those messages, I could easily spend most days working with nothing but email, and I would be completing less than a third of my actual job. AI isn’t doing it. Marketing floods our inboxes daily. In my personal accounts, I take advantage of Gmail’s AI features that sort Social, Updates, Forums, Promotions, and Spam, but those results aren’t consistent enough to do what I need at work. It simply isn’t acceptable for a message from the President of my college to show up in “forums” alongside true forums I use for professional development. For me, there is no point in having the filter if I need to check it daily to keep up with important mail. You are supposed to be able to train the AI to sort as you prefer. Now, I may just be a bad trainer (my dog’s behavior would support that hypothesis), but I cannot seem to make Google remember what I tell it. Just when I get all the marketing separated from invoices, some update or slip of the mouse messes up the system. Tips for Managing the Cluttered Inbox Casey Botticello gives some great tips in his post about spending less time on emails. There are three ideas from his piece that are particularly applicable to having a productively cluttered inbox. Use automated responses and templates. The automated responses are fairly commonplace now, but the templates in Gmail are a bit hidden, so read his instructions on how to set them up. Seriously, the templates can be life-changing. The last time I hired someone, the “Thanks, but no thanks” template cut the time needed at least in half. (It also kept me from sending snarky replies to people who think that “loving books” qualifies them for a library position requiring a master’s degree.) Manage your subscriptions. Yep, use that unsubscribe button on marketing emails. If you are in a position like mine where buying is part of your job and the marketing might be useful, see if your marketer has segments of emails you can unsubscribe from. Make use of starring and labeling emails. This is the final note I’m picking up from Boticelli. (hey, I’m a librarian — I’ve lectured undergraduates about plagiarism since before most of you were born. Over-zealous citation is in my blood.) In the process I describe below, you’ll see how I use these to my advantage. Tiptoe through the Messy Mailbox Rather than create a listicle telling you about my amazing mailbox tricks, I’m going to show you what I do when I sort my email throughout the day. Yeh, I’m not a one-time-a-day emailer, and I do check my email first thing in the morning. I turn off notifications, but I glance at mail whenever I am shifting from one task to another throughout the day. I like breaking the time management guru’s rules (selectively). My morning email routine goes like this:
https://medium.com/curious/zero-inbox-wastes-time-7d03b6a6e8b5
['Marie Jones']
2020-12-21 21:17:44.539000+00:00
['Technology', 'Email', 'Productivity Hacks', 'Gmail', 'Time Management']
1,996
Blockchains in Space with SpaceChain’s Alex Albano
In this two-part episode we talk to Alex Albano, COO of SpaceChain. Alessandra has been advising blockchain projects since 2017. She joined SpaceChain after having been COO of DEX, a decentralised data exchange, and having been part of the founding team at Ocean Protocol, a decentralised data sharing protocol, in charge its ecosystem growth and operations. Oxford Said Business School and INSEAD executive programs alumna, Alessandra specialises in sustainable growth strategies and operational excellence. She brings with her a wealth of experience in managing high-value contracts and large-scale teams in different sectors and at global level. As one of the early innovators in tokenization, she is an established thought leader in ecosystem growth, she’s a published author and the founder of AcchaLabs.ai, a company that advises traditional companies on incorporating tokenization in their business model through incremental innovation. Boasting more than a decade of experience in leading operations and marketing between Europe, India and Asia, Alessandra is in charge of SpaceChain’s global operations and strategy and leads its ecosystem growth. Keen to champion more diversity in the space and blockchain sectors, Alessandra is an advocate for women in leadership roles in STEM, a mentor, and a speaker.
https://medium.com/clayming-space/blockchains-in-space-with-spacechains-alex-albano-561ce43373b7
['Clayming Space']
2020-07-15 07:42:31.433000+00:00
['Blockchain', 'Ethereum', 'Space Exploration', 'Space', 'Blockchain Technology']
1,997
Spotify’s #2020Wrapped Is Sublime Marketing at Its Finest
Spotify’s #2020Wrapped Is Sublime Marketing at Its Finest #2020Wrapped goes big on bold fonts and wild colors. Source: Spotify. It’s that time of the year again. Social media platforms are abuzz with Spotify’s annual Wrapped roundup, #2020Wrapped. It’s a flashy rear-mirror view of the most popular songs of 2020, both on a platform level and on an individual level for its 320 million-plus monthly active users. And while it is certainly a heartwarming look at the artists you’ve loved and the music that helped you get through this whirlwind of a year, #2020Wrapped is more than that. Spotify gets to pull a not-so-secret Santa on its users, knowing full well what their wishlists look like. But the move also grants it an organic boost in visibility that far exceeds that of competing streaming services. Be it embarrassment or nostalgia, Spotify’s numbers elicit a personal reaction. It forms a connection with users on a deeper level this year, one of masked faces and long distance relationships with people across the street. Say what you will about Spotify meddling with the economics of the music industry. There’s no doubt that #2020Wrapped hit different. The best marketing appeals to the fundamental desire of who we are. And what better way to do it than to hand our music choices to us on a silver platter? Identity is cause; brand is effect Larry Ackerman’s quote on marketing does a good job of summing up how Spotify taps into our desire to express ourselves. Music is an acquired taste; one that listeners reinforce by finding favorite artists and genres to listen to. This very taste is further validated by Spotify’s personalized stats. It satisfies an itch in a way few other luxuries do. Sharing your music tastes with the world is like sharing a part of you, your personality, and your lifestyle. Spotify Wrapped is about you. Your artists, your songs, your decade. By making the consumer the hero, Spotify sings your story. You’re the muse. You made their year. By appealing to the human desire of being appreciated for who you are, Spotify lets you sing your own praises on social media. Being genre-fluid or having a personal list of favorite artists or titles appeals to a distinct sense of accomplishment intrinsic to how society perceives differences. Unintrusive organic marketing: the best form of marketing Tapping into the psychological needs of your users creates content that people just can’t help but share. Sure, there will always be that one person who doesn’t want to be basic or follow the herd. Or someone might just want to keep their unadorned tastes in music to themselves. But when your feed is filled with people sharing their music tastes, be it conventional or bizarre, the appeal is pretty hard to turn down. Spotify doubles down on the user’s intent of sharing a piece of themselves with the world by making the sharing process frictionless. A couple of taps is all it takes for a template to show up with the stats that matter in a format that is ready to share on your platform of choice. You even get to pick from a number of color palettes to better define your identity. And while the Spotify logo does show up, it’s not about them. It’s about you. Smooth transitions embody Spotify’s sense of flair. Source: Spotify. In 2020, Stories are everywhere Spotify’s nuggets of vivid colors and bold expression show up in a mini-train that is familiar to billions today. With platforms like YouTube and LinkedIn adopting the little circles from Snapchat, Stories are a mainstay in how we perceive social media as an extension of our lives. Call it inspiration or a knock-off; keeping things familiar helps keep things personal. While the colors reel you in, it’s the information that keeps you going from one screen to another. Getting to know how many hours I sunk into Krewella’s album gave me a sense of pride. The best part? I was among the top 0.5 percent of their fans in terms of time spent listening to the Pakistani duo’s sick beats. It’s no surprise that it showed up on my Instagram story shortly. Speaking of which, frictionless sharing really does oil the marketing machine like nothing else. Of known unknowns and unknown unknowns Populating your feed with #2020Wrapped posts and memes helps Spotify in more ways than you think. Sure, it helps their brand image. But it also ignites a sense of discovery in those who swipe past people showing support to their favorite artists and genres. While it might be a heyday for those who want to slide into people’s DMs, it’s a godsend when it comes to learning more about the kind of music that the people around you listen to. Finding a fellow Hamilton fan is a moment worthy of celebration in itself (I have yet to board said bandwagon). It’s a refreshing break of pace from boring feeds that compete with dwindling attention spans. I decided to embark on a little soul-searching by listening to the top songs that featured on my friends’ #2020Wrapped posts. The results surprised me. As someone who leans towards EDM and videogame soundtracks, I had to rethink my tastes after listening to a couple of obscure instrumentals. The experiment also helped me find like-minded individuals who shared my taste in music. All in all, it was a great excuse to dive into a medium of expression that has something for everyone. Spotify is where the party’s at The pursuit of the self is platform-agnostic but Spotify begs to differ. If you told me that something as simple as a roundup of annual stats would be enough to spark interest in a rival streaming platform, I’d have found the claim bizarre. Fast forward to a pandemic-stricken 2020 and the idea makes sense. Spotify Wrapped hit such a level of widespread sharing across platforms that competitors like Apple Music got left in the dust. The sense of FOMO was bad enough that CNET had to write an article on how Tidal, YouTube Music, and Apple Music users could join in on the fun. Wrapped didn’t just help Spotify’s brand image, it brought home bags of cash too. Stocks jumped up by 16 percent to an all-time high on a wave of powerful engagement from tens of thousands of Spotify users sharing their best moments on the platform. By opening up a generic roundup of 2020’s music to non-Spotify users, it tapped into an even larger audience of music enthusiasts. Spotify opens its doors to non-Spotify users too, albeit with a generic list. Source: Spotify. Visual language that goes loud True, your favorite songs playing in the background should be enough to warrant your attention. But Wrapped’s large text and interactive transitions set the stage for the numbers to shine. With elements that had a snazzy parallax effect between layers, Spotify added dimensionality to its simple-yet-bold visuals. Wild color combinations helped set the tone for a roundup that went big on self-expression. It’s safe to say that their design team knew what they were doing. The quizzes placed between the bits of information served both as a personal test and as a vehicle of curiosity. How well do you know thyself? After all, you are what you listen to. At least, that’s what Spotify would have you believe. Wrapped presents you with a rewind to remember This year is one for the books, for better or worse. The COVID-19 pandemic has shaped billions of lives, drastically altering the way we see our imminent future. Gone are the rosy dreams, replaced with shards of hope scattered about in the rubble. But that doesn’t mean that 2020 didn’t have its share of positive memories. And your playlists are a pretty good embodiment of that sentiment. True, my list was warped by a playlist that accompanied my workouts and your family members might have turned the suggestion algorithm on its digital head. Nonetheless, it still is a remarkable collection of memories. It is a reflection of how far you’ve come. And no, I’m not talking about old favorites that are now cringy to listen to. Music has the ability to evoke memories, both good and bad, as well as situations that you’ve faced in the past. It scratches the right itch by asking you to embrace 2020 instead of forgetting it. It is what it is, after all. Wrapped Spotify might have the best music algorithm on its hands but if they don’t get the show on the road, it isn’t going to keep its crown for long. Harnessing its strengths to create an organic marketing powerhouse lets Spotify reach an even bigger audience, one that regular marketing efforts might not reach. Finetuning the sharing experience and playing along with multiple distribution channels nets them a win that looks simple on the surface. With an annual roundup that makes it about you and your identity, Spotify gives you something you can’t help but share with the world. It empowers the individual while bolstering its brand image at the same time. Is that not what marketing is all about?
https://medium.com/swlh/spotifys-2020wrapped-is-sublime-marketing-at-its-finest-67d5d5a9c46e
['Antony Terence']
2020-12-04 07:30:57.227000+00:00
['Gadgets', 'Design', 'Social Media', 'Marketing', 'Technology']
1,998
The Work of Art in the Age of Algorithmic Reproduction
Anna Ridler’s Fall of the House of Usher unspools, rooms and bodies spreading half-seen across the frames of this 12-minute film like gossamer. A woman appears to walk down a hallway, then melts into a moonlit sky. A face appears in the dark, contorts into shapes. The animation is based on a 1929 film version of Edgar Allen Poe’s story, but its inky and strange visuals are the result of something altogether more modern: machine learning. Each moment of Ridler’s film has been generated by artificial intelligence. The artist took stills from the first four minutes of the 1929 movie, then drew them with ink on paper. These versions were then used to train a generative adversarial network (GAN), teaching it what sort of picture should follow on from another. The GAN uses this information to create its own procession of stills, based around a pair of networks that work in competition with each other — one as a generator, one as a discriminator, evaluating the work of the former like an algorithmic critic. ‘Fall of the House of Usher,’ by Anna Ridler. Photo: Anna Ridler The result is an AI-generated animation based on drawings that are based on the opening minutes of a 1929 film, which is based on an 1839 short story about a decaying lineage. It is a project that uses machine learning techniques not to showcase the technology, but as a way to engage with ideas of memory, the role of the creator, and the prospect of degeneration. It is primarily an artistic work, leveraging artificial intelligence as a medium in a way another artist may use acrylics or videotape. “By restricting the training set to the first four minutes of the film, I was able to control to a certain extent the levels of ‘correctness,’” Ridler explains. “As the animation progresses, it has less and less of a frame of reference to draw on, leading to uncanny moments that I cannot predict where the information starts to break down, particularly at the end of the piece. I deliberately take the ‘decay’ offered by making an image in this way and turn it into a central part of the piece, echoing the destruction that is so central to the narrative.” ‘Fall of the House of Usher,’ by Anna Ridler Ridler is part of a new wave of artists who are adept at coding and plugged into the nascent field of machine learning. If neural networks have largely been the domain of the computer science community, projects like Fall of the House of Usher are efforts to reframe these cryptic technologies as both artistic apparatus and important subject matter. After all, talk of adversarial networks may sound obscure, but these are techniques that lie beneath the interfaces we swipe and stroke on a daily basis, from video games to photo recognition on Facebook. “Given the implications on our society that machine learning already has, and will increasingly have, it is crucial that people investigate and question this technology from all possible angles,” says Mario Klingemann, currently artist in residence at Google Arts and Culture. “Artists tend to ask different questions than scientists, businesspeople, or the general public,” Klingemann adds. “Artists also might be in the right position to interpret or extrapolate the possibilities and dangers that machine intelligence harbors and express their findings in a language that many people can understand.” Bodies Built by Machines Klingemann’s work, much like Ridler’s, hinges between artificial intelligence and human bodies. His 2017 collaboration with Albert Barqué-Duran, titled My Artificial Muse, for example, resulted in an oil painting of a neural network–generated “muse,” itself based on a training set of classic paintings, including John Everett Millais’ Ophelia. Earlier this year, another project, Alternative Face, involved training a neural network to generate controllable faces based on the French singer Françoise Hardy. Klingemann used this to make it seem as if Hardy was speaking the words of Trump counselor Kellyanne Conway during her infamous “alternative facts” interview. ‘Alternative Face,’ by Mario Klingemann The ability to put the words of one person into the face of another is an unsettling illustration of the scope for machine learning to undermine the truth of what we see on our screens. Klingemann has continued to mine this seam, regularly posting neural network–generated faces on Twitter that bring to mind the abject self-portraits of Cindy Sherman. I asked him if he considers working with artificial intelligence in this way to be a form of collaboration. Klingemann told me that it’s closer to playing an instrument that he happens to build himself. ‘Alternative Face,’ by Mario Klingemann. Photo: Mario Klingemann “Admittedly it is a very complex instrument that at times seems to have its own unexpected behaviors, but with more practice and experience, outcomes that at first seem to be unexpected or surprising become predictable and controllable,” he said. For Klingemann and Ridler, engaging with artificial intelligence means understanding the meat and potatoes behind neural networks: algorithms and code. But other artists are also tackling these concepts and asking the same questions about invisible, intelligent systems, but from a different angle. Lauren McCarthy’s work, for example, frequently involves substituting software for herself. In Follower, volunteers were invited to download an app that granted them a real-life follower for one day. At the end of the day, the volunteer was sent a picture of themselves, taken by the follower. In a more recent project, LAUREN, McCarthy took on the role of an artificial intelligence assistant, akin to Amazon’s Alexa. Volunteers allowed the artist to install cameras, microphones, and smart sensors in their homes. Over the course of three days, LAUREN studies their habits, takes orders, makes recommendations, and controls everything from bathroom taps to door locks. It’s a purposeful inversion of the networks purported by Amazon, Google, and Apple—something McCarthy calls a “human intelligent smart home.” ‘Followers,’ by Lauren McCarthy. Photo: Lauren McCarthy/David Leonard “I’d say that I am trying to remind us that we are humans in the midst of technological systems,” McCarthy told me. “My interest is in people, not technology. What it means and what it feels like to be a person right now is changing quickly as the systems around us evolve, but there are also some parts of the human experience that remain constant through it all. I think this is the question as we think about ourselves in relation to machines. Where is the boundary of what we consider ‘human’?” Editorial Control One advantage of McCarthy’s approach is that by using human performance in favor of artificial intelligence, she skirts a potential pitfall in terms of patronage. Because AI is such an emerging area in art circles, there isn’t much set up for it in the way of funding streams. This means artists are often reliant on commercial companies to offer funding or technical expertise, and this, McCarthy suggests, could limit the subjects that tech-heavy projects are given license to confront. ‘LAUREN,’ by Lauren McCarthy. Photo: Lauren McCarthy “On one hand, I am happy to see corporations recognizing the potential for art to explore these topics and putting money toward it,” McCarthy says. “However, we need to be careful. Google and other companies providing these funding streams means that they have ultimate editorial control. It is unlikely that we will see work come from it that includes strong critique of AI, political provocation, or questioning of technologies developed by the companies.” This October, Arts Council England announced a new pot of funding for arts organizations working in the relatively new field of virtual reality, so it follows that machine learning–based projects could similarly be included in future publicly funded initiatives. There’s also the sentiment, however, that it’s ultimately more crucial to interrogate the systems of power this type of technology facilitates, rather than the technology itself — regardless of whether an artwork uses a generative adversarial network or a human sitting behind a monitor, watching a man brush his teeth. If AI is to be part of our lives, art should be there to meet it. “Most artists are dealing, in one way or another, with the experience of being a person right now,” McCarthy says. “Technology is a force that affects almost every aspect of this experience, whether we feel it directly or not.”
https://medium.com/s/living-in-the-machine/the-work-of-art-in-the-age-of-algorithmic-reproduction-bd3bd9b4e236
['Thomas Mcmullan']
2017-12-12 17:01:01.447000+00:00
['AI', 'Artificial Intelligence', 'Technology', 'Art', 'Machine Learning']
1,999
Cryptocurrency drop explained & why I expect it to drop further
It’s been a while folks! Counterintuitive isn’t it? Seeing an article suggesting a further drop in the Cryptocurrency market from a pro-Blockchain & Crypto guy is not what you would expect. My opinions over the long term viability of the solid projects associated with the technology are not altered (check the rest of my articles for more details around this), and in fact, I do consider the recent drops in the Cryptocurrency Market Cap to be positive for the long term potential of the sector. However, I do expect an even further drop in the Cryptocurrency Market for the short term. If you do not want to lose your investment, do read further! Market Capitalization of the Crytocurrency Market, courtesy of Coinmarketcap. Regulations, Regulations and even more Regulations Looking at the Cryptocurrency Market Capitalization chart above, shows a scary picture — from the market peaking at $820 billion on 7th of January, to dropping as low as $400 billion by the 2nd of February, we are talking about a more thn 50% drop in less than a month. Why did such a drop occur? As the title says, January has been a month where heavy regulatory pressure from governments has befallen Cryptocurrencies. From closure of exchanges in South Korea which do not comply with certain rules, to banning of Cryptocurrency trading in China and quotes of the Indian Finance Minister showing a negative stance of the Indian government towards Cryptocurrencies, has resulted in such a huge drop in the Market. As regulations result in more taxation of Crypto gains, less anonymity and more scrutiny it is normal that Cryptocycurrency traders are not reacting favorably to such news, and are quickly selling their coins to invest them in different asset classes. Will the Market drop further? Even though nobody can predict the future, I do expect the market to continue dropping in the short term. The reason is, that announcements of further regulations regarding Cryptocurrencies are expected to keep on coming. In the World Economic Forum at Davos that occurred last week, the US and UK Governments clearly expressed that they are planning on regulating Cryptocurrencies. And as recent history shows, this is going to have a negative effect in the market. Will the Market ever recover? With that being said, I do expect that in the long run, the solid Crytocurrency projects do have a future. As the market becomes more regulated, its legitimacy will increase and become more mainstream, allowing investors that are initially skeptical to join the market. The End of the Beginning To summarize, I do not expect that the market will instantly recover from the drops that it has suffered in January. Even if it does, I expect it to then soon retract to lower levels than it is today. Will that be the end of Blockchain and Cryptocurrencies? Definitely not, it will only mark the beginning of the asset’s widespread adoption. What is your take on the future of the Cryptocurrency Market? Feel free to share your opinion in the comments section or in a personal e-mail. Also, if you liked the article, don’t hesitate to share it with friends and family.
https://medium.com/cryptolinks/cryptocurrency-drop-explained-why-i-expect-it-to-drop-further-ff4579859e76
['Petros Leandros']
2018-02-03 17:47:05.568000+00:00
['Technology', 'Blockchain', 'Cryptocurrency', 'Ethereum', 'Bitcoin']