text
stringlengths
234
589k
id
stringlengths
47
47
dump
stringclasses
62 values
url
stringlengths
16
734
date
stringlengths
20
20
file_path
stringlengths
109
155
language
stringclasses
1 value
language_score
float64
0.65
1
token_count
int64
57
124k
score
float64
2.52
4.91
int_score
int64
3
5
The word “hacker” has negative connotations. You hear that word, and you think stolen data, lost money, lawsuits, and endless trouble. And it’s true: You do not want your business to get hacked. Unless you arrange for it to happen. Ethical hacking is a process by which a business hires an individual or company to act as a hacker would. An ethical hacker, or “white hat,” discovers system vulnerabilities that could be exploited by someone with ill intent. Then your business can strengthen these weak points before suffering negative consequences. How to begin the process Businesses should be cautious during the hiring process. Don’t give anyone carte blanche. Only select a certified white hat who will stay within the boundaries of the project. Your business should know how the ethical hacker will assess your system’s security. And once the job is done, the ethical hacker should recommend next steps. It’s a good idea to construct an agreement that outlines the scope and requires the ethical hacker not only to disclose discovered vulnerabilities and potential fixes but to ensure that sensitive data stays private. If that seems like a tall order, providers of IT support in Denver, Business Network Consulting Systems, can help your business craft a strategy. While it’s up to your business to stipulate the type of work an ethical hacker will do, there are certain security issues that they commonly discover. For example, SQL injection attacks are a common way for malicious hackers to access sensitive data. Ethical hackers look for security gaps in application code that would allow bad actors to inject their own code. Other typical vulnerabilities include poorly configured firewalls, inadequate security permissions, and weak or broken authentication processes. Fixing the problems Ethical hackers will report their findings to your company. Ordinarily, they will also include recommended methods for shoring up your security based on their investigation. Ordinarily, ethical hackers do not, however, implement the security improvements that they recommend. It is the organization’s responsibility to do so. This is sometimes where businesses drop the ball. It’s not enough that a business knows where it may be vulnerable; those gaps must be filled. Time is of the essence when it comes to fixing the identified problems. If an ethical hacker found vulnerabilities, malicious hackers will eventually do the same. Work with your existing IT team or add strength to its ranks by consulting with BNC. Its experts have encountered nearly every security weakness and keep abreast of new tactics malicious hackers use. Once isn’t enough Knowing that these chinks in your network security exist can be unsettling. Naturally, you want to plug whatever holes your ethical hacker may uncover. But even after you’ve gone through the whole process, you cannot rest on your laurels. That’s because malicious hackers don’t rest. They’re constantly seeking new ways to penetrate business systems. It can be difficult to determine how often your business should test its systems for vulnerabilities. BNC can guide you through the whole process from start to finish, as well as helping plan for the future.
<urn:uuid:8b7d9b5f-59d4-45cd-a3be-c99aef7a51bf>
CC-MAIN-2022-40
https://www.bncsystems.com/blog/what-businesses-should-know-about-ethical-hacking/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030333455.97/warc/CC-MAIN-20220924182740-20220924212740-00480.warc.gz
en
0.931575
654
2.546875
3
A group of researchers led by J. Alex Halderman, assistant professor of electrical engineering and computer science at the University of Michigan, have been working on a new anti-censorship system. It is called Telex and – in theory – it should allow users from countries that practice Internet censorship to not only reach their wanted Internet destination, but also hide from the censors the fact that they are trying to do so. “Telex operates in the network infrastructure — at any ISP between the censor’s network and non-blocked portions of the Internet — rather than at network end points,” explains Halderman. “This approach, which we call ‘end-to-middle’ proxying, can make the system robust against countermeasures (such as blocking) by the censor.” The advantage that Telex would have over other existing anti-censorship systems is the fact that the user doesn’t have to get ahold of ever changing encryption keys or IP addresses for proxy servers, but simply use the client software that would be made publicly available. “The client secretly marks the connection as a Telex request by inserting a cryptographic tag into the headers. We construct this tag using a mechanism called public-key steganography,” explains the researcher. “This means anyone can tag a connection using only publicly available information, but only the Telex service (using a private key) can recognize that a connection has been tagged.” The researchers believe that Telex could complement anonymizing services like Tor, but in order to work, the cooperation of ISPs outside of the censoring countries is needed. “Telex doesn’t require active participation from the censored websites, or from the non-censored sites that serve as the apparent connection destinations. However, it does rely on ISPs to deploy Telex stations on network paths between the censor’s network and many popular Internet destinations,” says Halderman. “Widespread ISP deployment might require incentives from governments.” He admits that that Telex is still far from ready to be used by real users, but that they have been testing it for the last couple of months using a mock ISP they have set up in the lab, and that they are satisfied with the system’s performance and stability. Of course, there are still a lot of questions that need to be answered, and a lot of things to be worked out in order to make this solution secure enough to use, which is probably why they have released details about their ongoing work on the Internet. Judging by the comments Halderman’s post has received, commenters are eager to point out what they consider the solution’s weak spots, and this is something that can only help the researchers in their future work.
<urn:uuid:4756a676-35fa-48f3-80a6-4551e80aaf55>
CC-MAIN-2022-40
https://www.helpnetsecurity.com/2011/07/19/a-new-approach-to-circumventing-state-level-internet-censorship/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334644.42/warc/CC-MAIN-20220926020051-20220926050051-00480.warc.gz
en
0.960576
584
2.671875
3
Data has become one of our most important assets. We usually store data in databases, so knowing how to secure those is of vital importance. This article can help you quantify the level of security of your databases on a scale of 1 to 10. CISOs and database administrators (DBAs) can use it to determine their security maturity level and identify steps to improve it further. Find your database security rating The table below summarizes security ratings on a scale of 1 to 10, with one being the lowest level of security and ten being the highest. All ratings are cumulative, so each rating includes all the requirements for all previous ratings. The order of the levels reflects both an increase in security and an increase in cost and complexity. Lower levels are achievable without additional software, but achieving higher ones becomes increasingly difficult and requires appropriate products. 1. No security Rating 1 is for databases that aren’t secured. All databases provide an inherent level of security, even if you made no effort to protect them. For example, they require credentials to connect and have roles and privileges. Managing data in a database is the first step toward securing it. 2. Standard security and least privileged Rating 2 is for databases where both the database and operating system are configured by following industry standards and best practices. This rating also requires that all database accounts are least privileged, which means that privileges granted to accounts are the minimum needed to perform their duties. As part of the requirement for Rating 2, there should be an effort to eliminate shared accounts. If shared accounts exist, they should not be in regular use, and their credentials kept secret. Limiting usage is especially true for privileged shared accounts built into the database. Database accounts such as SYS and SYSTEM in Oracle or SA in SQL-Server should not be used regularly, and their passwords should be kept securely with limited access. This requirement can be challenging when, for example, the application uses a privileged account or when shared accounts are part of how we operate. However, reducing and controlling the usage of these accounts is essential to security. 3. Change control and metadata snapshots Rating 3 is for databases that are under change control. That means that any changes to metadata such as users, privileges, configuration, and objects, should go through a change control approval process. As part of the requirement for Rating 3, a daily snapshot of the configuration, users, privileges, and object metadata need to be made. Changes between snapshots should be promptly investigated and approved. It is also advisable that these snapshots are cross-compared to similar databases to ensure consistent and uniform configurations, users, privileges, etc. The challenge with change control is that it can be cumbersome and seen as useless red tape. However, lack of control over metadata changes quickly becomes a lack of control over the data. 4. Session monitoring and review Rating 4 is for databases where all logins are monitored and regularly reviewed. Logins from unexpected users, programs, or machines, should be promptly investigated. One of the easiest ways to breach database security is through credential theft. For example, stealing a DBA username and password would grant an attacker unlimited access to the data. Monitoring logins allows mitigating this risk. Most databases allow auditing of logins and failed logins with minimal overhead. The implementation challenge is to provide an efficient review of the information via reporting. 5. Basic SQL auditing (DDL & DML) Rating 5 is for databases where high-risk SQL activity is regularly recorded, reported on, and reviewed. High-risk SQL activity includes: - All DDLs (including DCLs) – SQLs that modify the database configuration, objects, users, privileges, etc. - DMLs from unexpected sources such as privileged users and particular programs The objective of this requirement is to apply control to activity that is infrequent and high-risk. Auditing rare activities do not usually create a performance overhead and requires a minimal time investment. The implementation challenge is to allow for a prompt and efficient review of the activities. 6. Full SQL auditing and network encryption Rating 6 is for databases under full SQL auditing where all potentially risky SQL activity is regularly recorded, reported, and reviewed. That translates to auditing large amounts of activity, including queries. For example: - Access to sensitive tables with Queries and DMLs - All DBA and privileged user activity - All the activity from high-risk programs such a SQL Plus, Management Studio, etc. - Activity by the application account that is not from the application servers - Sensitive activity even when performed inside the database by stored procedures or triggers Rating 6 also requires all database network activity to be fully encrypted to prevent network sniffing and spoofing. The objective of this requirement is to start applying strict security measures on SQL activity and prevent many network attacks. Encryption of network activity comes built-in for free in most databases and is easy to turn on. The main implementation challenge in this requirement is that auditing too many activities without an appropriate solution can have a significant database performance impact. A secondary challenge is to achieve efficient reporting for prompt review of the information with minimal time investment. When searching for an auditing solution, note that some products do not avoid the database performance overhead, while others do not support network encryption. 7. Session anomaly detection and alerting Rating 7 is for databases with automated detection and alerting on anomalous activity sources. Unlike the manual session review performed in Rating 4, this rating requires automation capable of detecting and alerting on changes in activity source profiles. - Alerting on new users, programs, machines, or a combination of those connecting to the database - Detection of shared accounts (accounts used by several individuals) as well as individuals using multiple accounts The purpose of this requirement is to supplement the human review with automation that will notice and highlight anomalous activity. That helps avoid accidental oversight as well as ensure quick detection and response. Implementation requires an appropriate solution to perform the analysis. 8. SQL anomaly detection and alerting Rating 8 is for databases with automated detection and alerting on anomalous SQL activity. Unlike the manual SQL reviews in Rating 5 and 6, this rating requires automation capable of analyzing all the SQL activity in the database, including the application activity. - Unusual application behavior like potential SQL injections - Abnormal activity levels. For example, an unusually high number of SQLs executed or rows accessed - Activities at an odd time of day - New SQLs touching sensitive tables The objective of this requirement is far beyond avoiding accidental oversight and improved detection time. The aim is to apply control to the impossibly high activity volume in databases that cannot undergo human review. Even low activity databases can execute millions of SQL queries per day, and without automation, it is impossible to apply any level of control to them. Implementation requires software that can capture all the activity with low overhead and perform the analysis. 9. Proactive forensic reviews Rating 9 is for databases with regular proactive activity reviews. That means that a person familiar with the activity profile of the databases regularly inspects the activity (for example, once a month). The purpose of the review is to identify behaviors that might otherwise go undetected, including both internal abuse and external attacks. The inspection could also highlight gaps in the controls, risky practices, and more. Implementation requires a solution that can capture all the activity with minimal overhead, reduce and store it in a reasonable amount of disk space, and provide the forensic tools to analyze and review it. 10. Restrict access to DBAs and applications Rating 10 is for databases that restrict access to accounts with otherwise unlimited access to the data, for example DBA accounts, privileged accounts, and the application account. Such restrictions are not usually part of the native database capabilities and could include: - Preventing privileged accounts from accessing schemas, tables, or objects that they should not be accessing. For example, DBA accounts should not usually access the data. - Preventing access to an account from programs or machines that should not access it. For example, only the application program and the application server should access the application account. - Preventing access to an account on days and at times when it should not be in use. - Preventing accounts from accessing more data than is expected (rate-limiting) - Enforcing separation of duties by requiring security personnel to pre-authorize certain privileged activities. This requirement requires a solution to implement since it goes beyond built-in database preventive controls. Applying preventive controls to databases poses the operational risk of blocking legitimate activity. Deploying such measures must, therefore, be done carefully following appropriate best practices to minimize the potential for disruptions. If your security rating is not as high as you thought, don’t despair. Database security requires persistent pursuit that will gradually improve your rating. Database security is an essential investment and will give you the confidence that your data is safe, secure, and reliable.
<urn:uuid:745e133d-97ff-42a8-aed0-eb91420e789a>
CC-MAIN-2022-40
https://www.helpnetsecurity.com/2021/09/14/evaluate-database-security/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335004.95/warc/CC-MAIN-20220927100008-20220927130008-00480.warc.gz
en
0.915698
1,866
2.59375
3
A Virtual Private Server (VPS) is a cross-functional virtual operating system that resides within the host server and effectively mimics a dedicated server environment despite being on a shared physical server. The usage of VPS hosting has increased exponentially because it is less expensive than dedicated hosting and provides better security protocol, performance, affordability, accessibility, control, data protection, flexibility, growth, customization, and reliability than shared hosting. Tips and Tricks for securing VPS structure To avoid cybercriminals and fraudsters to attack your data files, Virtual Private Server (VPS) can be secured in the following ways: Disable Root Logins Cybercriminals and fraudsters try to unlock the login credentials and gain access to the server. Disabling the password from the “root” username enhances security protocol and the cyberattackers will not be able to guess your login details. Therefore, creating another username to execute root-level instructions is recommended. Use Strong Passwords It is easy to guess weak login credentials that contain information related to identity. Passwords containing upper case letters, special characters, and numbers can secure your account from any cyber threat. It is also recommended to not reuse the same login passwords. Bitwarden and Lastpass are online security tools that can also be used to create strong passwords. Configuring an Internal Firewall Configuring an internal firewall helps the user to avoid undesirable and malicious traffic to gain access to your system and also helps to defend the distributed denial of service attacks (DDoS). Change Default SSH Port Switching the SSH port number can avoid hackers to connect directly to the default number (22).To change the SSH port number, you will have to open up /etc/ssh/sshd_config for appropriate settings. Prioritize the use of SFTP instead of FTP Cybercriminals can cause a sniffing attack to gain access to your login credentials if an outdated File Transfer Protocol (FTP) is used, even while using “FTP” over “TLS” (FTPS). To avoid cyber attacks, use “FTP” over “SSH” or (SFTP). Install an Antivirus Software Setting up an internal firewall is the first line of defense that can deny access to any malicious activity, but it is not a foolproof security protocol. Installation of antivirus software is needed to further enhance security. ClamAV is an open-source antivirus that is most commonly used for the detection of any malicious activity. Connect to your VPS via VPN Most VPS are configured simply by exposing web or app services directly to the Internet. If your VPS is only accessed by end-users within your enterprise environment, you should consider implementing a LAN to LAN VPN between your IT environment and the VPS server. Then you should implement a firewall and only allow the VPN ports. Every other service should be tunneled via the VPN service. Avoid Anonymous FTP Uploads It is important to edit your server’s FTP framework to disable unidentified FTP uploads. Because it can cause a huge cyber threat and make the system vulnerable to security issues. In most cases, cybercriminals send malicious traffic through IPv6 to gain access to the system. And make the user more susceptible to hacking attacks. Upgrade the settings like NETWORKING_ IPV6=no and IPV6INIT=no. Securing Offsite Backups A significant data loss will occur if you keep the backup system on a similar VPS. To prevent further data loss and security breaches, it is recommended to store your backups on a remote server. Install a rootkit Scanner Rootkit is an important component of malware that works below other security tools and permits undetected access to a server. To fix this problem, reinstall the OS (Operating System). Use GnuPG Encryption It is important to encrypt transmissions to your server because cybercriminals attack data files while it is in transit over a network. Encryption can be done by using passwords, certificates, and keys. For that purpose, GnuPG, an authentication system, can be used to encrypt transmissions. Use SSL Certificates For Everything To ensure privacy, it is helpful to use SSL certificates that create an encrypted channel between the server and the client. Unlike VPC, Virtual Private Server (VPS) are exposed to the Internet, which means that these cloud products are exposed to a lot of cybersecurity threats that need to be defended to mitigate the chances of a security breach. It is very important to know about every perspective of security threats, especially on a self-managed VPS. Most companies running their business online have basic security plans, which are not effective enough to stop penetration attacks. Therefore, IT admins must know how to implement the best security measures such as disabling root logins, ensuring strong passwords, configuring a firewall, using SFTP instead of FTP, changing the default SSH listening port, using antivirus software, using VPN for your VPS, disabling IPv6, avoid uploading anonymous FTP, securing offsite backups, updating the system on regular basis and by using Spam filters. GDMS Infrastructure as a Service offerings (IaaS) solutions allow our customers to control their own data infrastructure without having to physically manage it on-site. Find out more about our VPC and VPS Services.
<urn:uuid:9f4094c1-988b-484b-a7e2-9edc91d883e3>
CC-MAIN-2022-40
https://www.global-dms.com/tag/vps/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335424.32/warc/CC-MAIN-20220930020521-20220930050521-00480.warc.gz
en
0.87943
1,120
2.5625
3
Phishers launched a record number of attacks in January 2006, as reported by the Anti-Phishing Working Group. These attacks often take the form of an email that purports to be from a trusted entity, such as eBay or PayPal. The email states that the user needs to provide information, such as credit card numbers, identity information, or login credentials, often to correct some alleged problem supposedly found with an account. Some number of users fall for these attacks by providing the requested information, which can lead to fraudulent charges against credit cards, withdrawals from bank accounts, or other undesirable effects. The first attempts at applying learning to these problems took the form of browser toolbars, such as the Spoofguard and Netcraft toolbars. Our research group is currently conducting a study to determine the accuracy of these and other toolbars more precisely. Preliminary results indicate that a large percentage of phishing emails make it past these toolbars. One of the biggest drawbacks from a learning perspective is that toolbars in web browsers have access to less information. In some sense, users have already partially fallen for the attack by clicking on a link in an email, and this could potentially expose the user to spyware and malware loaded by attacks on insecure web browsers. Furthermore, now that detection is being done in the browser, the contextual information and features from the email are no longer available. In theory, this reduced amount of information should cause a similar reduction in the accuracy that such systems are able to achieve. Aside from accuracy, toolbars suffer from a number of other problems. Ideally, phishing detection algorithms require minimal user interaction, either as a server-side filter to compliment existing spam filters, or as a filter running locally in the client. This approach has several benefits over other methods. First, by removing user interactions, there is no chance for the user to dismiss warning dialogs and proceed to provide information to an attacker. Second, contextual information is available in the email. Finally, by operating on the email rather than in the browser, server-side detection is possible. Server-side detection avoids paying a transmission cost for sending the email (and subsequent requests for images linked to in an HTML image) to the user, as well as the cost of evaluating certain features (such as doing WHOIS lookups, whose results can be cached for re-use). We present in this an algorithm, which we call “PILFER” – phishing identification by learning on features of email received. Our implementation is not optimal. It does not make use of all the information potentially available to a server-side filter. However, we obtain high accuracy rates, and posit that further work in this area is warranted. Click here to download the full whitepaper
<urn:uuid:93ee118e-de1f-494e-a1c0-dd8cf8ba2ee5>
CC-MAIN-2022-40
https://it-observer.com/learning-detect-phishing-emails.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337855.83/warc/CC-MAIN-20221006191305-20221006221305-00480.warc.gz
en
0.957601
559
2.8125
3
Why FTTO is Best for Future Green Networks Arafat Yousef, Managing Director, Middle East & Africa, Nexans Data Network Solutions, on how smart hybrid in-building networks can help reduce CO2. Rise of the digital economy The digital economy is growing fast. IDC has predicted that the digital economy will account for 60% of the world’s total economy by 2022. The United Nations Conference on Trade and Development estimates that the digital economy makes up anywhere from 4.5% to 15.5% of global gross domestic product (GDP). In the US, home of several dominant players, the digital economy already accounted for 6.9 percent GDP ($1.35 trillion) in 2017. In China, the digital economy accounted for a third of the nation’s GDP in 2018. In Europe, ‘digital’ is among the fastest growing sectors and “a major contributor to economic prosperity”, according to E&Y. However, as the digital economy grows, so does consumption of energy and production of heat and CO2. Hyperconnectivity and energy consumption The digital economy is based on hyperconnectivity. More and more people, devices and organisations are interconnecting, driven by developments such as 5G, Cloud, Wi-Fi 6, Internet of Things (IoT) and more. In all of these areas, vast growth is predicted. According to IDC some 212 billion IoT-enabled devices may be connected to the internet soon. Cisco predicts that by 2021, there will be 4.6bn Internet users worldwide and 271 billion connected devices. But as the number of devices, users, applications and networks grow, so will the global carbon footprint. According to the European Commission, the ICT industry generates up to 2% of all global CO2 emissions. The journal Environmental Science and Pollution Research the installation and the operation of new ICT devices can be characterised as highly energy intensive. A study carried out by the Global Enabling Sustainability Initiative (Gesso) and Deloitte shows that ICT can help solve a wide variety of sustainability challenges and climate change in particular. Big data, AI and intelligence built into devices and systems that consume power play an important part in this. Working from home facilitated by ICT, for example, reduced the need for travel. However, there is also a great deal to be gained by simply reducing the amount of power consumed by the devices, cabling and networks that provide the backbone for the digital economy. Furthermore, making sure systems can remain operational for as long as possible also reduces the burden on the environment. Less energy usage and longer operational lifetime One possible solution is the introduction of Fibre To The Office (FTTO) networks to connect devices throughout a building at Gigabit speeds. FTTO is a hybrid network consisting of fibre optic and twisted pair copper patch cords with connectors. In an FTTO network environment, fibre is laid from the central distribution switch right into the office floor, where it ends in active FTTO switches within the workplace. FTTO networks require significantly less equipment, consume less energy and require less cooling. In a traditional copper-based Ethernet network with 1,000 ports, the annual electricity consumption is estimated to be 82k kWh. However, using an FTTO architecture, the same network would consume some 30k kWh per year. A fast, low latency, responsive FTTO network with PoE (Power over Ethernet) capacity also supports the introduction of smart building management systems, helping reduce energy consumption by optimising lighting, air conditioning and heating usage. The use of fibre also allows compliance with 802.3az Energy-Efficient Ethernet’ standard. This allows each port on the switch to power down into a standby mode when no connected devices are active. Energy Efficient Ethernet (EEE) ports consume power only when data is transferred. Finding the right solution FTTO architecture can reduce energy consumption by as much as 70% lower, cut total cost of ownership by 40%, and reduce installation time by 60%. What’s more, the added flexibility means networks can accommodate changes in building layout and function and cabling can stay support consecutive generations of equipment. As a result, it isn’t necessary to dispose of large volumes of cabling and equipment every few years. The exact type of solution and best configuration depends on the on-site conditions of different buildings – we are happy to help you make the smartest decisions!
<urn:uuid:6ea9974e-b421-44d0-8c97-58fdfb48db56>
CC-MAIN-2022-40
https://www.cxoinsightme.com/opinions/why-ftto-is-best-for-future-green-networks/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337855.83/warc/CC-MAIN-20221006191305-20221006221305-00480.warc.gz
en
0.934829
913
3.015625
3
Infrastructure-as-a-Service (IaaS) enables enterprises to lease or rent servers to use for storage and computation in the cloud. Users have the freedom to run the operating system of their choice and any applications they like without worrying about the cost of maintaining the physical infrastructure. This gives companies the ability to access servers that are physically close to the end-users that need them, which can decrease latency issues and even allow for a serverless physical environment. Also, because IaaS can be scaled automatically in either direction (i.e., scale up or down), it gives users a greater sense of agility, particularly when the business undergoes change. What Is "The Cloud and the 3 Main Cloud Service Models"? The cloud refers to access to servers through the internet, as well as the databases and software that those servers run. A vendor gives its clients the ability to pay as they go as they access servers, storage, networking, and other resources inside the cloud. With Software-as-a-Service (SaaS), a provider delivers applications and software to users through the internet. Users can subscribe to the software and use it through the internet or an application programming interface (API) provided by a vendor. With Platform-as-a-Service (PaaS), the provider gives users cloud resources for building and delivering applications. Types of Infrastructure-as-a-Service (IaaS) Public IaaS includes solutions that run on a public cloud and are available to the general public, which means their resources are shared among many organizations. With a private IaaS architecture, your solutions are implemented on your own servers. Therefore, you have exclusive use of your IaaS and can benefit from dedicated access to its resources. Hybrid IaaS provides you with a combination of public and private cloud resources. Your IaaS solution is composed of multiple cloud environments, which may also include in-house servers. How to Implement Infrastructure-as-a-Service The implementation of IaaS can be through private, public, or hybrid cloud architecture. Customers have the freedom to alter the infrastructure through a graphical interface. They can also gain access to the infrastructure through an API key, which allows new servers to be incorporated when necessary. How Does Infrastructure-as-a-Service Work? With IaaS, customers use dashboards and APIs to gain access to the servers and storage they are renting. They may add resources as they see fit and experiment with different solutions to meet changing needs. What Are the Common Business Scenarios of IaaS? Customers can host a website using resources provided by an IaaS service. This eliminates the need for in-house servers. Test and Development Users can experiment with different environments as they design applications using Structured Query Language (SQL), Java, or any other language. This makes setting them up and taking them down easier than if they use their own, on-premises resources. Storage, Backup, and Recovery An organization can use IaaS for storage, backup, and recovery resources, which may be particularly helpful when the demand is unpredictable or if the needs are growing steadily as the business expands. With IaaS resources, you have access to powerful computing that may otherwise be economically infeasible. This enables the development, testing, and deployment of a larger range of applications, including software for memory-intensive applications like those for video editing, computer-aided design (CAD), and more. With IaaS, you can deploy some or all of your web applications within the infrastructure, making it easier to scale up or down as demand for the applications shifts. Big Data Analysis With greater access to more powerful resources, IaaS empowers users to examine huge sets of data. This includes scouring through data-sets to find patterns that can be used for artificial intelligence (AI). Advantages of IaaS Eliminates Capital Expense and Reduces Ongoing Cost With an on-site data center, an organization has to make considerable investments in setting it up and maintaining it. With IaaS, the provider handles those costs. When you are done developing a new application or offering, the infrastructure necessary for launch is ready almost right away, instead of reconfiguring your in-house systems over the course of weeks or months. Respond More Quickly To Shifting Business Conditions With the scalability of IaaS comes the agility to increase or decrease production quickly, such as during the holidays or seasonal events. Focus on Your Core Business Because you do not have to spend as much time with your IT infrastructure, IaaS gives you more time and energy to further core business objectives. Increase Stability, Reliability, and Supportability With an IaaS solution, you do not have to invest extra resources in supporting your infrastructure or making sure it is stable and reliable. The provider handles all of that for you. You can include security options in your service agreement with the cloud provider that are as powerful as any other cloud or multi-cloud security solution. Your IaaS security options may outperform what you can obtain for an in-house solution. Get New Applications To Users Faster There is no need to spend time preparing the infrastructure to deliver new applications. Your IaaS provider takes care of that for you. This allows you to deploy new solutions more quickly. Good Examples of IaaS Some of the top IaaS examples include: - Microsoft Azure: Microsoft Azure is known to fit a wide range of business solutions. - Amazon Web Services (AWS): Amazon’s offering is popular for its ease of use and impressive selection of tools. - IBM Cloud: IBM Cloud’s service stands out due, in part, to its Bare Metal-as-a-Service (BMaaS) option. This provides users with access to the hardware that supports their cloud solution. How Fortinet Can Help The Fortinet FortiGate next-generation firewall (NGFW) can protect an IaaS by shielding virtual machines from cyber criminals and malware. The FortiGate solution can be positioned between virtual servers and the internet. Using deep packet inspection (DPI), it searches through the data packets that attempt to go both in and out of the servers. FortiGate also incorporates AI, which allows it to detect zero-day threats, in addition to those that already have been profiled using cyber intelligence. In this way, the IaaS solution can be kept secure from a variety of threats, enhancing both security and the confidence of developers, IT admins, team leaders, and end-users.
<urn:uuid:c29d6dcd-a27f-46c1-ab9e-fb4a2fce622f>
CC-MAIN-2022-40
https://www.fortinet.com/tw/resources/cyberglossary/infrastructure-as-a-service
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337855.83/warc/CC-MAIN-20221006191305-20221006221305-00480.warc.gz
en
0.919073
1,447
3.015625
3
In a digitized world, one of the greatest conveniences is health information technology (HIT). Considering nearly everyone in the US is a consumer of healthcare, there are numerous benefits to adopting electronic health records. The evidence report, Costs and Benefits of Health Information Technology, outlines a few. Health information technology is a means of electronically storing, recording, accessing, or transferring a patient's medical records. This includes health and medical history. Not only is this paperless and perhaps more reliable, but it allows for "clinical decision-making and disease management." It also allows for prescription filling, test ordering, and care reminding. For example, the system can provide alerts for necessary patient vaccinations or send a prescription to be filled at a pharmacy convenient to the patient. Overall, health information technology improves the efficiency of healthcare - a highly profitable, nationwide business. However, adopting HIT is costly and requires change in the organization. It is considered an investment, but perhaps a necessary investment in terms of economic advancement. In non-financially focused studies concerning adoption of HIT, areas of improvement included increased productivity by the healthcare provider, improved patient safety and, subsequently, fewer adverse drug events (ADE) and time spent in hospitals to treat ADEs, and better physician decision-making. For example, the ability to reduce the "ordering of redundant clinical laboratory tests could produce an annual savings of $35,000 in laboratory charges." It is economically beneficial to improve the efficiency of healthcare. In a day and age where nearly everything is digitized, it is only fitting a business as widely used as healthcare should follow suit. Adopting health information technology will improve provider efficiency while increasing consumer centeredness. Electronic health records are more personalized, more organized and more efficient. Although implementation of HIT is an expensive adjustment, benefits for both provider and consumer are apparent. Healthcare efficiency is important to society as a whole. With efficiency of digital access to healthcare records comes the necessity of increased data security measures. The Federal Health Insurance Portability and Accountability Act of 1996, known as HIPAA, was passed to establish a national framework for security standards and protection of confidentiality with regard to health care data and information. Fortunately, Giva makes HIPAA compliance very easy for our customers. The data center, hardware and software infrastructure of Giva's cloud help desk & customer service software meet the very strict HIPAA compliance regulations. See 7 Key Elements of Giva's HIPAA-Compliant Cloud Help Desk Software for Electronic Health & Medical Records.
<urn:uuid:dccebd2c-f539-469e-87e1-a6982ecc209f>
CC-MAIN-2022-40
https://www.givainc.com/blog/index.cfm/2015/2/11/value-in-healthcare-information-technology-hit
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335286.15/warc/CC-MAIN-20220928212030-20220929002030-00680.warc.gz
en
0.943934
505
2.640625
3
Defining a Derived View¶ The administrator can use the base views of the system to define new relations. These new relations are called derived views. Derived views are created through the statement CREATE [ OR REPLACE ] VIEW <name:identifier> [ FOLDER = <literal> ] [ DESCRIPTION = <literal> ] [ <primary key> ] [ <tag list> ] [ ( <field properties> [, <field properties> ]* ) ] AS <select> [ USING PARAMETERS ( <parameter> [, <parameter> ]* ) ] [ ORDER BY <field name> [ ASC | DESC ] [, <field name> [ ASC | DESC ] ]* ] [ WITH [ CASCADED | LOCAL ] CHECK OPTION ] [ CONTEXT ( <context information> [, <context information>]* ) ] <field properties> ::= <name:identifier> ( <property list> ) [ <tag list> ] <parameter> ::= <fieldname:identifier> : <fieldtype:identifier> [<default:literal>] As you can see, a name and the query that defines it are specified, when creating a view. The query is specified using the syntax of the SELECT statement, which has been explained in detail in the section Queries: SELECT Statement. Therefore, the administrator can create new derived views by combining other existing views using operators such as unions, joins, cartesian products, selections, projections, group-by operations, intersections, the minus operation and the flatten operation. Furthermore, existing derived views can be also used to create new derived views, allowing view trees with as many levels as required. For example, considering the views A, B and R as base relations (those that directly access the sources to obtain their data) the administrator can define a view G as the join of the result of applying the union (A, B) with R, as can be seen in the figure below. ORDER BY clause indicates that when querying the view, the results will be ordered by those field(s). ASC sorts in ascending order and DESC, in descending order. If DESC are omitted, Virtual DataPort will sort in ascending order. The creation of a view also accepts the SQL standard clause WITH CHECK OPTION, which is related to the updating of view contents DELETE statements. The function of this modifier is described in detail in the section Use of WITH CHECK OPTION. The use of the OR REPLACE modifier specifies that, if there is a view with the name indicated, this must be replaced by the new view. Where, due to the change in view definition, the query capabilities of some derived views have been altered (e.g. due to the addition of another field or a query restriction that did not previously exist), Virtual DataPort will update the schema and query capabilities of the upper level derived views wherever possible. PRIMARY KEY clause sets the definition of the primary key of the view. See more information about primary keys in the section Primary Keys of Views of the Administration Guide. USING PARAMETERS defines the parameters of the derived view. See more about this in the section Parameters of Derived Views of the Administration Guide.
<urn:uuid:46b46652-4abf-4752-a529-691c301d6607>
CC-MAIN-2022-40
https://community.denodo.com/docs/html/browse/8.0/en/vdp/vql/defining_a_derived_view/defining_a_derived_view
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335444.58/warc/CC-MAIN-20220930051717-20220930081717-00680.warc.gz
en
0.861453
736
2.765625
3
As I’ve mentioned in a previous post, web applications are one of the most attractive targets for hackers because they can be easily reached, have virtually-countless attack vectors, and there are too many ways for naïve developers to goof and inadvertently introduce vulnerabilities in their source code. In this post, I’ll familiarize you with cross-site scripting (XSS), one of the oldest and most common types of attacks that can be staged against websites, and steps you can take to protect your site against it. What is XSS? Cross-site scripting is a security exploit in which the attacker tricks a web server into storing malicious client-side script and later insert them into webpages when other users request them. This type of attack has been around since the 90s and has at some point affected major websites such as Google and Facebook. XSS attacks can involve a wide range of malicious activities, including data theft, user session hijacking, malicious code execution, or they may be used as part of a phishing scam. With the introduction of rich internet applications (RIA), which rely heavily on client-side scripts and AJAX calls, the detection and prevention of cross-site scripting attacks has become even more complicated. Cross-site scripting is predicated on the need of websites to receive input users, and since you’ll find less and less websites that unilaterally deliver content to visitors without offering some sort of interaction mechanism, most websites are susceptible to have XSS vulnerabilities in one way or another. According to a survey by WhiteHat Security, 70% of website vulnerabilities are XSS-related. How does XSS work? The main difference between cross-site scripting and other types of attacks, such as SQL injection, is that they target the users of a website and not the host webserver itself. A simple XSS example Cross-site scripting is most effective in areas where users can post public comments or enter input which will be displayed to other users (such as profile information). For instance, the comments section of a blog or a forum usually includes text input which allow all users to write their opinion on the article or topic of discussion. Anything a visitor types in the input box will be stored in the server and later displayed to other visitors. A user with malicious intents can insert a malicious <script> tag in the textbox and submit the input. Subsequently, when other users request that page, the malicious code is sent to their browser and executed as part of the page. XSS attacks even work on websites that use SSL to encrypt their data exchange because they are executed in the context of the webpage’s DOM and do not account for malicious external data. How do you prevent XSS? In essence, the main cause for cross-site scripting vulnerabilities in websites is the lack of validation and verification mechanisms for user-generated input. Naively accepting all sorts of user input and later displaying them to other users will open the path for serious XSS attacks. Therefore the single most important step in XSS prevention is to assume all data received by the web application is from an untrusted source. This includes anything that comes from a client, including form data, cookies, files, images, headers… and it includes data from authenticated users as well. Not trusting doesn’t mean blocking the data, but rather verifying its format, type, length and range to make sure it is within safe and trusted boundaries, and to reformat, encode and escape the data to make sure it can’t harm other users when it is displayed to them. XSS prevention best practices Methods can be both applied on the client- and server-side to prevent script injection attacks. However, you should consider that in general, there’s no best practice on the client-side that can prevent script injection attacks, simply because all such safeguards can be circumvented by the attacker when sending requests to the server, and therefore client-side methods can only be relied on as ways to make XSS exploits harder for the attacker. Here are practices that can help mitigate and eliminate the threat of cross-site scripting attacks. Client-side validation cannot be relied upon as a security means to prevent XSS attacks but rather as a means to guide the user into entering correct input. Server-side validation, however, is a must-have when it comes to combating cross-site scripting. You should verify every single input being submitted by the user against a set of rules to make sure it is within the defined range. Validation includes using simple checks, regular expressions and encoding/escaping functions to check the format of the input and to replace characters and strings that might have malicious purposes. For instance, if the user has entered the word “<string>” in the input, it should be escaped properly in order to display it as a text instead of rendering it as an HTML tag. Escaping is especially useful if your site doesn’t need rich content. Controlling page charsets Defining the character set of your pages with meta tags can greatly reduce the number of possible forms script injection can be implemented. This is especially important since some browsers don’t specify default charsets for pages, and some servers don’t allow charset headers to be sent along with responses. So, for instance, if your site isn’t meant to display characters outside the ISO-8859-1 character set (which encompasses English and most European languages) you can use the following tag in your pages’ <header> in order to prevent charset-based XSS attacks: <META http-equiv="Content-Type" content="text/html; charset= ISO-8859-1"> Crossing boundaries policy A good XSS prevention practice, which is gaining popularity among websites with sensitive and financial information, is to require users to re-authenticate when desiring to access critical parts of the application. This way, even if the user’s browser has a stored cookie which will automatically log the user into the website, accessing a sensitive section such as the user account and financial information will require the user to reenter the credentials. This can prevent many forms of XSS attacks that steal or leverage cookies that are already stored on the server to hijack user sessions and make requests to the server with the user’s privileges. Limiting session IPs Another useful technique in fighting cross-site scripting is to prevent a user from being logged in to the same account from different IPs at the same time. This way, if a session cookie is stolen and sent to a remote location, it will not be usable by an XSS scheme. However, this measure can be overcome with IP spoofing, in which the attacker manipulates requests and packets to present requests as coming from the original user’s computer, but it does provide an extra layer of security that gives attackers a harder time in carrying out their XSS attacks. Using source code scanners Using an automatic source code scanning and auditing tool can help a lot in identifying security gaps in your code during development. These tools are becoming smarter and cap help spot, among others, XSS weaknesses and loopholes in your code. Also if you’re using third party libraries and packages in your website, make sure you make a quick search for known issues and vulnerabilities. And it never hurts to test the way they handle input by yourself to make sure they’re not prone to XSS. Disable scripts on the client
<urn:uuid:ecfb2299-ca5e-4ba9-b604-6c5260b3b7d2>
CC-MAIN-2022-40
https://bdtechtalks.com/2016/03/10/a-beginners-guide-to-cross-site-scripting-xss-attack-prevention/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337480.10/warc/CC-MAIN-20221004054641-20221004084641-00680.warc.gz
en
0.926875
1,842
3.1875
3
By Orla Forrest, Irish Telecom Net neutrality refers to the provision of a completely open Internet whereby any user can access and post online content that they please, in the knowledge that it will not be censored by Internet service providers (ISPs). For most web users, it is the preferred state of affairs, as their views will not be censored, while mega-rich corporations are prevented from pricing users out of an online presence. However, ISPs and a minority of users will tell you that online content should be prioritized in order of importance and that heavy users should be charged more so that bandwidth can be distributed more appropriately. The net neutrality debate has been a venomous one at times in the U.S., with the FCC becoming involved in running battles with other parties. While the Obama administration strongly advocated net neutrality, it appears likely that Trump’s troops, who oppose the concept, could undo much of the legislation that has been introduced in recent years. Were that to happen, a lot of people would be up in arms. It would be especially disastrous for small start-up businesses who are trying to gain a foothold in terms of having an online presence, as the abolition of net neutrality would deny them the fair competition that currently exists. Google is a prime example of a company that was able to grow exponentially because of an open Internet. Minority communities would also be greatly affected by the possible abolition of net neutrality, as their views could be censored by ISPs and they would likely be subjected to suppression. A survey by Know Your Target Market showed that most Internet users would demonstrably show their opposition if popular high-bandwidth websites such as Netflix and Pandora were blocked, slowed down or incurred a fee to use. Almost half of users would switch to a different ISP or complain to their current provider. More than 1 in 10 would go as far as cancelling their Internet service entirely, while 14% would complain to the relevant government authority. Only 1 in 5 admitted that they wouldn’t take any action. The net neutrality debate is one that will continue to polarize opinion, depending upon the best interests of the stakeholders involved. To learn more about the concept and determine which side of the divide you occupy, take a look at this infographic from Irish Telecom.
<urn:uuid:c2c4beaa-093f-4071-94d3-26668c656e5f>
CC-MAIN-2022-40
https://bdtechtalks.com/2017/04/08/why-should-you-care-about-the-net-neutrality-and-open-internet-debate/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337480.10/warc/CC-MAIN-20221004054641-20221004084641-00680.warc.gz
en
0.974338
458
2.828125
3
JDBC (Java DataBase Connectivity) is a Java data access technology from Oracle Corporation. JDBC provides an API for the Java programming language for database-independent connectivity, and it is based on the use of drivers for each database. A client application requires separate drivers, usually vendor supplied, to connect to different types of databases. Denodo includes a JDBC driver jar file named denodo-vdp-jdbcdriver.jar, and it is located under the In this section, you are going to see how to access to the Denodo server using a JDBC client. This information is valid for any Java-based application. For the example, we will use DBVisualizer (a generic database management tool for developers) but feel free to use any other JDBC client. The first thing that you have to do when connecting using JDBC is to add the Denodo's JDBC driver to the client application. To use the JDBC driver in your client, you have to add the .jar file to the classpath of your application. In DBVisualizer, you have to go to Tools > Driver Manager... and in the Driver Manager window go to Driver > Create Driver and then browse to the VDP's driver file. Use the following driver settings and close the window to save the configuration: Now that you have added the driver, you can configure a connection to your Denodo virtual database. Go to Create Database Connection and use the following settings for the connection: Click on the Connect button, and you will establish a connection to the tutorial database. In the left panel of the window, you will see that the base views are listed as tables and the derived views are listed as views. If you run any query from the SQL Commander, you will get the same results that you were seeing before from the Administration Tool. Try, for example:
<urn:uuid:25602bd1-8cbb-4404-b861-b284604dcf24>
CC-MAIN-2022-40
https://community.denodo.com/tutorials/browse/basics/4connect1jdbcclient
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337631.84/warc/CC-MAIN-20221005140739-20221005170739-00680.warc.gz
en
0.899765
413
2.71875
3
In this digital age, virtually all devices are running on some type of digital network, whether they are hard wired or wireless. Electronic security systems currently in place on many campuses are dated; they do not integrate or “talk” with today’s digital environment. The challenge of how to migrate these systems from analog to digital solutions utilizing an IT network can become overwhelming and costly for university, hospital and school district IT departments. Many of the current devices and systems are antiquated and the life cycle is near expiration. Replacing analog devices that are no longer being manufactured with new digital devices poses security planning and budgeting challenges. Migrating existing electronic security systems to a fully digital environment can deliver numerous advantages for any organization. New technologies offer many desirable features by allowing devices to “talk” to each other, unlocking capabilities only dreamed of a decade ago. As a new tablet has more computing power than a two-year-old desktop computer (at a fraction of the cost), electronic security systems have vast advantages over older, analog models. Unfortunately, campuses that currently have electronic security systems such as video surveillance, access control, intrusion detection, visitor management, emergency blue light photos and others often are utilizing disparate or independent ones that don’t communicate with each other. Security system migration planning can address this challenge, enabling the institution to more easily upgrade its technology. Conduct an Initial Security Assessment The first step is to observe and analyze the existing security structures, including an institution’s electronic security systems and components. It is important to address current and future security requirements and demands. Organizations that understand the need for initial security assessments must then decide if they will move to digital security enterprise systems. Once they make this determination, further challenges arise as vital decisions must be made in order to create a digital security system that will function efficiently and fluently while remaining future-proofed and adaptable to changing technologies. Additional challenges faced in these types of assessments are the existing campus IT network and determining if there is sufficient bandwidth to allow for the transmission of digital security data. Once the quantity of devices to be added to the network is identified, the data to be transmitted can be calculated, and the expansion or addition to the IT backbone/network can be defined. In many situations, the telecom rooms must be expanded to accommodate additional network equipment, and conduits and/or cable trays may need to be installed to manage the additional wiring. The evaluation of adequate electrical power and environmental control (temperature and humidity) for these locations is also necessary. Security command centers may face similar challenges to those experienced by the telecom rooms.
<urn:uuid:fbeebfc4-0e92-4d34-82ef-60d188b63efc>
CC-MAIN-2022-40
https://mytechdecisions.com/network-security/integrating-school-electronic-security-systems/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337631.84/warc/CC-MAIN-20221005140739-20221005170739-00680.warc.gz
en
0.932509
537
2.546875
3
All companies are exposed to some risk in their everyday operations. However, among these risks, financial risk is arguably the most concerning. The 2021 Financial Services Data Risk report showed that in about two-thirds of companies, more than 1000 sensitive files are accessible to all employees. Moreover, 60% of companies have above 500 passwords that don't expire. Such widespread accessibility of data and a lack of expiring passwords can pose security risks to the company. Apart from data-specific financial risks, companies face many other types of financial risks. Some include market risk, credit risk, liquidity risk, and operational risk. In this guide, you'll learn about different types of financial risks. Financial risk is a type of risk that can harm a company's financial performance. It can arise from different sources and lead to severe consequences. Financial risks can come from both internal and external sources. Internal sources include the decisions made by management, while external sources include the economic environment and market conditions. As mentioned earlier, there are different reasons that companies face financial risks. Some of these include: Changes in Government Policies: Changes in government policy can also lead to financial risks. For example, if the government imposes a new tax on a company's products, it can impact the organization's bottom line. Now that you know what financial risks are and some of the reasons they occur, let's look at some of the different types of financial risks. Keep in mind that some risks might be more impactful than others, but they're all detrimental to a company. Also known as systematic risk, this is the possibility that an investment will lose value due to factors that affect the market's overall performance. It involves the risk of changes in asset prices and is beyond the control of any single entity. Market risk can be further divided into two types: Market risk also includes the possibility of being taken out of business by competitors. Today's global market is more competitive than ever. Companies that keep up with the latest trends and make necessary changes to their value propositions tend to be more successful than those that don't make required adjustments to stand out in the market. Companies can take different measures to mitigate market risk. Some of them include: Operational risk is the possibility of losses due to problems with a company's internal processes. It can arise from different sources, such as human error, system failure, or natural disasters. The most effective way to mitigate operational risk is to have adequate internal controls. These controls help to ensure that processes are carried out correctly and that risks are identified and managed effectively. Another way to manage operational risk is to have insurance. It can protect a company from losses arising from accidents, natural disasters, or other unforeseen events. At Accountable, we offer all our clients who have completed the steps of compliance a $100K Accountable Compliance Protection Guarantee to cover any breach or noncompliance related costs. Liquidity risk refers to the possibility that a company will not be able to meet its financial obligations when they're due. The risk can arise when a company doesn't have enough cash on hand to meet its short-term obligations or cannot convert its assets into cash quickly enough to pay its bills. According to the WallStreet Mojo, some examples of liquidity risk include: For instance, startups are often at a break-even risk. If they don't get subsequent funding, they may experience a liquidity risk. Organizations must always have sufficient cash to meet their short-term obligations. They can do this by maintaining a solid cash position and keeping a line of credit open. Another way to manage liquidity risk is to invest in assets that businesses can easily convert into cash, such as short-term investments or marketable securities. A company can also face financial risks if its customers or clients don't make timely payments. This is called credit risk. When a customer doesn't pay on time, it can harm the company's cash flow and bottom line. To mitigate this type of risk, companies often do a credit check before extending credit to customers. A McKinsey report showed that the Covid-19 pandemic is bound to induce credit risk for almost all industries, including consumer goods, construction, automotive, logistics, real estate, telecommunications, etc. However, the crisis-induced shock to loss and profit will differ by industry and recovery paths. Credit risk mitigations mean using certain practices to minimize the risk of losses due to nonpayment. Some examples of credit risk mitigations are: As a business owner, you should take proactive steps to identify and mitigate financial risks. The most common financial risks are market risk, operational risk, liquidity risk, and credit risk. Although the impact can vary from one risk type to another, all financial risks can damage a company’s financial standing and inherent value. There are a variety of ways to mitigate financial risks. Some examples include limiting liabilities, diversifying investments, maintaining a solid cash flow, and performing extensive credit analyses of customers and clients.
<urn:uuid:f8024674-f41a-43c3-a01a-e9b7c1c08366>
CC-MAIN-2022-40
https://www.accountablehq.com/page/types-of-financial-risk
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337631.84/warc/CC-MAIN-20221005140739-20221005170739-00680.warc.gz
en
0.959985
1,020
2.875
3
McKinsey has predicted that 800 million workers could be possibly displaced in 42 countries — a third of the workforce, in the Fourth Industrial Revolution. While similar predictions came before every earlier change of era, could this one be? The world has seen four industrial revolutions till now, and as each one has come with terrible frightening job loss scenarios. But, to date, nothing has been as disruptive as what’s coming now. Disruptive technologies are indeed transforming the entire social, economic, and political systems in many unpredictable ways. While technology evolution is difficult to map because of its factorial or exponential, or higher growth rate, impact assessments are challenging but not impossible. Many repetitive, low-skilled jobs are being automated, starting in the high wage countries but quickly spreading to the developing world as well. With two-thirds of the world’s robots present in Asia, some expect this region to be susceptible to these changes. From an economic viewpoint, the cost of adding extra supervisors to low skill levels doesn’t outweigh the savings from automation; the reduction in the price of the final product would spur an increase in demand. If the rise in demand is significant enough, it may expand the number of jobs in factories that automate few processes, but not all, of their production. There will also be inter-industry effects. Productivity increase from new technology in one industry. It can also decrease production costs in others through output-input linkages, contributing to increased employment and demand across industries. Greater demand and more production in one sector raise demand for other sectors, and on it goes. The widespread pessimism about the next industrial disruption It is more pragmatic to see how existing jobs may be lost to automation than it is to imagine how new ones emerge will fare. It is more sensational to highlight the job displacing possibilities rather than the job-creating ones. The effects will vary not just between developing and developed countries, but also within each group, which then shapes our overall negative perception. When there is enough uncertainty, it is safer to overstate compared to understate the potential cost to innocent victims of change. But there could be a silver lining. It is not just the extent of reskilling that will vary, but also the type, based on a host of factors which broadly correlate with various levels of income. If negativity leads to efforts to reshape and reskill the workforce to better adapt to change, that is precisely what is required. Ironically, this pessimism could well generate for the next job apocalypse
<urn:uuid:03669d39-d258-4ad3-ac1c-746156c79258>
CC-MAIN-2022-40
https://enterprisetalk.com/featured/industries-fight-automation-hypochondria-in-the-tech-industrial-age/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337889.44/warc/CC-MAIN-20221006222634-20221007012634-00680.warc.gz
en
0.955088
518
2.640625
3
Your IT systems are no longer safe unless you have strong data security measures in place. Data security has become more than a legal obligation to follow a bunch of compliances. When the security of your customer’s sensitive information is threatened, it’s time to take a step back and analyze what you can do to improve safety. The data threats can be of any form: data loss through cyberattacks, phishing collection of personal information, virus attack, or insider theft. You can strengthen data security by enforcing strict access controls, protocols, strong firewalls, and antivirus and antimalware protection systems. This blog will take you through various things you need to know to improve your data security. The Importance of Data Security Before you look into the details of what data security means and how you can enhance it, let’s start by analyzing the main concerns: Data security is defined as the laws and regulations that ensure the safety of people’s personal information (e.g., credit cards) and information systems in various business, medical, and technical fields. It’s time to think of data security as a responsibility and not a means to comply with regulations. You have data from users, end-users, and business partners, all of which contain private and sensitive data. Before choosing any third-party service provider, let’s break down each type of information and what you need to know. End-User Data: They include information such as identity and authentication details, geo-locations, credit card information, bank details, phone numbers, etc. These may be maintained internally but will often leak to another party. Application Data: This is the code that runs on your computer. For instance, when a website requests a particular piece of information on your computer, this is what you are typing in. When such data is hacked, it can threaten the very foundation of your business and, at the least, lose your customers’ trust. This is why you need to give the most to such data. Reasons for Weak Data Security The main reason for weak data security is human error. Your employees may accidentally click on an insecure link or email attachment that contains malware. Other possible causes include lack of training, lack of control on employees, weakly trained staff, and employees who have a poor understanding of network administration. If this sounds like your environment, start doing these three things: - Add passwords to every new login. This way, when a user gets a password for a new account, they won’t reuse the same password. - Install virus protection and antivirus software on all computers and mobile devices. - Don’t email sensitive or confidential information to any external sources. That includes emails, text messages, and social media platforms. How to Improve the Data Security An effective data security system takes an in-depth look at data and its security needs while changing data processing systems or building new ones. Here are some of the ways to improve your data security. 1. Create Strong Passwords With the increasing usage of mobile apps, password leaks have been a frequent occurrence. It’s unfortunate, but passwords are one of the weakest links to protect our data, exposed more than 30,000 times per day, according to SplashData. We are all guilty of not changing our passwords often enough. This is because most people do not make a conscious effort to change their passwords often enough, or even at all. The damage can be difficult to recognize and stop. However, there are steps we can take to improve the situation. The first step is not to use a shared password. 2. Set up Firewalls Your firewall’s purpose is to stop unauthorized devices from accessing network resources and the internet. Make sure you have enabled port forwarding for all your access points to ensure the best results. This allows port forwarding for critical destinations to work. You can specify the port ranges in which access to these destinations is allowed. 3. Enable Access Controls Using Single Sign-On (SSO) solutions, SSO is a secure system in which one login system can provide you with multiple access points such as web, mail, data, and so on. It is an easier way to protect data and grant access to authorized users. You can create multiple accounts for each of your employees. You can also configure password management systems for each of these accounts. 4. Create Encryption Encrypting your data helps ensure that nobody can access and misuse the data stored on your systems, even if they get their hands on it. You can use standard encryption to add“secret” information into a file, creating a code that a person cannot read without the appropriate decryption key. And you can also simply enable 2FA for your login attempts. Data security is no longer about whether or not you’re complying with rules. Instead, it’s if you’re implementing a security approach that a knowledgeable insider or hacker can’t circumvent. Being hacked often happens because a person or company is unaware or uneducated about some rudimentary security protocols. Educate your employees to adopt the right security practices to keep your organization safe.
<urn:uuid:91c35e94-78d8-4375-8a6f-31337ff03d67>
CC-MAIN-2022-40
https://l1n.com/ultimate-guide-on-data-security/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334871.54/warc/CC-MAIN-20220926113251-20220926143251-00080.warc.gz
en
0.913858
1,087
2.59375
3
What is an Ansible Playbook? Ansible operates by connecting to what you wish to automate and pushing programmes that perform manual instructions. These applications make use of Ansible modules that are developed depending on the endpoint’s connection, interface, and instructions. Ansible then runs these modules (through regular SSH by default) and removes them after they’re done (if applicable). Ansible is the most straightforward method for automating everyday IT activities. It’s intended to be simple, consistent, safe, and dependable, with an incredibly short learning curve for administrators, developers, and IT managers. Ansible playbooks are an essential feature of Ansible and the foundation of every Ansible setup. If you’re not familiar with Ansible, have a look at our Ansible tutorial to learn more about the power and potential of this popular DevOps tool. Table of Contents: - What is an Ansible Playbook? - How do Ansible Playbooks Work? - How to Build an Ansible Playbook? - How to use Ansible Playbooks? - What are the Variables Of Ansible Playbooks? - Example of Ansible Playbooks Check out this video for more information about Ansible Playbook What is an Ansible Playbook? An Ansible playbook is a file in which users create Ansible code, which is a structured collection of scripts that define the work of a server setup. They define a series of stages in a general IT process or a policy that your remote systems must follow. Playbooks are collections of one or more plays that are performed in a certain order. A play is an ordered sequence of tasks performed against hosts from your inventory. The task to be done is defined by plays. Each play has a list of hosts to configure and responsibilities to complete. There are no standardized plays; each play must be written by an administrator. YAML, a human-readable data processing language, is used in playbooks. “YAML ” is a sequence acronym that stands for “YAML Ain’t Markup Language.” How do Ansible Playbooks Work? Ansible modules carry out tasks. A play may be created by combining one or more Ansible jobs. An Ansible Playbook is made up of two or more plays. Ansible Playbooks are collections of tasks that run automatically against hosts. Your Ansible inventory is made up of groups of hosts. Each module in an Ansible Playbook is responsible for a certain task. Each module has metadata that indicates when and where a job is performed, as well as which user does it. There are lots of Ansible modules that execute various IT tasks, including: - Cloud management: oci_vcn creates, deletes, or updates virtual cloud networks in Oracle Cloud Infrastructure environments. - User management: selogin maps Linux operating system (OS) users to SELinux user and gitlab_user creates, updates, or deletes GitLab users. - Networking: Dozens of modules handle application programming interfaces (APIs); Cisco IOS, NXOS, and IOS XR devices; as well as F5 BIG-IP services. - Configuration management: pip manages Python library dependencies while assemble consolidates configuration files from fragments. - Security: Openssh_cert generates an OpenSSH host or user certificates, and ipa_config manages global FreeIPA configuration settings. How to Build an Ansible Playbook? Ansible playbooks have the ability to get pretty massive and sophisticated. But, whether you’re writing a short and sweet playbook or a massive epic, here’s how you do it. Every playbook is divided into the following standard sections: - Host: The host section specifies the computers on which the playbook will be executed. This data is derived from the Ansible inventory file. As a result, the host section is a list of devices. - variable: The variable section is optional and provides any variables required by the script. If at all, it can be as huge or as little as needed. - Tasks: The task section describes the usage of Modules and identifies the tasks that the target machine must do. Every job is given a name, as well as a brief description of what it accomplishes, and is listed when the playbook is executed. How to use Ansible Playbooks? Ansible makes use of the YAML syntax. YAML stands for yet another markup language or YAML ain’t markup language, depending on who you ask (a recursive acronym). YAML file extensions are also varied, but totally acceptable:.yaml or.yml. Ansible Playbooks can be used in two ways: through the command line interface (CLI) or through the Red Hat Ansible Automation Platform’s push-button deployments. - From the CLI: To launch Ansible Playbooks, just use the ansible-playbook command after installing the open source Ansible project or the Red Hat Ansible Automation Platform. - From within the platform:The web-based user interface of the Red Hat Ansible Automation Platform offers push-button Ansible Playbook deployments that are utilised as part of bigger tasks (or job templates). These deployments provide extra precautions that are especially beneficial to users who are new to IT automation or who may not have as much expertise working with the CLI. If you want to Ace the interview, here’s a golden chance for you Intellipaat Ansible Playbook Interview questions! What are the Variables Of Ansible Playbooks? Ansible utilizes variables to assist users in dealing with system variances, as no two systems are exactly the same. Variable names are made up of letters, digits, and underscores, but they must always start with a letter. Variables also never include blank spaces. Variables may be defined directly in playbooks by using the “vars:” command. Variables can be anything from proper nouns to ports to web servers, or even a specific command. For example, you may create a playbook in which you greet individuals with “How’s it going, eh?” by defining a variable called “greeting” with the value “How’s it doing, eh?” When the user runs the playbook, the message is shown on the terminals. Variables are used to store values, and you may build practically any type of variable you require. There are variables for groups and hosts, inventory files, array variables, dictionary variables, and special variables. Special variables are built-in variables that cannot be changed by the user and are always overridden by Ansible. You may alternatively maintain variables in a separate file and import them as needed with the vars files command. It is important to note that your playbook does not have to include variables if they are not required. They are completely optional. Example of Ansible Playbooks: Ansible can communicate with a wide range of device types, including cloud-based REST APIs, Linux and Windows systems, networking devices, and much more. This is an example of two Ansible modules that automatically update two different types of servers: - name: Ansible hosts: servers become: yes become_user: root tasks: - name: ensure an apache is at the latest version yum: name: httppd state: updated version - name: ensure apache is running service: name: httpd state: started This short ansible-playbook example should be enough to have your Apache installation completed and ready. I sense your frustration since I just provided a text with no explanation of what they do. I’ve already described what each line does. name The playbook’s name hosts A collection of hosts that are often grouped together as a host group and defined in an inventory file. To tell Ansible, this piece must be performed with higher privileges. become_user is the user name we wish to switch to, similar to sudo su – user. tasks a collection of activities to be completed, all tasks would be stated below this Then we have two jobs with two modules, the first of which is yum and the second of which is service. The latest variable in the first yum job signifies that the above mentioned package httpd should be installed if it is not already installed (or) updated to the latest version available if it is already installed. If you do not want it to be updated if it is present, modify the state:start. Courses you may like Ansible believes that you should not require another framework to validate the fundamentals of your system. This is because Ansible is an order-based system that will fail instantly if a host encounters an unhandled error, preventing further setup of that host. This pushes mistakes to the top of the list and displays them in a summary at the conclusion of the Ansible run. However, considering Ansible is a multi-tier orchestration system, it is quite simple to integrate tests at the conclusion of a playbook run, either via loose tasks or roles. When combined with rolling updates, testing processes can determine whether or not a machine should be returned to a load balanced pool. You can visit the Intellipaat DevOps Community page to get your doubts resolved!
<urn:uuid:d2c42fe2-b56f-471e-997d-668fbf521c1d>
CC-MAIN-2022-40
https://www.businessprocessincubator.com/content/what-is-an-ansible-playbook/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334871.54/warc/CC-MAIN-20220926113251-20220926143251-00080.warc.gz
en
0.902187
1,986
3.078125
3
What are ITU Regions? The International Telecommunication Union (ITU), in its International Radio Regulations, divides the world into three ITU regions for the purposes of managing the global radio spectrum. Each region has its own set of frequency allocations, the main reason for defining the regions. - Region 1 comprises Europe, Africa, the former Soviet Union, Mongolia, and the Middle East west of the Persian Gulf, including Iraq. - The western boundary is defined by Line B. - Region 2 covers the Americas including Greenland, and some of the eastern Pacific Islands. - The eastern boundary is defined by Line B. - Region 3 contains most of non-FSU Asia east of and including Iran, and most of Oceania. - Line B is a line running from the North Pole along meridian 10° West of Greenwich to its intersection with parallel 72° North; thence by great circle arc to the intersection of meridian 50° West and parallel 40° North; thence by great circle arc to the intersection of meridian 20° West and parallel 10° South; thence along meridian 20° West to the South Pole. Another chart showing the regions: The definition of the European Broadcasting Area uses some of the definitions of Region 1. About the ITU The International Telecommunication Union (ITU; French: Union Internationale des Télécommunications (UIT)), originally the International Telegraph Union (French: Union Télégraphique Internationale), is a specialized agency of the United Nations (UN) that is responsible for issues that concern information and communication technologies. The ITU coordinates the shared global use of the radio spectrum, promotes international cooperation in assigning satellite orbits, works to improve telecommunication infrastructure in the developing world, and assists in the development and coordination of worldwide technical standards. The International Telecommunication Union is active in areas including broadband Internet, latest-generation wireless technologies, aeronautical and maritime navigation, radio astronomy, satellite-based meteorology, convergence in fixed-mobile phone, Internet access, data, voice, TV broadcasting, and next-generation networks. The agency also organizes worldwide and regional exhibitions and forums, such as ITU Telecom World, bringing together representatives of government and the telecommunications and ICT industry to exchange ideas, knowledge and technology. For Further Information For More Information on Microwave Links, Please Contact Us
<urn:uuid:a38f179e-feda-4b5c-95d8-3ddf1da471b0>
CC-MAIN-2022-40
https://www.microwave-link.com/microwave/itu-regions/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335058.80/warc/CC-MAIN-20220927194248-20220927224248-00080.warc.gz
en
0.889012
494
3.40625
3
In the earlier article “Secrity, Democracy And Digital Identity” , we made clear that we should look for ‘something other than the text password’ in the domain of ‘Secret Credentials’ and referred to the proposition of Expanded Password System. We would like to talk more about what Expanded Password System specifically offers in this article. What our Remembrance Brings for Digital Identity In the matrix below, there are several known images. We can easily find all of them right away. Or, rather, these known images jump into our eye. And, only we are able to select all of them correctly. This is Expanded Password System. We can use both images and characters. It’s easy to manage the relation between accounts and the corresponding passwords – Comfortable and even fun! The idea of using pictures for passwords is not new. It’s been around for more than two decades but the simple forms of pictorial passwords were not as useful as had been expected. Unknown pictures we manage to remember afresh are still easy to forget or get confused, if not as badly as random alphanumeric characters. Expanded Password System is new in that it offers a choice to make use of known images that are associated with our autobiographic/episodic memories. Since these images are the least subject to the memory interference, it enables us to manage dozens of unique strong passwords without reusing the same password across many accounts or carrying around a memo with passwords on it. And, handling memorable images makes us feel comfortable, relaxed and even healed. Torturous login is history. Well, let’s talk about some major problems that use of our own episodic memory enables us to solve. Relation of Accounts with Passwords Being able to recall strong passwords is one thing. Being able to recall the relation between accounts and the corresponding passwords is another. When unique matrices of images are allocated to different accounts, those unique image matrices will be telling you what images you should pick up as your password for this or that account. When using images of our episodic memories, Expanded Password System will thus free us from the burden of managing the relation between accounts and the corresponding passwords. So far, only texts have been accepted. It was, as it were, we have no choice but to walk up a long steep staircase. With Expanded Password System, we could imagine a situation that escalators and elevators are provided along with the staircase. Or, some of us could think of all those ladders we have for climbing in Donkey Kong. Where we want to continue to use text passwords, we could opt to recall the remembered passwords, although the memory ceiling is very low, Most of us can manage only up to several of them. We could opt to recognize the pictures remembered in stories where we want to reduce a burden of textual passwords. The memory ceiling is high, that is, we would be able to manage more and more of them. Where we choose to make use of episodic image memory, we would only need to recognize the unforgettable images, say, known images. There is virtually no memory ceiling, that is, we would be able to manage as many passwords as we like, without any extra efforts. Hard-to-break text passwords are hard-to-remember. But it’s not the fate of all the secret credential. It would be easily possible to safely manage many of high-entropy passwords with Expanded Password System that handles characters as images. Each image or character is presented by the image identifier data which can be of any length. Assume that your password is “CBA123” and that the image ‘C’ is identified as X4s& eI0w, and so on. When you input CBA123, the authentication data that the server receives is not the easy-to-break“CBA123”, but something like X4s&eI0wdoex7RVb%9Ub3mJvk”, which could be automatically altered periodically or at each access where desired Brain-Computer-Interface and Its Security A simple brain-monitoring has a problem in security. The authentication data, if wiretapped by criminals, can be replayed for impersonation straight away. Therefore the data should desirably be randomized as the onetime disposable ones. An idea is that the authentication system allocates random numbers or characters to the images shown to the users. The users focus their attention on the numbers or characters given to the images they had registered. The monitoring system will collect the brain-generated onetime signals corresponding to the registered images. Incidentally, the channel for showing the pictures is supposed to be separate from the channel for brain-monitoring. If intercepting successfully, criminals would be unable to impersonate the users because the intercepted data has been disposed of. No-Cost 2-Factor Authentication A very strong password supposed to not be remembered and written down on a memo should be viewed as ‘what we have’, definitely not ‘what we remember’, so it could be used as one of the two factors along with a remembered password. We could then turn a boring legacy password system into a two factor authentication system at no cost, just by verifying two passwords at a time, one volitionally recalled and the other one physically possessed. When those two different passwords are used as two factors, we could rely on the strength of a remembered password against physical theft and the strength of a physically possessed long password against brute force attack, although it is not as strong against wiretapping as token-based solutions armed with PKI or Onetime Password. This configuration could be viewed just as a thought experiment or could actually be considered for practical application in between a single factor authentication and a costly heavily-armored 2-factor scheme, or, as a transition from the former to the latter. It goes without saying that Expanded Password System could be brought in for generating a remembered high-entropy password. President, Mnemonic Security, Inc. Profile: Advocate of ‘Identity Assurance by Our Own Volition and Memory’, Hitoshi Kokumai is the inventor of Expanded Password System that enables people to make use of episodic image memories for intuitive and secure identity authentication. He has kept raising the issue of wrong usage of biometrics and the false sense of security it brings for 17 years. Mnemonic Security Inc. was founded in 2001 by Hitoshi Kokumai for promoting Expanded Password System. Following the pilotscale operations in Japan, it is seeking to set up the global headquarters. Appendix – Fighting Threats to Security and Democracy from Within Where the digital identity platform was built without the secret credentials made from our memory, we would have to see the necessary level of security lost. Where the secret credentials, for which our will/volition is indispensable, are removed from the digital identity platform, we would have to see erosion of democracy that our ancestors have won through heavy sacrifices. On this front we are not optimistic; too few people are taking the correct course towards the correct objectives. Too many people, with professionals, researchers, politicians and journalists included, are badly distracted and straying off the course. More and more people are expected to join our efforts.
<urn:uuid:e754e234-b0c6-4174-a9c1-50b01ac563e7>
CC-MAIN-2022-40
https://informationsecuritybuzz.com/articles/digital-identity-and-episodic-memory/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335304.71/warc/CC-MAIN-20220929034214-20220929064214-00080.warc.gz
en
0.951686
1,539
2.671875
3
IT asset management is not yet another buzzword but a real set of practices that can help your organization maximize the use of IT resources and avoid potential risks. Sounds interesting? If you have wondered what is ITAM, this post will introduce you to IT asset management definition and help you navigate your way towards a tangible IT asset management strategy. What is Asset Management In Its Essence? You might have heard the term asset management from financial advisors and market analysts who mostly use in relation to stocks and real estate. But what is an asset anyway? The asset is literally anything that has some value and can be traded, exchanged, or used for business activities. For example, stocks, a house, a laptop, copyright are assets because you can sell them or utilize them to get money. As an owner, you take care of the resources available to you and ensure they are profitable. For example, you’d want to repair a roof of your house, renew car insurance, or sell the car if it’s no longer a good investment. This is what asset management is all about — around the clock care about the resources that are available to you. On the company level, asset management deals with real estate, machinery, equipment, intellectual property, devices, technologies, etc. To keep these assets profitable, the company should be diligent about building maintenance strategies. The Definition of IT Asset Management But what is the definition of ITAM? Are IT assets and approaches to managing them somehow different? Actually, they are a part of the bigger picture. IT asset management (ITAM) aligns with corporate asset management but adds a special flavor to it. Information Technology Asset Management is a set of business tactics and strategies aligned to sustain life cycle management and decisions made for the IT environment. Some IT assets such as computers and servers, hardware components, peripheral devices, disks are physical just like traditional assets. Software, licenses, files, databases, Cloud-based storage, user profiles and settings are also considered to be assets even though they are not feasible. These assets are intellectual property and have great business value in the digital economy. The asset management techniques that are specifically focused on IT assets and keeping them secured and cost-effective are known as IT asset management. Why Do You Need ITAM? If you are still unsure whether you need IT asset management, then look around. Does the company you work for rely on communication technologies, data storage, planning or finance apps? It’s a safe bet that the vast majority of company business operations is held over the network and on computers. The amount of data can be overwhelming and can easily spiral out control if the IT assets aren’t supervised. Unlike traditional assets, IT assets have a higher level of mobility. Laptops are passed from one employee to another, financial files are shared between departments, and licenses are floating. At the same time, most IT assets require regular maintenance and servicing to keep them running. With no management, if something goes wrong in this melting pot, it will be hard to remediate the problem and trace its cause. IT asset management safeguards business operations and streamlines the work of the IT department. On top of that, it facilitates overall corporate governance ad service management. Types of ITAM ITAM is not a newcomer and has been here for quite some time. It has a professional organization (https://www.itamstandards.org) and standards, such as ISO/IEC 38500 and ISO/IEC 19770. ISOs provide guidelines on managing, labeling, auditing of specific groups of IT assets: - Management of hardware and devices, including mobile devices. Servers, cellphones, printers should be accountable and scheduled for regular service checkups. - Management of software and licenses. For this category, it’s important to keep track of installed software and its license status to validate compliance with industry regulations as well as roll out updates and security patches. - Information management. This category is translucent but needs your special attention. Business documents, corporate email addresses, user profiles are subject to privacy regulations. This data tends to be a valuable resource and should be secured against misuse and access violation. - Cloud management is a recent yet important add-on as more and more businesses rely on cloud computing. IT Asset Management Processes IT asset management processes center around the asset lifecycle and work towards optimizing each stage while mitigating risks and potential loss. Here are the asset lifecycle stages and IT asset management practices associated with them. - Planning is a preliminary stage when the asset doesn’t yet exist. Based on the stakeholders’ interests and corporate policies, IT manager plans and budgets for adding new assets to the IT infrastructure. For example, soliciting requirements for a new server, finding the right configuration of hardware, setting up a budget and timeframe, finding the right supplier, and finally acquiring the asset. - Deploying in the IT environment is the stage when the asset is introduced on a companywide level. IT manager has to set it right, ensure all licenses are in place, and configure access permissions. At this early stage, asset management is especially important. New assets should be accountable and easily traceable. For example, at this point new corporate cellphones should be labeled, added to the device database, assigned to users, their configuration should be described and approved. With no governance and supervision, devices soon will be scattered among employees, have malicious software installed, or violate internal security standards. - Maintenance and support is the longest stage of the asset lifecycle. During its lifespan, the asset requires multiple examinations, updates, repairing. For software assets, you’d want to renew licenses or transfer them to other users, roll out version updates and security patches. The goal of the management efforts on this stage is to keep the asset profitable as long as possible. As IT manager, you’ll also want to reevaluate assets individually and all together and consider a replacement if the asset no longer fits the company requirements. IT asset management practices will always convey higher-level governance and compliance principles employed in the company and prioritize certain decisions over others. For example, when it comes to reducing costs, your company may prefer buying a new device instead of fixing existing an old one and vice versa. - The disposal stage is the last stage of the asset lifecycle within a given company. Keep IT asset hardware inventory to track outdated devices that are no longer used can be recycled, put to the storage, or sold. Deprecated technologies or compromised software products are taken out of use, files are archived. Even though the retirement stage is the last one, it is critical for IT asset management to keep it accountable. For example, when a laptop is put to the storage, it should be clearly labeled as retired. IT Asset Inventory Tools Keeping a neat inventory is a bread-and-butter of IT asset management. How can one possibly manage IT asset inventory without information on what assets are available, where are they located, who owns them, etc. While you can create an inventory manually in any form of a spreadsheet, there are inventory tools that significantly speed up the process. Basically, there are two approaches to building inventory and inventory tools that help you with it. Descriptive tools such as many service desk applications provide you with a straightforward interface for cataloging your managed assets. It means you’ll have to enter them yourself, one by one. On the contrary, the discovery tools can automatically locate your assets, acquire their configurations, and provide you with detailed reports. Neither of them is sufficient without another as alone they can provide misleading results. With the descriptive tools, you get an ideal snapshot of your IT system: the assets you have, configurations, assigned users, versions. Unfortunately, it is extremely difficult to keep this list up-to-date as IT assets keep circling. This is where the discovery tools come to help. They show the real situation if some assets were changed or redistributed. If something falls through the cracks, you’ll find it with discovery tools. But discovery tools provide no clue on missing assets as they are simply not in the scope. Combining both approaches and respective tools helps you gain insight into what assets you have and where they actually are and identify missing or uncatalogued assets faster. IT Asset Management Best Practices Obviously, the best possible outcome is when the asset management plan you have on the paper and the way your assets are distributed are nearly identical. Sticking to these basic best practices will help you archive that: - Inventory-first approach. First things first, add an item to the asset inventory as soon as it’s acquired. - Make assets accountable. Stick labels and barcodes to your hardware assets or assign tags in the software inventory tool. - Mark & map. Clearly mark the status of your IT assets and map assign them to users or locations. You should have a feasible way to check whom the laptop belongs to, how many users run the software, or who has custody of the shared folder. The chances to roll out security updates on a laptop that goes from one employee to another are diminishing if you don’t have a clear system to assign assets to people. Alternatively, you can map less mobile devices such as printers, IP phones, etc. to locations, not people. - See the bigger picture. Consider creating clusters of assets such as hardware plus software and manage them as one. You can also identify relationships between IT assets and traditional assets and take them into account as well. IT asset management is tightly integrated with IT service management and you can make great use of business process management routines. Summary — the Benefits You’ll Get by Employing ITAM Practices IT asset management can be an independent activity or it can be a part of the companywide governance framework. Either way, it’s beneficial for IT departments to utilize IT asset inventory tools since it ensures that information resources and related technology are used in the most transparent, risk-free, and cost-efficient way. When performed right, ITAM brings visibility in IT assets usage patterns and thus helps mitigate risks and plan upgrades in a timely manner.
<urn:uuid:80d45923-922f-4dae-8fd7-4e13db28b49d>
CC-MAIN-2022-40
https://www.action1.com/what-is-it-asset-management-defining-itam/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335469.40/warc/CC-MAIN-20220930113830-20220930143830-00080.warc.gz
en
0.934525
2,078
2.515625
3
System X History. System X was a leading edge UK digital narrow band Switch (Telephone Exchange) that was incepted and developed during the 1980s. System X was designed by a consortium of companies that included: General Post Office (Now British Telecommunications, or BT), Plessey, GEC and STC. It was a joint engineering collaboration with the main telecommunication incumbent (GPO/BT) and leading engineering companies that had specialised competence with Digital Switching. Around 1984, BT was privatised. The Switching Fabric was designed by Plessey, GEC designed the Processor, STC designed the Concentrator, the GPO (now BT) took control of the overall Management System. Overall, the whole System-X development Project was backed by the GPO/BT. Soon the consortium merged and one company was formed – GPT – GEC Plessey Telecommunication. STC soon dropped their involvement with System X. VBR/ Wallis Dudhnath
<urn:uuid:ac34128c-abcf-400c-ac1c-3b5510beb08a>
CC-MAIN-2022-40
https://www.erlang.com/reply/69977/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334515.14/warc/CC-MAIN-20220925070216-20220925100216-00280.warc.gz
en
0.981056
202
3.046875
3
Thousands of black holes A team of astrophysicists has recently reveals a study on black holes in our milky way. The study explains, thousands of black holes gathered around Sagittarius A*, the supermassive black hole in the center of our Milky Way. More than two decades ago, researchers searched unsuccessfully for evidence to support a theory that thousands of black holes surround supermassive black holes (SMBHs) at the center of large galaxies. Columbia Astrophysicist Chuck Hailey, lead author on the study, says, there are only about five dozen known black holes in the entire galaxy, including that extensive unsuccessful search made for black holes around Sgr A*. Black holes around Sgr A* According to researchers, the Sgr A* surroundings covered with a halo of gas and dust that set up a stage for the birth of massive stars, which live, die and could turn into black holes there. While most of the trapped black holes remain isolated, some capture and bind to a passing star. These isolated and mated black holes in the Galactic Center forming a density cusp which gets busier as distance to the SMBH decreases. It’s an obvious way to search for black holes, but the Galactic Center is so far away from Earth those radio bursts are solid and bright to see about once every 100 to 1,000 years. Binary black holes To detect binary black holes, researchers need to look for the unclear, but steadier X-rays emitted after the initial bonding, when the binaries are in an inactive state. Hailey and colleagues turned to archival data from the Chandra X-ray Observatory to test their technique. They searched for X-ray signatures of black hole low mass binaries in their inactive state and able to find 12 objects within three light years of Sgr A*. At last, the scientists analyzed the properties and spatial conveyance of the identified binary systems and estimated 300 to 500 black hole low mass binaries and about 10,000 isolated black holes in the area surrounding Sgr A*. This finding affirms a noteworthy hypothesis and the suggestions to related studies.
<urn:uuid:407f313b-8ed8-414b-a788-7ec76d1322a1>
CC-MAIN-2022-40
https://areflect.com/2018/04/10/thousands-of-black-holes-gathered-at-the-center-of-our-milky-way/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334912.28/warc/CC-MAIN-20220926144455-20220926174455-00280.warc.gz
en
0.924408
437
4.09375
4
Mountains, Volcanoes & Deserts - Quiz Questions with Answers See how much you know about mountains, volcanoes, and deserts with this quiz. - Which African peak is known as “The Mountain of the Cold Devils”? - Mont Blanc lies within the territories of which three countries? - Which range of mountains in Africa takes its name from a mythological character condemned to support the universe on his shoulders? - Two continents are separated by one range of mountains; Name the continents and the mountains. - Where are these? (a) Atacama Desert (b) Thar Desert - Where are the following volcanoes: (c) Mt. Meru - What is the name of the two largest Passes through the Austrian Alps? - (a) Name the great desert between China and Mongolia; (b) Name the two largest deserts in Africa, one being in the north and the other in the south. - Which mountains form the backbone of the Italian Peninsula? - What is the name of the famous active volcano on the island of Sicily? - Which mountains are the backbone of: (a) North America (b) South America - Where are the following mountains? (a) The Pennines - In which range of mountains is Everest? - On which of the Italian Lipari Islands is there a famous volcano? - What well-known mountain range in turkey is also the name of a sign of the Zodiac? - Where is (a) Mt. Egmont (b) Mt. Erebus - In which mountains do the following rivers are? (a) The Amazon (b) The Porise - All continents have deserts except one, which continent? - Where the following mountain ranges are? (a) K-2 Mt. Godwin Austen - Which is the longest mountain range in the world? - Mount Kilimanjaro - France, Switzerland, and Italy - The Atlas Mountains - Europe and Asia; the Ural Mountains - (a) Chile - (a) Ecuador - Brenner and Semmering - (a) The Gobi Desert (b) (i) Sahara (North), (ii) Kalahari (South) - The Apennines - Etna (the highest volcano in Europe) - (a) The Rockies (b) The Andes - (a) England - The Himalayas - (a) North Island, New Zealand (b) Ross Island, Antarctica - (a) Andes (b) French Alps - (a) Karakoram - The Andes in South America is the longest mountain range – 15,000 miles long.
<urn:uuid:a4d84285-3541-4fae-ab76-3a031c36eae6>
CC-MAIN-2022-40
https://www.knowledgepublisher.com/article/717/mountains-volcanoes-deserts-quiz-questions-with-answers.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334912.28/warc/CC-MAIN-20220926144455-20220926174455-00280.warc.gz
en
0.905187
660
2.890625
3
Article written by Ben Herzberg, Chief Scientist of Satori We live in a data-driven world. From the first moment we learn to speak, we are inundated with data: numbers, measurements, and labels. When we are in need of assistance, numbers guide us to the proper resources. When we want to purchase something new, numbers tell us what we need. In every step of our journey, data is there to help us make decisions. Today’s corporations rely on data to make critical business decisions—decisions that affect their bottom line and can influence whether they will be successful tomorrow. Data is the lifeblood of a company, and to make informed decisions, business leaders need to be data fluent. So, what is data fluency? And how can you empower your business to become data fluent? This article will explore the definition of data fluency and its key components, along with some tips on how to improve your company’s data fluency. What is Data Fluency? Data fluency is the ability to understand, interpret, and act on data. It’s being able to ask the right questions of data and use that information to make sound decisions. Data fluency is also about communication—the ability to take data and turn it into a story that everyone can understand. Data fluency in a business setting links users from various teams via a common set of standards, procedures, technology, and lingo. Data fluent users may turn raw data into usable insights since they understand how to read it, know the data that is and isn’t accessible, as well as how to make effective use of it. The idea that only a few people are gatekeepers of data is rejected. Instead, information is spread, data access across an organization is increased, and decision-making for everyone improves as a result of data fluency. Key Components of Data Fluency There are three key components of data fluency: - Data literacy: the ability to read and understand data - Data interpretation: the ability to make sense of data and see patterns in it - Data actionability: the ability to use data to make decisions All three of these components are essential for making informed decisions. So, let’s take a closer look at each one. Data literacy is the ability to read and understand data. To be literate in data, you need to be able to extract information from data sets and understand what it means. This includes being able to identify the different types of data (numeric, text, categorical, etc.), and knowing how to handle them accordingly. Data interpretation is the ability to make sense of data and see patterns in it. What do your numbers really mean? What insights can you find within a dataset, and what decisions will they help inform? This component also includes being able to tell a compelling story with data (e.g. data storytelling). Data actionability is the ability to use data to make decisions. What questions can you answer with your data? What hypotheses can you test? What actions can you take based on what you’ve learned? This component is essential for turning data into knowledge and using that knowledge to improve business outcomes. All three components of data fluency are essential for making knowledgeable judgments, and all three are critical to the successful operation of an organization. Why is Data Fluency Important for Data Democratization? When it comes to data democratization, data fluency is an essential component. Data democratization refers to the process of breaking down data silos and moving valuable business analytics from a departmental level to an enterprise-wide level. The key requirements of data democratization are that users will have access to data (making data accessible), and to help make users more data fluent, so they know how to use this data. Data democratization provides people with access to relevant data so they can make better data-driven decisions. And in order for this system to work, everyone in the organization needs to be data fluent. That’s why data fluency is so essential for data democratization. Improving Your Data Fluency So, with all of the above in mind, how can you improve your data fluency? Well, there are several ways. Let’s take a look at just some of them next. It’s important to learn the basics of data literacy and interpretation so that you can understand and make use of the data that’s available to you. Moreover, you can use data to drive insights and tell a story with it. Data literacy is essential in the digital economy we live in today, and being educated in knowing how to read and understand data is a critical skill. There are many different ways to get educated in data, and it’s crucial to find one that works best for you. There are online courses, boot camps, and certificate programs available, as well as traditional educational institutions that offer degrees in data science and analytics. The important thing is to start somewhere and make sure you’re constantly learning and keeping up with the latest data trends. Data sharing also plays a crucial role when it comes to your data fluency. When you share data with others, you open yourself up to new insights and perspectives. The value of data is amplified when it’s shared and used collaboratively. There are many different ways to share data, but the most effective approach depends on the type of data you’re sharing and the people with whom you’re sharing it. The key is to find the right balance between security and accessibility when sharing data and to make sure you’re all on the same page when it comes to how and why data is being shared. Tools to Process Data Data communication and processing tools are other helpful resources for improving your data fluency. But, as we’ve already discussed, it’s not enough to just have access to data—you need to be able to understand and interpret what you see as well. Data visualization and BI tools can help with that by allowing users the ability to explore insights in a more interactive way that helps them better understand what they’re viewing. So if you’re looking to tell a story with the value of data, data visualization is a powerful way to do that. When your users learn how to create informative visualizations easily, they can simplify the communication of the data-driven insights to those who need it most. Simplified Access Control Lastly, simplified access control can help improve your data fluency. If your users have to jump through hoops to get access to data, this will be a barrier to data fluency and data democratization. This includes the application of dynamic fine-grained access control to enable access to data that may contain sensitive information. An example of this is applying dynamic data masking on data, to enable access to the data without expositive sensitive information to those unauthorized to view it. When it comes to data fluency, there are several different ways to improve your users’ skills. Education is essential, as is data sharing and the use of tools to help you process data. Simplified data access control is critical to driving data in a secure way fast enough for data consumers. When users in your organization are data fluent, they can make better decisions and use data to drive more business value. Improving your data fluency is essential for the success of your business—and for the success of data democracy in the digital age. Ben is the Chief Scientist of Satori, the DataSecOps platform, simplifying access to sensitive data. - Big Data - Business Analysis - Business Architecture - Business Change - Business Intelligence - Business Process Management - Change Management - Chief Data officer - Company News - Data Governance - Data Management - Data Modelling - Data Quality - Data Warehouse - Digital Transformation - Enterprise Architecture - Enterprise Design - Enterprise Innovation - Information Strategy - IT Strategy & Management - Master Data Management - Portfolio Management - Reference Data Management
<urn:uuid:72c81ca4-95f6-416b-9b24-dbd3d30977a8>
CC-MAIN-2022-40
https://www.irmconnects.com/data-fluency-empowering-your-data-driven-organization/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335491.4/warc/CC-MAIN-20220930145518-20220930175518-00280.warc.gz
en
0.926131
1,704
3.28125
3
The MITRE ATT&CK Framework The MITRE ATT&CK framework is designed to build awareness and understanding of how cyberattacks work. To accomplish this, it organizes information into a hierarchy, including: - Tactics: MITRE ATT&CK Tactics are high-level objectives that an attacker may wish to achieve during a cyberattack. This includes stages of an attack like gaining initial access to a system, compromising user accounts, and moving laterally through the network. - Techniques: For each of the high-level Tactics, MITRE ATT&CK defines multiple Techniques for achieving the goal. For example, an attacker can gain access to user credentials via a brute force guessing attack, stealing them from the operating system, and other methods. - Sub-Techniques: Some MITRE ATT&CK Techniques can be achieved in various other ways (called Sub-Techniques). For example, a brute force password attack could be accomplished by cracking password hashes, credential stuffing, or other means. MITRE ATT&CK’s Tactics, Techniques, and Sub-Techniques drill down to a specific way in which an attacker can achieve a goal. For each of these techniques, MITRE ATT&CK includes a description of the attack, as well as the following: - Procedures: Procedures describe specific examples of using a technique. This includes malware, hacking tools, and threat actors known to use that particular technique. - Detection: For a given technique, MITRE ATT&CK recommends methods of detecting the technique. This section is invaluable for designing cybersecurity defenses because it outlines the types of information that need to be collected to detect a particular attack. - Mitigation: The mitigation section describes steps that an organization can take to prevent or reduce the impact of a particular technique. For example, the use of multi-factor authentication (MFA) is a common mitigation for techniques designed to achieve access to user accounts. Leveraging MITRE ATT&CK for Cyber Defense The MITRE ATT&CK framework is designed as a tool, not solely a repository of information. Security operation center (SOC) teams can operationalize the MITRE ATT&CK matrix in a number of ways, including: - Designing Defenses: The MITRE ATT&Ck framework outlines methods for detecting and mitigating different cyberattack techniques. This information can be used to ensure that an organization has the right defenses in place and is collecting the information required to detect a specific threat. Threat intelligence can be used to prioritize the techniques that an organization focuses on. - Incident Detection: The MITRE ATT&CK framework describes the ways in which a particular threat can be detected. This information should be used to develop detection rules in a security information and event management (SIEM) solution, next-generation firewall (NGFW), and other security solutions. - Incident Investigation: The MITRE ATT&CK framework describes how a particular attack works, and the malware that uses certain techniques. This information is invaluable for incident investigation because it allows an investigator to identify the MITRE ATT&CK Technique in use and take advantage of the additional data provided by the framework. - Infection Remediation: The MITRE ATT&CK framework describes how a particular technique is carried out and the capabilities of different malware samples and threat actors. This can help with remediation efforts since it outlines the actions that an attacker has taken and that must be undone to remove the infection. - Reporting: By standardizing terminology, the MITRE ATT&CK framework makes reporting simpler. Tools and analysts can generate reports referencing particular techniques in the framework, which provides additional detail and mitigation steps if needed. - Threat Hunting: The descriptions and detection information provided in MITRE ATT&CK can be invaluable for threat hunting. By performing a MITRE ATT&CK evaluation and working through each of the techniques described in the framework, threat hunters can determine if they have been targeted by attackers using a particular technique and whether or not existing security solutions are capable of detecting and preventing these attacks. Check Point and MITRE ATT&CK The MITRE ATT&CK framework is a valuable tool for improving communication and understanding of cyberattacks. CheckPoint has integrated MITRE ATT&CK’s taxonomy into its entire solution portfolio, including Infinity SOC and Infinity XDR. Mappings to MITRE ATT&CK techniques are included in forensic reports, malware capability descriptions, and more. This provides a SOC analyst with a number of advantages. When analyzing a particular attack, the use of MITRE ATT&CK makes it easy to understand the root causes, attack flow, and the attacker’s intent in each stage. By understanding what the attacker is trying to achieve and how, a SOC team can easily understand the scope of an attack, any necessary remediation, and how to improve defenses for the future. By integrating MITRE ATT&CK, Check Point Infinity SOC makes cyberattacks more transparent and comprehensible. To see for yourself, check out this demo video. You’re also welcome to sign up for a free trial to see how Check Point and MITRE ATT&CK can simplify and optimize incident detection and response.
<urn:uuid:5d67a1d1-cf9b-4021-b38d-f693d9afc707>
CC-MAIN-2022-40
https://www.checkpoint.com/cyber-hub/threat-prevention/what-is-mitre-attck-framework/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030336978.73/warc/CC-MAIN-20221001230322-20221002020322-00280.warc.gz
en
0.886011
1,076
2.90625
3
Object storage is an increasingly important category of storage, with nearly all the big vendors jostling in the space along with numerous smaller players. With its ability to handle large amounts of unstructured data, object storage fits the bill for some of the key storage challenges facing organisations today. So, what are the benefits of object storage and what products have the big players brought to market? Traditional file systems are based on blocks and files, storing data in hierarchical, tree-like directory structures. As the numbers of files and users grows, so does the number of directories and with it the complexity of the tree structure. As a result, it takes ever longer to locate a particular file. There comes a point where this hits performance significantly or the file system reaches the limit of the numbers of files, directories and hierarchy levels it can manage. Object storage systems were designed to resolve this problem. Implementations vary and there is no single recognised definition (see the analysis of the major vendors’ offerings below) but they do share several broad attributes. Files have no meaning in an object-based storage system. Instead, all data is broken down into objects addressed by a unique identifier, and stored in a flat address space – there are no subdirectories. Objects are retrieved using the identifiers contained in an indexed database and assembled at a higher level into meaningful data, such as files. Read more about object storage - Definition: object storage - Object storage: An architecture for the petabyte era? Objects consist of metadata, which provides contextual information about the data, and the payload or actual data. In file-based storage systems, metadata is limited to file attributes, but metadata in object storage systems can be enriched with any number of custom attributes. As very large systems, object stores and their associated databases are often distributed across multiple geographies (often using techniques such as erasure coding), that make traditional access protocols such as CIFS and NFS unsuitable. Instead, access is usually via a REST API over HTTP. Commands sent over HTTP to object storage are simple: “put” to create an object, “get” to read an object, “delete” to purge an object, and “list” to list objects. Why object storage? The consolidation of data into ever larger storage systems, a trend accelerated by the growth of cloud computing, has highlighted the limitations of traditional file storage systems. Object storage works well as a highly scalable data store for unstructured data that is updated infrequently, so it is well suited for cloud-based file content, especially images and videos. It is not so well suited to transactional data, such as database queries, due to its slower throughput. Web-scale companies such as Facebook and Google hit the limits of block and file storage some time ago. They now use object storage to surmount performance and capacity barriers. They are not alone. Cloud storage providers have been among the most eager adopters of the technology, using it to improve performance and scalability at the back end, and to ease access for users at the front end through multi-tenancy features. Vendor approaches to object storage EMC’s Atmos cloud flagship EMC Atmos Cloud Storage is the company’s mainstream object storage product line. It is aimed at multi-tenant environments, and offers a global namespace, distributed architecture, REST API, metering and chargeback. Connectivity is offered via HTTP, CAS, web services and file-based access. It can be deployed as one of two hardware-based editions – the Light Edition for single locations, the Complete Edition for distributed locations – and as a virtual appliance for VMware environments. The Atmos’s active/active architecture offers automatic replication for easier scalability, as well as versioning, compression, data deduplication and disk drive spin-down. The hardware versions are all 40U or 42U systems. The WS2-120 permits up to 360TB from 120 3TB 7,200rpm SAS disks, the WS2-240 doubles the numbers of disks and the total capacity, and the WS2-360 trebles the number of disks for a total capacity of 1,080TB. The smaller two configurations provide space for other computing systems to co-exist in the same rack, while the WS2-360 does not. The 40U Atmos G3 series was launched in December 2012 and includes support for Amazon’s S3 API. It is split into four editions, each with increasing levels of density and capacity. The densest, the Dense-480 node, consists of up to eight nodes, each housed in a 4U enclosure and carrying 60 disks, and eight servers interconnected via 10 Gigabit Ethernet. When populated with 4TB disks, total capacity is 1,920TB. Scalability is achieved by adding new systems and EMC says there is no capacity limit. The company has not quoted performance figures. EMC also offers object storage via its acquired Isilon product line (positioned as scale-out NAS), through Centera (described as a “content-addressable storage platform for data archiving”) and in its ViPR storage virtualisation/big data environment. Dell ditches object storage Dell’s DX object storage hardware platform was based on Caringo’s CAStor. Then the company switched to selling the product as software-only in 2013. Now, however, Dell’s website points potential buyers to its Compellent SC8000 Storage Center Controller, a traditional SAN product, which seems to indicate the company no longer sells object storage. Hitachi Content Platform aims at cloud and archive Hitachi Data Systems’ Hitachi Content Platform (HCP) is a distributed, object storage system aimed at public and private cloud providers to enable data sharing, synchronisation, analysis and retrieval. The system uses redundant nodes – Hitachi CR220 servers – and provides support for Multi-tenancy with the ability to subdivide each tenancy into multiple namespaces, and for multiple protocols and scalability up to 80PB. Each server includes five disks in a RAID5 configuration. Entry to the HCP series starts with the HCP 300 with a minimum of four and a maximum of 20 nodes connected over 1GbE and offering up to 140TB. The highest capacity of the five-strong series is provided by the HCP 500XL, which consists of up to 80 2U nodes with four 10GbE and two 4Gbps FC ports. The system is accessible using a REST API, as well as traditional protocols such as NFS and CIFS. Uniquely, SMTP is also provided to allow email archiving. Configurable data protection by redundancy provides for up to four copies of each piece of data, along with continuous data integrity checking. It includes data retention policies and WORM features for compliance, along with data shredding where appropriate. Archiving features include disk spin-down and tiering, and versioning over HTTP REST only. Metadata search is integrated into the system. Security features include layered access control for administrative, tenancy management and end user purposes. Tenants’ passwords are inaccessible to system admins. HP’s archive appliances Bringing together HP’s StoreAll 9320 and 9730 storage systems with its StoreAll 8800 storage node, the StoreAll 8800 series was launched in December 2012 and is aimed at organisations that need to store and access archives of unstructured as well as production data. It uses an object store to manage a maximum 16PB of capacity with billions of objects, scaled by adding nodes and controllers, in what HP describes as “a single hyperscale, economic, ultra-dense appliance”. Disks are arranged in multiple pairs of nodes, each node consisting of a 2U enclosure accessible over a range of protocols including HTTP, WebDav, REST API, OpenStack object storage API, NFS, CIFS and FTP. In each 2U enclosure are 36 or 70 7,200rpm SAS disks up to 4TB capacity, depending on model. Network connectivity is over 10GbE, with 1GbE ports provided for management purposes. Features include snapshotting, replication, data deduplication, and policy-based data tiering and retention, plus continuous data integrity checking. Using technology from the firm’s acquisition of Autonomy, the system also includes automatic indexing and fast retrieval of data for analytic purposes, and aims to provide real-time access to and querying of data. IBM’s Elastic Storage and SoftLayer IBM offers two routes to object storage. It does not yet produce an on-premises system itself but has pledged to do so. The first is its Elastic Storage software, launched in May 2014. The technology is based on the company's General Parallel File System (GPFS) and the Watson supercomputer. It uses Watson's cognitive technology to handle workloads generated from cloud, analytics, mobile and social media, and can be deployed on-premises or via IBM's SoftLayer cloud. In similar fashion to EMC’s ViPR, it works as a control plane that offers block, file and object storage access with automated tiering, guided by analytics, using patterns, storage characteristics and the network to determine where to move data. It also includes automated backup and snapshots, using one copy of the data for snapshots and their replication, so reducing storage consumption and costs. Elastic Storage sits above OpenStack Swift, so users can access and manage data across private and public clouds. Open-source OpenStack Swift is accessed through a REST API and can scale horizontally to petabytes of data through the addition of nodes, which typically equate to servers. IBM's second route to objects is SoftLayer’s object storage-as-a-service offering. SoftLayer was acquired by IBM in 2013 for its cloud-scale technology. Charged on a GB per month basis and scalable as required, the service is aimed as those that want to store static data such as VM images, media and email archives. SoftLayer claims every facet of its platform can be automated and controlled by a single management system with its own API. It can be managed and controlled using a number of access technologies. Developer access is via a REST API, customer access is via a web portal or mobile application. The API provides more than 2,200 documented methods across 180 discrete services, and supports, among others, SOAP and XML-RPC interfaces, which SoftLayer says provides full customer access to all the services available to the service provider. The search service, for example, allows requests to be made to search an entire account, a particular container or a specified path, based on the URL entered at the time of search. Higher-level management is provided through a customer web portal that enables server and storage control, performance metrics and account management. A mobile application supports ticket creation, basic server management and bandwidth monitoring. For security, SoftLayer offers a multi-layered approach, with IPS and IDS for the network and servers, and detailed scanning and logging capabilities. It also offers security services including Citrix NetScaler’s ICSA-certified Layer 7 attack signature detection, McAfee anti-virus, endpoint protection, data encryption and malware detection, and Nessus Vulnerability Scanner. NetApp’s archive-targeted appliances In 2010, NetApp acquired Bycast, a developer of object-based storage software whose technology is now the basis of NetApp StorageGrid. NetApp combines the StorageGrid software as a VMware-based virtual appliance with NetApp E-Series storage systems to offer an object storage appliance. The system provides a distributed global namespace with automated, policy-driven data lifecycle management, and is aimed at organisations that need to store and manage archived data, especially large datasets. Hardware support is offered for NetApp's E2600 Series controllers, DE1600 and DE6600 shelves with 2TB or 3TB NL-SAS drives, and 500-plus tape drive, robotic library and autoloader products. Features include versioning, rules-based AES-256 or SHA-256 encryption, compression, and provision for multi-tenant gateway deployments. It provides a notification feed for third-party application billing and QoS monitoring, along with an audit feed for chargeback, search integration, security diagnosis, compliance events and customised reporting. Access is provided both via traditional protocols such as NFS and CIFS, as well as native object command sets over an HTTP API.
<urn:uuid:cd4c4ff7-00a3-4fea-96e8-487c5e09259a>
CC-MAIN-2022-40
https://www.computerweekly.com/feature/Big-players-target-object-storage-at-cloud-and-archive
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030336978.73/warc/CC-MAIN-20221001230322-20221002020322-00280.warc.gz
en
0.922806
2,618
2.546875
3
This was one of our most popular blog posts of 2018. We’ve updated this post with new information to continue providing valuable content. Artificial intelligence is on the rise and changing how historically human-powered business is done. One example of this is the effect of AI on trading. The Coming of Age of Artificial Intelligence Chances are you use forms of AI every day. Talking to the virtual assistant in your phone or home to get the weather or figure out the name of that actor who’s in everything is one example. A form of AI you may not use every day but has definitely shaped your life is the auto-pilot function on commercial airliners. Artificial Intelligence and Machine Learning have recently become a major part of trading. AI in Trading Artificial intelligence works by setting up computer systems that operate like the human brain and nervous system. This enables the system to learn behaviors and patterns and thus predict trends. That means it can analyze data to make profitable decisions. This analysis and prediction model is useful in short- and long-term trading strategies and is becoming increasingly visible. That said, the use of the technology is going slowly. It requires capital to invest in tech and people. But more AI is being applied because it offers a quicker pace thanks to no limit on how many transactions can be run. There are also security features inherent in the technologies that come with Artificial Intelligence. Increased use of Artificial Intelligence in trading could yield many benefits. Because AI replaces humans it reduces the time needed for analysis and removes human error. It also removes the second guessing and emotional reactions that come with humans. Thus, it’s not unlikely that increased AI in trading will bear markets that are more efficient and less volatile. AI As A Hedge Fund Tool Sentient Technologies, an AI company based in San Francisco that runs a hedge fund, developed an algorithm for trading. It scans huge amounts of data noting patterns and forecast trends. These patterns and trends allow it to make successful trading decisions. Of course, most of these decisions are currently run as simulations. When trades are sucessful they are moved to live trades where the artificial intelligence can continue to learn and improve. They are successful not only at predicting solid trades but at doing so with impressive speed. Sentient reduced 1,800 days of trading to a few minutes. Next Steps in Trading Using Tech It’s not surprising that there are many reservations when it comes to using computer intelligences to control markets. But there are massive benefits to this implementation. Fast-paced data analysis and predictions provide insights that can help the human brains behind trading make financially sound decisions. Throughout 2018, technology has continued to affect the world of trading. Over sixty-one million men and women are involved in online share trading today. Plus, that number continues to rise. High-frequency automated trades are already taking place using AI. This allows human traders to focus on more complex areas that can’t be predicted by calculations. Investment banking firm Goldman Sachs is a leader in these cutting-edge developments. Since 2000, they have replaced almost all of their cash equity traders while increasing their tech staff. Giving AI more of the simpler work frees traders up to predict the next big tech boom. Experts agree that while machine learning has a small margin of error, it’s equal to the impact of human error. Plus, AI will likely continue to handle time-consuming tasks. Overall, these advancements are changing the industry in a way that’s sure to impact the future. Leave a Comment You must be logged in to post a comment.
<urn:uuid:09154e6c-b546-4ba2-8d2b-713792a57ff8>
CC-MAIN-2022-40
https://www.consultcra.com/artificial-intelligence-trading/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337516.13/warc/CC-MAIN-20221004152839-20221004182839-00280.warc.gz
en
0.944584
737
2.71875
3
With the introduction of cloud technology, workforce accessibility, and increased inter-organizational cooperation, the threat surface of network attacks against device networks is increasing. Around the same time, traditional perimeter defence techniques dependent on trust can't fully cope with the security demands of the current decentralized enterprise. This combination makes the network unacceptably fragile. A new network security model is required that replaces the concept of a trusted network inside or beyond the corporate perimeter. The purpose of a software defined perimeter (SDP) is to implement perimeter features for dynamically configured perimeters for cloud, hybrid environments, and on-site data centre infrastructures. There is always a complex tunnel that is created automatically during the session. That is a one-to-one mapping between the requesting party and the trusted network. What is software defined perimeter? Software Defined Perimeter (SDP), also known as the Black Cloud, is a network technology strategy that emerged from the research performed by the Defence Information Systems Agency (DISA) under the Black Core Network Global Information Grid (GIG) program around 2007. The Cloud Security Alliance (CSA) has developed a software defined perimeter architecture to monitor identity-based access to services. Connectivity in the SDP technology is based on a need-to-know paradigm under which the device's status and identity are checked before access to the network infrastructure is granted. Application infrastructure is essentially "black" (a DoD term meaning that the network cannot be detected) without identifiable DNS or IP addresses. The inventors of these technologies say that the Dynamically Specified Perimeter mitigates the most popular network-based attacks, including server inspection, access denial, SQL injection, operating system, and device bug exploits, pass-the-hash, man-in-the-middle, pass-the-ticket, and other attacks by unwanted users. Why does security need a software defined perimeter? The decentralization of industrial businesses is an established reality. Cloud benefits — cheap computing, pay-per-use billing, disaster recovery, and on-demand resources — are expected to continue to fuel usage rates for some time. Equally evident is the truth of security breaches, which tend to be growing in regularity (if not severity) with each new attack. There is a disappointment with the number of data leaks in the cloud, which negatively affects the use of perhaps the most creative technology of this time. In order to strengthen the cloud value proposition, there is a growing need for a security model that is as robust and latent as the very possibilities and challenges cloud computing offers. This should reduce the surface area for threats while preventing detection by intruders; it should be profoundly rooted within an entity to protect the data as the properties of the enterprise they are. Software-Defined Perimeter (SDP) is a radical defence architecture that provides these and other advantages. When fully deployed, the network layer secures gateways to and from clouds for unassailable authentication with veiled micro-tunnels that malicious hackers cannot see or track. Components of a software defined perimeter: Access management software helps you to monitor when individuals can join. Day and day are regulated by the role of the calendar. Once an individual joins the access control program, you will appoint them to a group, and then arrange the group with the days, hours, and holidays you like. Many people use access management to do so than ever before — and the industry has made improvements in access control technologies to thank for that. Functions that were formerly hardware-based can now be supported by cloud-based applications, making access control more affordable; and application programming interfaces (APIs) allow access control applications to be combined with other programs, such as video or enterprise requirements planning (ERP), making access control more efficient and scalable. Performance management and reporting Performance management includes tracking, documenting, and managing change – from people who work with a business and the enterprise as a whole – with a view and enhancing results. Examples of performance improvement systems or resources include performance evaluation, key performance indicators (KPIs), and improvement dashboards. Essentially, performance improvement is what companies are doing to make them more efficient and remain ahead of their rivals. Since it covers such a wide field, performance management is typically split into two sub-disciplines: organizational performance management and human performance management. Security compliance and policy management Security Policy Management is the method of defining, enforcing, and maintaining the rules and procedures that all entities will follow when accessing and utilizing the IT assets and services of an entity. The purpose of these network protection policies is to resolve security risks and enforce techniques to minimize IT security weaknesses, as well as to determine how to recover from a system breach or when a network intrusion occurs. In addition, the policies offer guidance to employees about what to do and what not to do. They also determine who has access to what properties and services and what the implications are for not complying with the laws. Increasing the use of network security policy management tools is one of the key developments in the security policy management industry. Trends and innovations in the technology Over the years to come, technological innovation and growing investments are expected to drive the global market. Significant developments in telecommunications and increasing internet and broadband penetration have further accelerated the development of the IoT industry. It is observed that tech giants and corporate investors make important investments in the Internet of Things market. Employees have shown a greater tendency to use a single device for both personal and professional purposes to prevent the hassles created by bringing separate dedicated devices like theft and inconvenience. The need to bring multiple devices, or switch between work and personal devices, is completely removed in such an environment. It has been predicted that increased penetration of smartphones and the roll-out of services such as 4G LTE would drive market growth in both the developing and developed economies. As employees get more familiar with their own devices, they can perform tasks more effectively, and in doing so, improve productivity at work. The Software Defined Perimeter Market has established technologies that offer a complex architecture that reshapes the potential of network defence. Our software-defined perimeter system provides easy cloud migration protection, smooth access to infrastructure and safe connectivity to cloud environments – like Infrastructure as a service (IaaS), Platform as a service (PaaS), and more. Using the Software-Defined Perimeter (SDP) architecture, enterprises can safely provide customers using access to their IaaS services without impeding enterprise customers or IT productivity. Nevertheless, when correctly implemented, SDP deployment will be a tool for improving how network security is done across the whole enterprise – both on-site and cloud-based. Free Valuable Insights: Global Software Defined Perimeter Market to reach a market size of USD 10.7 billion by 2025 With SDP, companies will provide a standardized and policy-driven network management framework that protects their entire networks (both on-site and cloud) and their entire user base. This is a convincing dream – but one that can be easily accomplished with the SDP. Many companies around the world have used SDP to improve their protection, reduce their threat surface, increase the efficiency of a company and IT workers, and their burden of enforcement – while saving money.
<urn:uuid:7061b22a-1474-40f4-a6ff-fb30aa449cdd>
CC-MAIN-2022-40
https://globalriskcommunity.com/profiles/blogs/all-you-need-to-know-about-software-defined-perimeter?context=tag-network
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334528.24/warc/CC-MAIN-20220925101046-20220925131046-00480.warc.gz
en
0.940984
1,489
2.59375
3
Thank you for Subscribing to CIO Applications Weekly Brief Amazingly Simple Tips to Ensure the Privacy of Students Online Students' digital presence is more ubiquitous than ever thanks to distance learning. FREMONT, CA: Most learning experiences now include educational technology. When one add in social media identities, today's pupils' digital footprints have become almost as crucial as their physical ones. As a result, it's even more critical for teachers and parents to do everything they can to help pupils be safe online. The terms "digital citizenship" and "online safety" are frequently helpful in the educational world. Industry and academic leaders, of course, are working relentlessly to ensure that all students have access to materials that support a safe and secure online presence. Calm the Storm and Work Methodically In the era of remote learning, there are four phases to recognizing that pupils are at high risk. The initial stage is becoming conscious, followed by panic. This terror might manifest in various ways, but it almost always results in a state of paralysis and lethargy. Allow oneself to experience this before moving on to the third level: acceptance. Accept the fact that this is now the new normal for students. Then, in the fourth step, action, safeguard pupils promptly — and quietly. It's critical to develop a plan at this point but also to remain flexible and patient as one navigates unfamiliar seas. Create a Safe and Secure Digital Space Begin with a strategy and a clear understanding of the objectives. For example, decide if one wants pupils to have free access to social media or if a more regulated approach is preferable. Decide how much supervise kids' internet activities and what digital tools they'll be allowed to utilize. Communicate and Collaborate with Your Students It's critical to express the vision for students' digital lives to them directly once formed one. Common Sense Media, a non-profit company, has pictures and videos that can help with this conversation. Collect Information About the Digital Risks It's also crucial to educate oneself on the various online hazards that students face. Again, Common Sense Media has a wealth of materials to help keep us informed about these issues, from misinformation to cyberbullying. Connect with Organizations That Can Assist Numerous organizations are rich in resources and ideas. Collaboration with local organizations and state authorities that perform research and arrange events to share and develop innovative ways to protect students online and spread awareness about digital safety and citizenship is one of the unique ways to stay connected. To uncover more resources, consider reaching out to other schools or professional learning networks.
<urn:uuid:100b64b2-36f0-45de-abcb-d73846b8d4f9>
CC-MAIN-2022-40
https://www.cioapplications.com/news/amazingly-simple-tips-to-ensure-the-privacy-of-students-online-nid-8244.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334528.24/warc/CC-MAIN-20220925101046-20220925131046-00480.warc.gz
en
0.942632
529
2.796875
3
8.12.2022 • Blog Do you have a firewall implemented into your network? Maybe your network is secure and your documents are protected thanks to your firewall. This is great but if you need a client or a remote employee working at your company to have access to a document or your network then you'll have to find another way for them to access it. Port forwarding can help with this and allow you to send out and receive data or access blocked websites while your firewall is still in play. Without port forwarding, it can be difficult for you to grant your clients access to these documents. Whether it be client agreements or contracts, you want to keep them safe but you also want your clients to be able to have access to these documents as well. The idea of port forwarding is to be able to access private networks from outside. Without port forwarding, external devices cannot see internal IP addresses. Here at Coeo, we pride ourselves on giving our audience information on all things telecommunications. Educating our future and current customers is always a goal of ours. By the end of this article, you will know what port forwarding is, the three different types of port forwarding, and how to set up port forwarding. This article will allow you to decide if you should use port forwarding or not. What is port forwarding? Port forwarding, also known as port management, allows remote servers and devices on the internet to be able to access devices that are on a private network. If you are not using port forwarding, only devices on that private internal network can have access to each other or your network. Before we get any further, you may be asking, “What is a port?” A port is a communication endpoint and is identified as a port number that is always associated with an IP address. Of course, the whole reason you have security on your network is to keep outside sources from penetrating your network. If you are not able to grant access to an outside source, chances are you have good security over your network. However, there are some times when you need to grant access to an outside source whether it be to your clients or business partners to your documents or other elements within your network. Port forwarding will help you do that. The way that port forwarding works is you configure an internal IP address on the firewall. Then you associate that IP address with an external IP address that other devices on the Internet know how to reach. This allows you to be seen on a private network while still being protected from outside sources. 4 ways to mitigate the security impact of port forwarding 1. Only forward ports to devices in a DMZ Forwarding ports through a DMZ gives you an extra layer of security and protects your network from being vulnerable. A DMZ or demilitarized zone is a perimeter that protects and adds an extra layer of security to an organization’s internal network from untrusted outside traffic. 2. Only forward the specific ports that are required Forwarding to specific ports that you need will make your network much safer. If you were to forward to lots of different ports that may not be necessary, you could be opening up your network to cyber-attacks and data breaches. 3. Restrict the IP addresses that are allowed to access the port forward. If you aren’t restricting IP addresses, your server could be accessed by a malicious source. You can allow access to specific IP addresses so that only those IP addresses have access to your server. If you have a vendor that requires access to a server that you manage, then you can restrict access to the port forward for that server to only the IP addresses that your vendor uses. 4. Don't port forward at all. Instead of port forwarding, you can use services like SASE which uses secure VPN tunnels to allow authenticated and secured access to resources. How to setup port forwarding Port forwarding can be set up through your firewall. Pretty much any firewall is capable of doing port forwarding. Whoever manages your firewall will determine who sets up and manages port forwarding. If you have your own firewall managed by you or your IT team then you or your IT team will have to set up port forwarding. If you have a managed firewall outsourced by someone else, they will do the port forwarding for you. Port forwarding can be dangerous if you don’t do it correctly so make sure you or your IT team has the knowledge to set it up. Likewise, if you have a managed firewall you want to make sure that they can safely do this process for you. If not done correctly, you could be opening up your network to other outside entities that could lead to viruses, data breaches, and cyber-attacks. Here is a short tutorial that you can use to set up port forwarding. Next steps for port forwarding In this article, you learned about port forwarding and the different types of port forwarding. This will help you receive or send out access to clients or employees that don’t have this access but need it. Here at Coeo, we pride ourselves on giving you all the information about anything telecommunications. The first thing you should do after reading this article is to figure out whether or not you need port forwarding. You can start by asking yourself: - Do I have remote working employees who need access to my organization’s office network but can’t? - Do I have clients that need access to our office network or other important documents but can’t? - Do I feel safe enough to start port forwarding? If you answered no to these questions then port forwarding may not be for you but if answered yes to these questions, you may want to consider trying port forwarding. If you want to start port forwarding, you can read “How to Port Forward.” Once you have read the tutorial, you will have a better idea of how you can start port forwarding safely. If you have made the decision that you want to start port forwarding, talk to whoever manages your firewall, whether that be your IT team or an outside source, to determine the next best steps for you. If you still want to continue with port forwarding, you can talk to our sales team by calling (844) 300-2636.
<urn:uuid:6fff188f-c7f5-4f0d-869f-a0bffb34ad24>
CC-MAIN-2022-40
https://www.coeosolutions.com/news/what-is-port-forwarding?hsLang=en
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334528.24/warc/CC-MAIN-20220925101046-20220925131046-00480.warc.gz
en
0.94966
1,289
2.640625
3
Safer and Cleaner Cities with Innovative & Sustainable Mobility Solutions | SPONSORED With CO2 emissions and road fatalities on the rise, the need for sustainable mobility solutions is now more than ever. Currently, the CO2 emissions in the transport sector are about 30% in the case of developed countries and about 23% in the case of the total man-made CO2 emissions worldwide, according to UNECE. There is a widespread agreement to reduce CO2 emissions from transport by a minimum of 50% at the latest by 2050. Also, with 1.3 million annual road deaths, the UN wants to halve the number of fatalities by 2030. Best way forward to address CO2 emissions and road fatalities is to adopt sustainable mobility solutions driven by innovative mobility platforms. Connected vehicles, autonomous driving, AI-driven smart technologies, etc. are shaping the future of transportation and AR transportation one solution at a time. At the heart of all these technologies lie efficient mobility platforms that connect data from the vehicle to the Internet in near real-time. This panel will focus on how innovative technologies will revolutionize low-emission, sustainable mobility – an important part of any smart city project. Sponsored by HiveMQ
<urn:uuid:ccc928e1-55a9-419c-ad8a-588b42c86429>
CC-MAIN-2022-40
https://www.iiot-world.com/smart-cities-buildings-infrastructure/smart-cities/safer-and-cleaner-cities-with-innovative-sustainable-mobility-solutions/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334915.59/warc/CC-MAIN-20220926175816-20220926205816-00480.warc.gz
en
0.923536
247
2.59375
3
23 Nov Connectivity & 5G Can Assist in Decarbonizing Industries The pandemic era has proven that emission levels and company practices can change quickly. Recent research has shown that connectivity and 5G can assist the decarbonization of various industries. A recent MIT Technology Review Insights Report supported by Ericsson has found supporting data that the telecommunications infrastructure can assist in the decarbonizing of industries. The report specified how connectivity, especially with the upgrade to 5G, within the telecom infrastructure are supportive of this. In August 2021, a climate change report by the UN’s Intergovernmental Panel on Climate Change’s or IPCC, “Estimates of the global emissions outcome of current nationally stated mitigation ambitions as submitted under the Paris Agreement would lead to global greenhouse gas emissions in 2030. Avoiding overshoot and reliance on future large-scale deployment of carbon dioxide removal (CDR) can only be achieved if global CO2 emissions start to decline well before 2030.” Global greenhouse gas (GHG) emissions must be cut nearly in half to fully abide by the Paris Agreement. This means organizations and industries must begin planning and implementing their GHG reduction now. Decarbonizing With the Telecom Industry Of all the various technologies available to enterprises to assist in reducing GHG and the carbon footprint, connectivity and 5G can be one of the most powerful of these tools. 5G telecom infrastructure can assist in greener efforts with data-based decarbonization through more efficient digital channels. “5G and other digital mobile technologies can generate a transformational acceleration of decarbonization efforts.” The following 5G enhancements can provide the buildings blocks to reducing emissions by half by 2030. - Increase Transmission Speeds - Provide Ultra-Low Latency - Enhance Real-Time Analytic Capabilities - Manage Digital Devices and IoT Networks at Larger Scales and Flexibility - Advance in Digital Transformation - Increase of Automated Systems - Increase of Mobile Applications - Improve Energy Efficiency - Enhanced Remote Monitoring - Provide Sustainable Operational Processes - And Many Other Digital Improvements “5G-enabled cellular networks are a clear enabler for each of the so-called ‘four Ds’ of renewable energy: decentralization, decarbonization, digitalization, and—increasingly important for this fast-changing and entrepreneurial space—democratization.” Antonello Monti, Professor, RWTH Aachen University Areas That Can Benefit Almost every industry can benefit from the previously mentioned improved digital channels. Even those industries that have the largest impact. The energy, manufacturing, and transportation industries have the potential to reduce GHG by half with the proper implementation of 5G connectivity through various digital channels. Connectivity and 5G can assist various industries with their impact on decarbonization. Reducing an industry’s carbon footprint will not only impact that industry but will also positively impact our entire planet. If your organization wants to learn more about 5G infrastructure, please contact the experts at GeoTel today. Written By: The Experts at GeoTel
<urn:uuid:7eaa95b9-0f99-4c8c-b184-de2668a43bd1>
CC-MAIN-2022-40
https://www.geo-tel.com/connectivity-5g-decarbonizing-industries/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335504.22/warc/CC-MAIN-20220930181143-20220930211143-00480.warc.gz
en
0.892817
643
2.515625
3
Artificial intelligence and allied technologies are driving processes which were till now available only to humans such as learning, problem solving, pattern recognition and decision making. Curated data can be used to train models using statistical methods for solving problems related to classification of data and predicting outputs. Natural language processing, entity extraction, machine learning and deep learning are a few areas in Artificial intelligence which are making an impact in the world of technology. Futuristic scenarios such as Terminator aside, AI is already here and making definitive difference in domains such as Finance, Medicine, Science and Technology just to name a few. The customer is a primary player in the discount card market in the UK and Europe region. The products offered by these customers include student cards which open a world of discounts for the student along with other benefits such as proof of identity, age, and place of study. The customer has been in business for a period of two years and has data on about two million users who are either free or paid members of this card. The card is also available as a Mobile application based digital card removing the need for carrying a plastic card around. The customer had a lot of data being collected on a day-on-day basis which includes but not limited to - Student data - Card sales data - Discount and Offer redemption data. - Customer help desk data - Customer community interaction data The team from Arrk group did an “EmbArrk” journey with the customer to understand sources of data and the type of information which they were interested in. The Data science team at Arrk studied the various sources of data and did a cross relation of the significant data bits which were required to build a meaningful model. Date engineering and integration specialists worked with the customers to fetch this information in a Redshift Data Warehouse. Here the data was collated into meaningful views which would act as a data feeder to the Recommender Engine model. Once the data pipeline was estabished, the Data scientists on the team built a deep learning model using Python and Tensorflow. The model takes current offer adoption data along with demographic and purchase data as an input to train the model. The model trains of hundreds of thousands of such adoption data and infers the deep relations between various features. The created Tensor flow model was validated against test data and achived 70 % accuracy during trials. The Python model was served as a library to a Lambda function which was sitting behind an API Gateway. Devices and collaborating services made direct calls to this API end point with user data to get the best possible recommendations for the data. The model keeps learning every fifteen days on the latest adoption data to ensure that it is as accurate as possible. Replacing an older static recommender with a Intellegent recommender has improved the customer offer adoption by approximately 30% as can be observed in the monthly offer reports. The models accuracy will improve over time as we also feed it the adoption levels of the recommended data. The DataScience team at Arrk can work with you to use your collated data to build such intellegent recommenders which will give features based recommendations and help improve your sales and profitability. Please reach out to one ouf our Data experts below to help you start up on this journey.
<urn:uuid:410f49b4-3e13-4eec-9e37-b95c173ab2a4>
CC-MAIN-2022-40
https://www.arrkgroup.com/thought-leadership/the-intelligent-recommender/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337244.17/warc/CC-MAIN-20221002021540-20221002051540-00480.warc.gz
en
0.970914
668
2.625
3
The rise of the Internet of Things (IoT) has created a magnitude of new connected devices and in turn, a magnitude of entry points waiting to be breached. An even bigger issue is that many IoT devices are easier to target than conventional devices, making them the new endpoint of choice for cybercriminals everywhere. Security can often be an afterthought with many IoT projects, while innovation, agility and cost savings take priority. While these objectives are important, without a robust IoT security strategy firmly in place, businesses are at risk of exposure to a growing number of security threats that directly target IoT systems. According to research, 84% of organisations have experienced an IoT-related security breach and trends reveal that the increased risk of data breaches remains one of the greatest barriers to IoT adoption. The threats are real In October 2016, the world experienced the largest Distributed Denial of Service (DDoS) attack to date. Known as the Mirai botnet, the attack caused havoc across the internet by exploiting unsecured IoT devices and using them to cause widespread disruption of a number of services including Twitter, CNN, Reddit, Spotify, and Netflix. More recently, Cisco Talos uncovered a Russian botnet that was affecting at least 500,000 vulnerable network-access storage (NAS) devices and routers located around the world. These vulnerabilities were made possible with malware known as VPNFilter, which allowed hackers to control infected devices, including the ability to take them offline. Additionally, the malware allowed hackers to snoop on data as it passed between affected routers and learn more about the software used to manage the critical infrastructure. What is notable about these high profile attacks, remarkable even, is that they were carried out via simple, innocuous IoT devices. By compromising devices ranging from home routers to air-quality monitors and surveillance cameras, criminals found a way to exploit systems, hijack networks and steal sensitive business information. IoT presents a major opportunity for businesses and organisations everywhere to fuel digital transformation and compete more efficiently in today’s digital economy. IoT presents a major opportunity for businesses and organisations everywhere to fuel digital transformation and compete more efficiently in today’s digital economy. The number of connected devices worldwide is expected to exceed 20 billion in 2020 and we are becoming increasingly reliant on the advance of life-sustaining and practical use cases for IoT. From control systems that deliver power and water to self-driving cars and smart meters, this growth in dependency on network-connected technology is outpacing the means to secure them effectively. While Mirai’s actions may have prevented users from accessing their favourite movies or playlists, the threat of more devastating attacks remains. As the number of IoT devices increases exponentially, here is a brief sample of the mayhem that nefarious agents could cause if they took control of IoT devices: - Control of temperatures that causes servers to overheat and malfunction - Disruption of critical infrastructure, transportation, telecommunications and the power grid - Aiding criminals in infiltrating buildings by taking control of electronic security systems and devices - Hijacking medical devices with the intent to injure, kill or hold organisations to ransom - Infiltration of smart city devices causing widespread disruption to parking and traffic safety systems Working toward a more secure IoT framework When developing a security strategy, it is important to take a multi-layered approach and prepare for unexpected threats and challenges. This means simulating IoT-specific breaches, conducting risk assessments and creating security playbooks that allow the organisation to respond quickly and effectively to breaches while still maintaining efficient operations and positive customer experience. Secure products and advanced security functions and features should be the foundation of this strategy. However, traditional blacklisting and antivirus products are not going to be enough to secure IoT devices and protect them from the avalanche of threats that dynamic IT environments face. One example is a centralised management platform delivered over a standards-based secure mobile network that offers the option of private end to end connectivity of IoT devices without any exposure to the Internet. No magic bullet IoT devices can be situated anywhere on a distributed network. They operate across a variety of systems and in different locations. Most are managed remotely. For this reason, a comprehensive and consistent security strategy is required. There is no magic bullet solution or all-encompassing security solution for IoT. There is no magic bullet solution or all-encompassing security solution for IoT. Taking an existing IT security strategy and applying the same elements to your IoT infrastructure could be a costly mistake. Business leaders should instead focus on creating an IoT security ecosystem that integrates with all aspects of security, including physical security, cybersecurity and operational technology security. In doing so they can experience the opportunities and possibilities that IoT has to offer without unnecessary exposure to risk.
<urn:uuid:72d85fdf-fefa-47e4-b9f4-fb0d6308ae88>
CC-MAIN-2022-40
https://insights.ckhiod.com/insights/the-security-of-your-iot-network-is-a-clear-and-present-danger
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337415.12/warc/CC-MAIN-20221003101805-20221003131805-00480.warc.gz
en
0.940996
975
2.765625
3
Universities are divided into the business side, the academic side, and, in some cases, the research side, so those networks and systems should be compartmentalized as such. "One thing we're going to look at is more, proper segmentation across all of our campuses," Nebraska's Mauk says. That would mean keeping the business systems separate from the academic systems and student life, for example, he says. "Proper security between those segments [poses] larger challenges," he says. "Universities were not built that way 10 to 15 years ago." |5 Tips For College Kids To Avoid ID Theft 1. Set unique passwords for each account -- and don't give it to your boyfriend or girlfriend. That means not reusing the same password for multiple accounts. All it takes is one of those accounts to be compromised, and the attacker can reach the others as well. Create a minimum eight-character password with a mix of symbols, upper- and lowercase letters -- and no dictionary words within it. "Pick a secure password, and don't tell it to anyone else," says Paul Kenyon, co-founder and COO of Avecto, a Windows privilege management provider. Aaron Massey, a postdoctoral fellow in the School Of Interactive Computing at Georgia Tech, says that's his No. 1 tip. "Passwords are the No. 1 thing to talk to students coming in. They haven't established a strong routine for" creating strong and unique passwords, he says. 2: Run antivirus software, and keep it updated. "AV is seen as a tax you have to pay to use computers, but I think it's a necessary one," Kenyon says. "And keep it up-to-date." Oh -- and if it doesn't automatically scan on a very regular basis, execute the scan yourself. 3: Don't visit sketchy or unknown websites. "Students have a tendency to go to websites with unlicensed and uncopyrighted music, where they download quite readily," for example, and that is risky, he says. 4. Don't click on links or attachments in emails. 5. Keep an eye on your financial statements. Have a comment on this story? Please click "Add Your Comment" below. If you'd like to contact Dark Reading's editors directly, send us a message.
<urn:uuid:4301ba83-f6bb-4e39-8e1f-8e395ba669ef>
CC-MAIN-2022-40
https://www.darkreading.com/attacks-breaches/lessons-in-campus-cybersecurity
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337415.12/warc/CC-MAIN-20221003101805-20221003131805-00480.warc.gz
en
0.958244
495
2.625
3
ZDNet: HGST (formerly Hitachi’s disk unit) has announced that it has developed a helium-filled disk drive. The helium units use 23 percent less power than air-filled drives. In addition, they run 4°C (7°F) cooler, which reduces the power necessary for cooling in a data centers and results in a total 45 percent reduction in watts-per-TB. Encasing the drives in helium also allows the company to put two more platters in their disks, increasing the total storage space. Why helium-filled drives? As a disk drive spins, it causes air turbulence and drag that hinders the performance of the drive. Because helium is lighter than air, it reduces the drag on the spinning disks. The article also notes, “No, helium-filled drives won’t float away if they aren’t tied down.”
<urn:uuid:31b6570b-617c-4e16-84fe-648045de30e1>
CC-MAIN-2022-40
https://www.datamation.com/trends/hgst-announces-helium-filled-disk-drives/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337415.12/warc/CC-MAIN-20221003101805-20221003131805-00480.warc.gz
en
0.921083
185
2.953125
3
It is important for malware authors to keep a solid network connection between their malware on compromised computers and their own servers so that the malware can receive commands and be updated. However, communication between the malware and the malware servers may be filtered by a gateway or local firewall, or blocked by an intrusion prevention system (IPS). Consequently, malware authors try to find more secure methods of providing communication between the malware and the servers. For example, I wrote a blog last November detailing how Backdoor.Makadocs uses the Google docs viewer function as a proxy to maintain a solid connection between the malware and its servers. More recently, I discovered a Trojan horse that uses Sender Policy Framework (SPF), which is an email validation system designed to prevent email spam, to achieve the same goal. Basically, SPF consists of a domain name server (DNS) request and response. If a sender’s DNS server is set up to use SPF, the DNS response contains the SPF in a text (TXT) record. Figure 1. How SPF works The following matrix contains some SPF examples of major legitimate sites. Table 1. SPF examples of major legitimate sites The point for the malware author is that domains or IP addresses in SPF can be obtained from a DNS request and this DNS request doesn’t need to be requested from a computer directly. Usually the local DNS server is used as a DNS cache server. The DNS cache server can send a request instead of the computer. Discovery of a Trojan horse using SPF Recently, I discovered a Trojan horse (detected by Symantec products as Trojan.Spachanel) that uses SPF. Basically, it hijacks a Web browser to inject malicious content into every HTML page. The process of how the malware carries out an attack is illustrated below. Figure 2. How Trojan.Spachanel carries out an attack Below is a captured SPF record that is received from the attacker’s DNS server. Figure 3. A malicious SPF record Why did the attacker use SPF to get the malicious domains or IP addresses? My guess would be that it is because the attacker wants to hide communication in legitimate DNS queries. If this malware connects to the attacker’s server by a higher port number using the original protocol, it may be filtered by a gateway or local firewall, or blocked by an intrusion prevention system (IPS). In some cases, specific domains are blocked by a local DNS server, but this malware generates a domain that is rarely filtered. Furthermore, DNS requests are generally speaking not sent directly. Usually there is a DNS cache server in the network or in the ISP network, which makes it difficult for a firewall to filter it. Therefore, this is the attacker’s attempt to maintain a solid connection between the malware and the attacker’s server. What happens after infection? Figure 5. Legitimate site appearing to display malicious content So far, we have seen the following four types of pop-up windows. Figure 6. Four examples of malicious pop-up windows From what we have seen, if a button on the “PC Speed Test” or “PC Performer Test” pop-up window is clicked, the browser redirects the user to a security risk download site. The “how fast can you build your muscle mass?” pop-up window looks like an advertisement, but at the time of writing nothing happens if the button is clicked. We have only seen the “captcha” pop-up window in one attack and we have not yet analyzed it to see what it does. Evidently the purpose of these attacks is to make money for the attacker by selling security risks and clicking advertisements. To stay safe, please ensure that your computer has the latest software patches installed and always keep your antivirus definitions up-to-date.
<urn:uuid:38586cef-5f2a-4932-afda-5371acf1d939>
CC-MAIN-2022-40
https://community.broadcom.com/symantecenterprise/communities/community-home/librarydocuments/viewdocument?DocumentKey=d48dfc93-7997-4cc6-abb6-41ba1513a996&CommunityKey=1ecf5f55-9545-44d6-b0f4-4e4a7f5f5e68&tab=librarydocuments
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337723.23/warc/CC-MAIN-20221006025949-20221006055949-00480.warc.gz
en
0.904893
958
3.09375
3
Georgia’s massive Internet victory last week is truly a game changer when it comes to the Internet and communication, especially for local service in remote areas that often have poor to no Internet service. According to The Wall Street Journal, a group of Georgia mayors, county representatives, and activists banded together to convince state lawmakers to defeat a telecom-backed bill that would have prevented municipalities from creating their own broadband networks. This means that state-run municipalities will be able to build their own Internet service and offer it to the local citizens. This creates cheaper prices and better Internet access to rural areas. Remote areas with better Internet means better meansof communication, especially with phone services like VoIP (Voice over Internet Protocol). VoIP is Internet calling and is cheaper than landline calling and offers more features like IP faxing and IP video conferencing. The spread of Internet access to rural areas is especially important as the imminent death of the PSTN (the public switch telephone network) creeps closer. Experts estimate that the PSTN, which works as the nationwide landline phone service infrastructure, will become obsolete by 2018. Without landlines, users will be left with the options of expensive cell phone service, or the cheaper VoIP alternative. Needless to say, this Georgia victory might very well spread to other counties that have low Internet availability. According to a study done by PEW, roughly one-third of Americans who make less than $30,000 don’t have access to broadband Internet. If the local counties decide to create their own communication service, these places will not only have access to the Internet, but will also have access to cheaper phone service, cheaper international service, and better work tools for remote working. VoIP is the modern means of cheaper communication. A quarter of the US population already use VoIP, with this number growing every year. Many of these VoIP users are downloading VoIP apps to their smartphones, which allows them to take their cheap phone plans wherever they go. Cities with better local networks will make it easier for mobile VoIP users to stay connected at home, at work, or on the go. Other states and cities beyond Georgia are leading the way for cheap, if not free, Internet access. For instance, Google has recently worked with Kansas City, Kansas and Palo Alto, California to bring free WiFi throughout the two cities to test out a larger public WiFi service. High tech cities with cheap VoIP access is the wave of the future, and legal successes like the one in Georgia prove that the need for better service will win out over the interests of the restrictive and expensive competition. Jennifer Cuellar writes about technology, VoIP, and phone service developments for MyVoIPProvider.com
<urn:uuid:b020f1d8-383e-4057-b84e-0120b582530f>
CC-MAIN-2022-40
https://www.myvoipprovider.com/blog/georgia-internet-service-voip
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337723.23/warc/CC-MAIN-20221006025949-20221006055949-00480.warc.gz
en
0.941895
558
2.65625
3
Failover is a backup operational mode in which a role (system function) is automatically transferred from its primary server to a secondary server that is on standby. This transfer between servers occurs only if the primary server becomes unavailable, either through failure or through scheduled downtime. Role failover is managed by the Directory role. How role failover works in Security Center - Primary server - Server that normally hosts a role for it to work on the system. - Secondary server - Servers on standby that are assigned to a role to keep it running in case the primary server becomes unavailable. There is no limit to the number of secondary (or standby) servers you can assign to most roles. However, the more servers you add, the less cost-effective it might be for you. Before failover, a role is hosted on the primary server and connects to a database server hosted on a third computer. When the primary server fails, the role automatically fails over to the secondary server and reconnects to the same database server.
<urn:uuid:9029772d-6933-4bd7-80c5-4719a4879d63>
CC-MAIN-2022-40
https://techdocs.genetec.com/r/en-US/Security-Center-Administrator-Guide-5.9/Role-failover
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334579.46/warc/CC-MAIN-20220925132046-20220925162046-00680.warc.gz
en
0.936989
209
2.75
3
Ransomware operators are willing to notify corrupt traders of cyberattacks so that they could short companies’ stocks before the attacks are made public. Financial traders live and breathe by the information they’re able to digest. Indeed, the USP of fintech companies is that they use technologies such as AI to help traders make sense of the huge quantities of data that can influence the movement of a stock. Most of the time, these kinds of companies are harvesting publicly available information, with their value being in helping you to make sense of it all. Of course, often the most valuable information is not in the public domain at all. Research from the University of Technology Sydney Business School suggests that roughly four times as much insider trading takes place than is actually caught by regulators. The researchers explain that the US Securities and Exchange Commission generally prosecutes around 50 cases of insider trading each year, but that this is the tip of the iceberg, with insider trading occurring in around 20% of all mergers and acquisitions, and 5% of all quarterly earnings announcements. Of course, one group of people with potentially highly valuable inside information is cyber criminals, and earlier this year evidence emerged that the hackers behind the Darkside ransomware were looking to partner with equally crooked traders to give notification of the attacks before the victims managed to release the information to the public. The hackers reasoned that once news of the hacks was released into the public domain, the share price of the victim would fall, thus providing an opportunity for traders to short the stock ahead of the official announcement and make an easy killing. This is a reasonable assumption to make. After all, the Equifax hack in 2017 resulted in a fall of around 30% in the company's share price. Similarly, customers affected by the SolarWinds attack saw their share price fall by up to 20%. Research from Edinburgh Napier University found that things are not always so straightforward, however, and that for around 40% of businesses, their share price isn't affected by disclosure of a breach at all. This doesn't appear to be deterring the Darkside team, however, who stated that they were willing to notify corrupt traders of attacks so that they could short the company's stock before the attack was made public knowledge. They're believed to be the first cybercriminals to explicitly use this as a formal means of making money from each attack. What's more, the mere threat of leaking the attack to traders could be used to increase the pressure on companies to pay the ransom demand as quickly as possible to minimize not only any negative press but also any financial hit they may take in the markets. Shady business models Of course, as I wrote in an article last year, it's by no means certain that a cyberattack will result in a company's share price going down, which is the essence of the shorting business model. Indeed, research from MIT highlights various strategies companies can employ to ensure that doesn't happen. The paper finds that the reaction of the markets is often a reflection of how professionally the victims approach the attack. Far and away the worst strategy is to pretend the attack didn't happen, with the taking of minimal remedial actions not far behind. Indeed, in another article published last year, I argued that reporting attacks early is usually the best approach, as this not only shows that you are taking matters seriously, but it also significantly helps other organizations in your industry and in your ecosystem to take any remedial actions to secure their own systems. Interestingly, the presence of more legitimate businesses may also work to inhibit the actions of the cybercriminals themselves. In a paper from Harvard, researchers argue that the presence of short sellers in a market tends to discourage those with actual inside information from selling their stock. The authors suggest that without short sellers in the market, those with inside information will feel confident that they will be able to capitalize on the information they have about the potentially value-destroying event. When short sellers have potentially got access to the same information, however, then they offer competition to the insiders and therefore increase the likelihood that the information will be revealed to the market. So, it seems that all roads ultimately lead to earlier disclosure of any cyberattack on an organization, which is probably a good thing for all of those on the legitimate side of the fence, even if less beneficial for those seeking to further capitalize on each attack.
<urn:uuid:a2fde6cd-4d6e-4dcf-99c5-eb6d4d9522a6>
CC-MAIN-2022-40
https://cybernews.com/security/cybercriminals-turn-to-short-selling-to-further-capitalize-on-their-attacks/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334942.88/warc/CC-MAIN-20220926211042-20220927001042-00680.warc.gz
en
0.967911
904
2.515625
3
“Espionage” – a term one may associate closely with James Bond. However, espionage is more than action-filled stunt sequences. With today’s increasingly advanced technology, computers and software systems are used as a tool for spies to hack their way to sensitive information. In a recent cyberattack, hackers attempted to steal sensitive data from several key sectors, including United States defense contractors. Over the course of several weeks, foreign hackers targeted the US and other countries, attacking the defense and energy sectors among other important areas. While the goal of this attack does not appear to have been related to the disruption of activity, the true purpose may be related to espionage. These hackers scan the Internet for vulnerable software and pick out targets from the vulnerable pool that may grant them access to sensitive data and information, as well as intellectual property valuable to competitors. In this instance, hackers stole passwords with the intention of maintaining long-term access. Once access has been gained, hackers then dig deeper into the system, looking for pieces of information and communication including emails. This type of cyber-surveillance poses a serious threat to businesses and governments alike. Security agencies with bipartisan support have worked to spread public awareness and warn similarly vulnerable corporations in an attempt to combat the rise of cyber espionage attacks of this nature. These prominent cyber-spying attacks serve as a potent and timely reminder of the importance of cybersecurity for organizations at every level. The damage from these attacks can be catastrophic and irreparable, no matter what sector or industry an organization is involved in. Sensitive data in the wrong hands can be used to significantly disrupt the lives of everyday citizens. You can protect yourself. At Blue Ridge Networks, we know how important cybersecurity is to ensuring the safety of your private and sensitive information. Our unique, patented approach means that breaches are prevented before they have a chance to occur. We offer a wide variety of cybersecurity solutions perfectly tailored to the exact needs of your organization, business, or enterprise.
<urn:uuid:a62d0a14-23a4-42cc-9a91-b8e8304d1234>
CC-MAIN-2022-40
https://www.blueridgenetworks.com/cyber-spying/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335124.77/warc/CC-MAIN-20220928051515-20220928081515-00680.warc.gz
en
0.941383
403
2.515625
3
Malware is any malicious software purposefully designed to cause harm to a computer, human, service, or network. There are a wide range of malware, including adware, worms, Trojans, malware, spam, and spyware. Any computer that connects to the Internet and has an active web browser is at risk for the invasion of malicious programs. These viruses make their entries into computers through freeware, shareware, or other applications that may have been downloaded without the knowledge or permission of the end user. Once the program has been installed, the operator can install additional malicious software that will propagate and continue spreading rapidly. In some cases, malware that spreads through a computer system can be removed with a clean up tool that is purchased. The key is not to panic and to be diligent. In most cases, users can regain control of their computer systems after they use the appropriate removal tools. Malware can be divided further into two categories: Malicious-ware and Antimalware. Malicious software includes software such as spyware, adware, and viruses that are used to hack into computers and interfere with computer processes and information. They are sometimes used to monitor Internet activity and send information back to outside sources. They can be used by criminal hackers to obtain credit card numbers, passwords, and other personal information. Other malicious software transmit or spread through networks and can infect other computers. They may spread through email attachments or files, and in some cases they may install viruses onto machines. Many of these programs were developed by cybercriminals looking to monetize websites that contain harmful content. Other types of malware transmit or are created by companies that develop the software. One type of malware that is commonly associated with spam and online attacks is malware that executes when you run certain types of programs. For example, if you install a game on your Facebook page or an app on your Blackberry, chances are good that some form of malware might be placed on your device. It is possible for cybercriminals to trick you into downloading a piece of software by posting a fake application or sending you a fake email attachment. As long as you trust the website or app in question, it is relatively easy for them to install malware on your devices. However, there is also a separate category of malware that can install on mobile devices. This is referred to as “adware” or “spyware.” These malicious programs are designed to monitor the behavior of the user and send information back to the authors or attackers of the malware. Most malware that infections mobile device functions in the same way as the software installed on a personal computer or laptop. There are different types of malware that can infect your device, but most of them operate in similar ways. Examples of software that can be considered malware include Trojans, worms, and viruses. Malware can be created by any person or group that has access to your network, including hackers, telemarketers, and even government agencies. While some malicious functions may have been developed by government agencies, many of the viruses and worms being distributed online have been developed by regular people using the internet. Fortunately, the vast majority of malware and other virus programs are quickly removed by security vendors. There are several common ways that Malware is spread on the internet. Common methods include downloading files from p2p networks, executing unknown software applications, and opening email attachments that are infected. Some Malware programs are also able to bypass firewalls and enter networks. - Some malicious software programs are able to bypass network protocols and communicate back to attackers in the host system. - Once Malware is loaded onto a computer or a network, the attacker can use it to perform additional malicious activities including stealing financial or personal information. - The most common victims of Malware include Online casinos, Online shopping sites and Internet email accounts.
<urn:uuid:a8e87684-93fa-42d4-83b2-7dc978b4965b>
CC-MAIN-2022-40
https://www.ekoob.com/2021/04/05/advantages-of-using-it-support-in-malware-attacks/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337529.69/warc/CC-MAIN-20221004215917-20221005005917-00680.warc.gz
en
0.946857
785
3.3125
3
North Carolina's Transportation and Water Quality departments use SAS Analytics to analyze huge volumes of geographic data to build roads faster, reduce costly land surveys and minimize environmental impact. North Carolina’s Transportation Department is using analytics software to build roads faster and for less money while minimizing environmental disruption. NCDOT is analyzing geographic data to help narrow the choices of possible road corridors and, at the same time, is able to reduce costly land surveys. The process can save $500,000 per project and shave 20 percent off the time needed to select and plan a road, according to North Carolina officials. NCDOT and the North Carolina Division of Water Quality are working collaboratively on the project, using SAS Analytics as the engine to analyze volumes of geographic data. How data wizardry can revive America’s cities Planners need to protect water sources as they assess transportation projects. As a result, they have to propose ways to avoid or minimize environmental impacts. Road builders often verify geographical surveys by sending surveyors and water quality experts into the field to document streams and wetlands. However, a large project, such as a bypass, requires substantial time and manpower to survey thousands of acres in order to identify environmental issues. “You might have hundreds of possible combinations for one road,’’ said Morgan Weatherford, environmental program consultant with NCDOT’s Natural Environmental Section. “It’s a major challenge to comply with federal and state environmental regulations in a manner that is beneficial for the environment and taxpayers as well.” A new data source has emerged in recent years with the potential to eliminate some costly field work. Light detection and ranging (LIDAR), which uses laser pulses to record the distance between two points, is useful for charting land elevation, which is key to locating wetlands and streams. LIDAR data is used extensively to update flood maps and is considered more detailed than geological survey information. However, LIDAR produces large volumes of data. One transportation project might involve upward of 30 million records with 30 different attributes per record. Prior to using SAS analytic software, nobody at NCDOT had used LIDAR data to predict stream and wetland locations for construction planning purposes. The NC Division of Water Quality built models to predict headwater streams and tested the accuracy with field surveys.“The models were 85 to 95 percent accurate, depending on the terrain,’’ said Periann Russell, environmental senior specialist with the NC Division of Water Quality. NCDOT also used LIDAR data for a much larger project that includes predicting stream and wetland locations for an entire county. The data will help transportation planners choose a corridor for a 20-mile bypass. It will also be used on several bridge modernization projects and be available to private developers for their proposed developments, NC officials said. NEXT STORY: Va. county puts all land data into one site
<urn:uuid:96db9a98-375c-49d4-be97-0f027adaa24a>
CC-MAIN-2022-40
https://gcn.com/data-analytics/2011/11/state-uses-analytics-to-build-roads-faster-cheaper/282050/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337731.82/warc/CC-MAIN-20221006061224-20221006091224-00680.warc.gz
en
0.936032
600
3
3
The battle against counterfeit drugs is one that is of high stake for the pharmaceutical industry. However, the fight against counterfeit drugs is a complex and daunting challenge. It not only involves huge losses for pharmaceutical companies but also poses dangers to those taking counterfeit drugs. The gravity of the situation can be understood by the fact that approximately 10% of the world’s pharmaceutical products could be counterfeit, as per the estimate of the World Health Organisation (WHO). This percentage could be much higher, about 30%, for developing countries, according to WHO. Further, WHO estimates that approximately 1 million people die every year due to the intake of counterfeit medicines, which is quite alarming. The proliferation of counterfeit drugs can be attributed to a large extent to the complexity of the pharma supply chains with additional touch points, which makes it possible for counterfeit drugs to infiltrate supply chain operations. Many times, counterfeit drugs are manufactured in multiple locations. For example, the chemical synthesis of products may take place in one country, while the addition of filler can be carried out in another country. This makes it increasingly difficult to track counterfeit medicines in the supply chain. Moreover, counterfeiters are adopting sophisticated technology for manufacturing and packaging their products, which makes it all the more difficult to identify such products. However, the pharmaceutical industry is also adopting sophisticated technologies to identify, secure, and track products, and anti-counterfeit packaging is playing a crucial role in it. Some of the innovative anti-counterfeit packaging employed by the pharmaceutical industry are, holograms, RFID, 2D codes, micro-text and microscopic tagging, advanced barcodes, serialization, embedded images, digital watermarks, laser codes, and invisible printing. Holograms is one of the most widely used methods, as these can be identified easily. This method can be made more effective with nano-text and hidden images, which can be added to increase security features of holographic images. Such holographic images are almost impossible to replicate by employing traditional printing methods. Colour-shift inks are used widely for anti-counterfeit packaging. Here, different colour combinations are used in a way that specific colours can be detected when viewed from different angles. Covert technologies like microtext and microscopic tagging are very effective, as they are invisible and require special detection devices to identify. For example, in UV micro-text print, the text characters used are invisible to the naked eye. Another latest technology to be explored by the pharmaceutical companies is RFID (Radio Frequency Identification Devices) with tags. Pharma giants like Pfizer and GlaxoSmithKline have already introduced RFID tags on the packaging of some of its drugs. The development of 2D barcodes has also helped companies to tackle the problem of counterfeit products. Thousands of characters and scanning techniques can be accommodated by 2D barcodes. Tackling counterfeit products requires a multi-pronged approach, in which anti-counterfeit packaging plays an important role. To provide a higher level of security, many companies are combining the overt (e.g., holograms, security graphs, colour-shift inks) and covert technologies (micro-text and microscopic tagging, laser codes, digital watermarks, embedded images). Security design tapes and security tear tapes are examples of combining these two technologies to provide better anti-counterfeit packaging. In the future, we can expect more robust and counterfeit-proof packaging for pharmaceutical products. Nanotechnology is one such technology that has huge potential for improving the quality of pharmaceutical packaging. Forensic anti-counterfeit packaging is another emerging technology. With growing investment in anti-counterfeit packaging technologies, we can expect to see more innovative packaging techniques in the coming days.
<urn:uuid:f9a3a002-3991-4fe9-a81c-44b4a0d89fbe>
CC-MAIN-2022-40
https://www.alltheresearch.com/blog/anti-counterfeit-packaging-tackling-counterfeiting-pharmaceuticals
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337731.82/warc/CC-MAIN-20221006061224-20221006091224-00680.warc.gz
en
0.937847
758
3.109375
3
Sometimes readers ask us how to send an anonymous email or how criminals and scammers manage to send anonymous emails. Since this is not an easy question to answer, because, for starters, there are several ways to interpret the question, I’ll try to give you some information here. Interpret the question Sending an anonymous letter via snail-mail was easy. You forgot to add the sender address, filled out a false one, and if there was a chance the receiver could recognize you by your handwriting you used newspaper clippings to construct the sentences. And snail mail had the advantage that you could drop your message in a mailbox that gathered mails from various senders before starting the delivery process. So, not even the carrier had any way to identify the sender. The place of origin is hidden except maybe roughly by looking at the post stamp to see from what postal district the letter came. Unless the sender went through the trouble of driving halfway across the country to post the letter. As you can see there are a few sides even to this low-tech version of an anonymous mail: - No sender address - False sender address - Masking the content > encryption - Origin masking What is a spoofed email? Since sending an email without a sender address can result in errors and will certainly raise suspicion, it is easier to spoof a sender address. Spoofing is sending an email with a false sender address. Spoofing an address is relatively simple since the Simple Mail Transfer Protocol (SMTP) does not check the information in the “From”, “Reply to”, or “Sender” fields. The only reason it is possible to track back an email with a spoofed address is because the email headers will include the sending IP address. So, to pull off a completely untraceable spoofed email the sender will have to use a VPN to mask their IP address or use a compromised system to send the emails from. Compromised servers are popular with people running malicious email campaigns. How can I send and receive an encrypted email? A very different concern is to hide the content of an email from anyone except the intended receiver. This requires some type of encryption that only the receiver can decrypt. Encrypting emails like this—known as end-to-end encryption—has historically been difficult, although the tools for achieving this kind of encryption are getting better and easier to use. Most emails are encrypted during transmission, but they are stored in clear text when they are at rest, making them readable by third parties such as email providers. But there are some providers that provide end-to-end encryption and zero access encryption to secure emails. This means even the service provider cannot decrypt and read your emails. If you want to have full control and not depend on a provider you will need to exchange public keys with the parties that you want to start encrypted communications with. Once you have exchanged keys, most email clients will offer you the option to encrypt emails on a per-message basis. How can I send an email anonymously? I wrote a blog post on how to send encrypted mails a long time ago. Some things have become considerably easier since then. Some carriers offer you the option to send end-to-end encrypted email for free. Personally I have only tried Protonmail which allows you to come up with your own email address, and even the free version is free of advertisements. You only need to provide an existing email address if you want to use that as a recovery method in case you forget your credentials. If you do not need that option the sign-up procedure is completely anonymous. Is ProtonMail really anonymous? Protonmail is a secure email provider that does not solicit any information from you to use the free version, as long as you don’t chose to use the recovery option. For any legitimate use case Protonmail can be considered secure and private. This is considering that for any legitimate use cases it should be enough to send an encrypted email, so that the intended receiver is the only one that can read the content of the message. Protonmail can even be used in combination with a VPN so that even your IP address remains hidden. Unfortunately this also makes the service very popular amongst ransomware peddlers who sometimes create individual Protonmail accounts for every single victim. Can email be traced? Even hardened criminals make mistakes, so you should always be weary of the fact that an email you sent can be tracked back to you. On the other hand it is virtually impossible for anyone to trace back an email that was sent using all the techniques we have described above. As so often, it is wise to assume the worst possible scenario. We have seen script kiddies that thought they could use a Gmail account as a means to send anonymous emails. Maybe the receiver will not be able to trace it back, but the police certainly will, with some help from Google. If you need plausible deniability don’t put it in writing. For legitimate use we hope to have handed you some useful tips. I have received an anonymous or spoofed email. What should I do? How you deal with any mails you receive normally depends on the content. As with any email, it is advisable to scrutinize whether the email and the sender are who and what they claim to be. If you recognize the sender but don’t trust the content, contact the sender through other means to verify they sent it. Do not send read receipts or other confirmations that you have read the mail before you are sure you can trust the sender. You can find some tips on how to recognize and deal with unsolicited mail in this blogpost about recognizing and disposing of malicious emails and this article about phishing. If the mail has the character of an extortion email you may find our post describing what to do when you receive an extortion email helpful. Depending on where you live it may be prudent, or even mandatory, to inform the proper authorities about any extortion attempts.
<urn:uuid:15537ca0-1bd8-4db0-b627-0c538d079973>
CC-MAIN-2022-40
https://www.malwarebytes.com/blog/news/2021/07/how-to-send-an-anonymous-email
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335254.72/warc/CC-MAIN-20220928113848-20220928143848-00080.warc.gz
en
0.951742
1,245
2.890625
3
What Security Does a Firewall Provide? What is Firewall Security? You may be familiar with the term “fire wall” as a wall inside a building that is made specifically to protect something from fire. In IT, firewall security fulfills a similar purpose, except the fire is the myriad cyberthreats trying to break through your cybersecurity. Implemented correctly, firewall security can even protect two different parts of your network from each other. There are two types of network security firewalls: hardware firewall and software firewall. While most people have used a software firewall, which is a program that is installed on a single workstation that protects the machine from within the network, your primary defense should actually be a hardware firewall. Like the fire wall that protects your office building, this is a physical device that is programmed to stop a threat before it even gets into the network. A hardware firewall protects not just computers, but printers, servers, and any other device on the network. It is constantly on guard against intruders and puts up a good defense by acting as a gatekeeper, blocking unauthorized access to the network. When a hardware firewall is configured correctly, input from your IT provider informs the hardware what is safe and what is not, and it combats the threats while allowing the authorized users to access what they need. Enterprise-grade firewalls permit each user account to be configured and managed, allowing for specific securities per user. This provides different levels of access to confidential data and tells your firewall how to act in regards to usage. While a hardware firewall is meant to protect your network from the outside, it’s always best to have a layered security system, just in case one element fails. This is where a software firewall comes into play. A software firewall is an application on your computer or server that alerts the user or IT manager to any problem, if a security risk gets through the hardware firewall. This is what many people associate the term firewall with and would recognize on their computers. Windows Defender Firewall, or any firewall program that comes with an operating system is considered a software firewall. In combinations with other security measures, an enterprise-grade firewall can help your business fulfill security standards §164.308(a)(4) Information Access Management and §164.312(a)(1) Access Control. It also provides traffic logs that can help you audit (§164.312(b)) your network activity or find the source of a breach should one occur (§164.308(a)(6)). Why is a Hardware Firewall Important? The problem with having a software firewall without a hardware firewall is that by the time it has been flagged by the software firewall, the attack is already inside your network. Think of a hardware firewall as a security guard that keeps people from entering the building, while a software firewall is the alarm system that goes off only when the door opens. By the time the software firewall is activated, the hacker or malware has full access to your system until the problem is resolved. It’s not enough to simply have hardware and software firewalls, though. Both need to be continually updated, patched, and managed to handle the constantly evolving and increasingly sophisticated cyberthreats. How do Firewalls Help HIPAA Compliance? User- or Role-defined authorization and authentication are crucial aspects of HIPAA compliance, as well as newer cybersecurity architecture, such as the Zero Trust model. Having the hardware capable of performing these type of security checks helps ensure that the only people accessing your data are your employees, and then, only the employees who need the access. An enterprise-grade firewall is part of a thorough and complete HIPAA compliancy program. Having a software firewall without a hardware firewall is not secure. Your business has sensitive, irreplaceable data, and protecting it from attack and unauthorized access is critical. Malicious actors can find entry points, or your employees could unwittingly download a virus, ransomware, or other malware, which can jeopardize your company for days or weeks. Unauthorized access to your network can also mean serious downtime from cyber attacks, halting business function and doing permanent damage to your reputation and the relationship you have with your clients. For many industries, a breach means automatically reporting to the government and coping with the consequences. A lot of businesses worry that a firewall will impede their employees’ ability to do their jobs by slowing the network. If a firewall is poorly implemented, this could be a valid concern. Firewalls installed with limited configuration and the manufacturer’s defaults abound with cybersecurity vulnerabilities, unnecessary exposure, and business risks—and with it comes frustration from a slow “bottlenecked” connection. Firewalls should be tailored to your particular business environment to provide maximum security with optimal functionality. A well-configured firewall combines security with productivity without sacrificing one for the other. Firewall security is essential and acts as a barrier between your network and a variety of threats, including the threat of human error. Anderson Technologies emphasizes employee education for the clients we serve, but mistakes happen no matter how well-versed in cybersecurity your employees may be. One errant click can instantly create a vulnerability in your network, but firewalls provide a safety net. Don’t put your company at risk by thinking a firewall won’t be efficient. Work with your IT company to make sure you’re protected from every angle. By Founding Principal Amy O. Anderson Take a Look at the Small Business Monthly Issue Thank you! After being featured
<urn:uuid:4ce6fc12-5729-46bb-abe5-7377af0c0ce8>
CC-MAIN-2022-40
https://andersontech.com/learn/what-does-a-firewall-do-for-a-network/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335365.63/warc/CC-MAIN-20220929194230-20220929224230-00080.warc.gz
en
0.934815
1,136
3.15625
3
Government agencies such as the Canadian Center for Cybersecurity (CCCS) and the National Institute of Standards and Technology (NIST) in the U.S., not to mention cybersecurity insurance companies, all stress or mandate the need for cybersecurity awareness training. The reason for this need/requirement is very concrete. In addition to cybercriminals exploiting vulnerabilities in the technology organizations use, they also prey on and exploit people’s trust, behaviours and emotions to gain access to their systems. A cybersecurity awareness training program is a cost-efficient way to teach employees to recognize, avoid and report threats, which helps reduce cyber risk for an organization and creates an effective human firewall. Here are some best practices that can be used as a base when looking at implementing a cybersecurity awareness program within a business. Cybersecurity Awareness Training Best Practices 1. Emphasize that cybersecurity is not just about technology Technology is important to a strong cybersecurity strategy. Properly configured firewalls, endpoint protection, patching and updating and many other technologies that help reduce cyber risk are essential. However, one click on a malicious link can bring an organization to a halt. According to 2021 Data Breach Investigations Report, 85% of all data breaches involve human element. It’s important that everyone in the organization understands their importance and their responsibility when it comes to cybersecurity and reducing cyber risk. 2. Deliver training often and make the content easy to digest Delivering highly engaging content that employees can used right away at work or at home has produced better results than longer training sessions. Training delivered on a regular basis throughout the year and in bite-sized portions so people can complete them in short periods of time has proven to be highly effective in raising employee cybersecurity awareness. 3. Simulate and gamify Making the education program as real as possible and focusing on phishing simulations that mimic real-life attacks will reinforce the policies and procedures being taught. Delivering these simulated attacks to an employee’s desk allows for an organization to see how they react in their normal work environment. Depending on their responses, very targeted additional training can be delivered to address any vulnerable areas quickly. 4. Don’t punish employees Cybersecurity is an ever-changing landscape as cybercriminals hone their craft and sophistication levels increase. The training that is provided is meant to give employees a safe space to learn and to fail. Punishing employees that don’t perform well on the training is not the right approach. Understanding the areas that need reinforcement and providing additional training will yield the results organizations should be looking for. Testing is a key part of a cybersecurity awareness training initiative. It allows organizations to determine the knowledge level of each person within the business, including executives, to provide the appropriate training for each individual. Cybersecurity awareness training can help businesses stop many attacks by arming their employees with the knowledge they need to act as your best line of defense against cyber risks. MicroAge can help you implement a cybersecurity awareness training solution that is right for your organization. Contact us today.
<urn:uuid:6c4da5c0-0e16-424a-9436-ab34ff543cf7>
CC-MAIN-2022-40
https://microage.ca/thunderbay/best-practices-for-cybersecurity-awareness-training-programs/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337537.25/warc/CC-MAIN-20221005042446-20221005072446-00080.warc.gz
en
0.949142
628
2.78125
3
If you think your email is safe from hackers, think again. A lack of sufficient email security protocols can lead to data theft, unauthorized access to sensitive information, and successful malware attacks. Here are some tips to secure your email account from unwanted intruders and the many troubles that come with them. A lot of people get a handful of spam in their email inboxes every day. While spam can be a nuisance, it only takes a few minutes to delete or block spam. But if you receive tens of thousands of spam all at the same time, a huge chunk of your time and energy will be wasted on dealing with them — and they might actually be hiding telltale signs that you're being attacked by cybercriminals. Every day, you receive dozens of email messages, including spam, which is usually harmless, unless there’s a malware attached to it. But even the harmless ones can become more than just an annoyance if you have thousands of them flooding your inbox. Everyone uses email as the central hub for their personal internet activities, but this also makes it an attractive target for cyberattacks. The importance of email security is vital to your company’s survival, so applying these simple tips can dramatically reduce your exposure to hackers and malware. Cybercriminals are fairly experienced at avoiding detection. By the time you notice they’ve infected your computer with malware or hijacked your account, serious damage has most likely already been done. To make matters worse, they have another way to hide their illegal activities, and it involves sending thousands of spam emails.
<urn:uuid:57a05d11-fe84-4978-891c-1d23da4a0719>
CC-MAIN-2022-40
https://www.datatel360.com/tag/email-security/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030338280.51/warc/CC-MAIN-20221007210452-20221008000452-00080.warc.gz
en
0.961931
318
2.578125
3
Cryptanalysis is the decryption and analysis of codes, ciphers or encrypted text. Cryptanalysis uses mathematical formulas to search for algorithm vulnerabilities, attempting to break into a cryptographic system. The goal of cryptanalysis is to find weaknesses in or otherwise defeat encryption algorithms. This research is used by cryptographers to improve and strengthen or else replace irreparably flawed algorithms. An example of cryptanalysis improving an encryption protocol os Transport Layer Security (aka: TLS). It has been tested and broken and fixed and upgraded from version 1.0 to 1.1 and between 2017 and 2019 to version 1.2. The older versions are no longer recommended because of their inherent flaws; however, the current version is thought to be very secure and well trusted. It’s important to have this research done to our algorithms in the hopes of staying one step ahead of hackers who themselves are improving their techniques on cracking our encryption protocols. Additional Reading: Outdated Encryption: Is Your Data Secure?
<urn:uuid:977688d3-7a52-4c21-afe9-78cedac51c33>
CC-MAIN-2022-40
https://cyberhoot.com/cybrary/cryptanalysis/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030331677.90/warc/CC-MAIN-20220924151538-20220924181538-00281.warc.gz
en
0.930529
199
3.296875
3
Although the threat of cyber espionage has largely been considered a recent phenomenon, foreign spies have been targeting the networks of rival governments for years. The two main goals of these operatives is to steal valuable state secrets, technology research and other valuable information and to disrupt operations at critical installations. Both of these endeavors place universities and other academic institutions firmly in the sights of state-backed hackers. System administrators at these schools should stay apprised of emerging threats created by cyberspies to prevent their networks from becoming compromised and sensitive information from being stolen. Cybersecurity researchers recently identified an espionage campaign conducted by approximately 50 individuals that has been in operation for the past eight years. In that time, the group has successfully breached more than 350 targets across 40 countries. The group has traditionally prioritized government and diplomatic targets, but have also attacked research centers and universities. Recently, the criminal syndicate has been observed allocating more resources to breach assaults against institutions that may contain valuable information regarding advanced technology. Some of the organization's areas of interest include nanotechnology, energy production and space exploration. These fields have been extensively studied at research labs across numerous universities, meaning school networks may be targeted with greater frequency moving forward. Spear-phishing: the weapon of choice The syndicate's preferred malware tool, dubbed NetTraveler, utilizes spear-phishing email campaigns to attach malicious files in the guise of Microsoft Office documents. Once a user's system has been compromised, the malware tool begins gathering information including documents, sensitive data and key logs. Researchers determined that the amount of stolen data contained on command and control servers was upwards of 22 gigabytes, however, this did not include files that been wiped from storage over the years. The cybersecurity research team noted that some of the victims identified in their study had previously been targeted by the famous Red October cyber espionage campaign, which managed to operate undetected for five years. With hackers becoming more adept at concealing their activities, universities and other institutions can no longer rely on government agencies to identify cybercriminal syndicates before they launch widespread campaigns. Comprehensive measures are needed to protect critical networks, including the deployment of layered security protocols. One component of a healthy cybersecurity platform is the use of application control technology. If IT administrators are concerned about malicious programs – such as those sent through the NetTraveler campaign – launching on their systems, they can use whitelisting tools to prevent anything except predetermined software from running on a computer. This can prevent hackers from compromising academic networks and stealing vital information.
<urn:uuid:86f414d8-f1b0-4433-ae2c-9837be47d241>
CC-MAIN-2022-40
https://www.faronics.com/news/blog/cyber-espionage-campaign-targets-governments-academic-institutions
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334620.49/warc/CC-MAIN-20220925225000-20220926015000-00281.warc.gz
en
0.961655
507
2.953125
3
Overview of Security Analyst Jobs Jump to>> The 20 Best Online Bachelors in Cyber Security Degree Programs Security Analysts are by far the most common job title of cyber security professionals. And one of the most important. Daily, security analysts probe the systems they are meant to protect to discover security weaknesses. Once the weaknesses are discovered, a security analyst will help to implement creative ways to prevent others from exploiting these weaknesses. Security Analyst Job Description: - Planning and implementing security measures for computers, networks, and data. - Staying up-to-date with new technologies, and leading efforts to implement new security technologies - Creating, implementing, and practicing disastear recovery and mitigaton plans - Helping with encryption, firewall, and other general security-related tasks in your organization - Establishing and training other non-cyber security staff on security protocols. Differences Between Security Analyst and Security Administrator Positions Security analysts and administrators often work in tandem, though their roles focus on different aspects of cyber security. Where a security analyst will help to discover security flaws and recommend changes to help tighten up cyber security, administrators take a more active role in implementing technical changes. Also, security administrators are the main party responsible for keeping a system going. The number one focus of many security administrator positions is to ensure system up time. While the main focus of security analysts is to ferret out security weaknesses and come up with solutions to these weaknesses. Educational Background Needed for Security Analyst Positions While there are those who have worked their way up through IT ranks and learned information assurance skills less formarly, most security analyst positions require a bachelors degree in computer science, programming, or information technology. Furthermore, some employers prefer that security analysts have more specialized undergraduate degrees such as bachelors in cyber security or information assurance. Depending on the seniority of the role, many employers also prefer technical certifications which coupled with any technical undergraduate degree can help to show competency in cyber security topics. For more information on these educational requirements, check out our ranking of top online bachelors degrees in cyber security, as well as our guide to the most lucrative and in-demand cyber security certifications. Technical Skills Needed for Security Analyst Positions While this skillset may vary from organization to organization, many of the most requested include the following: - Data and Traffic Analysis - Anti-virus and anti-malware advanced knowledge - Security configuration knowledge - Data Management knowledge - Computer networking, routing and switching knowledge - Commonly some combination of C, C++, C#, Java or PHP programming knowledge - Linux, Unix, and Windows operating system knowledge - Security information and event management - Wide ranging device configuration knowledge - Some incident response knowledge While that list may seem overwhelming to those looking for their first job as a security analyst, note that very few security analysts are experts in most of the above disciplines, and on-the-job learning is definitely part of the security analyst experience. Job Outlook and Salary Information for Security Analysts As with most cyber security roles, employment prospects for security analysts are great. According to the Bureau of Labor Statistics the number of open jobs in security analyst positions are set to increase by 18% nationwide by 2024. This signals that demand for security analysts is booming. With the scale of cyber threats quickly ramping up in recent years, it’s safe to say that demand could even greatly outpace the BLS estimate. Earnings for security analysts are a great indicator for cyber security job earnings in general, as security analyst is one of the most common positions in information assurance teams. According to the Bureau of Labor statistics, the median salary for a security analyst was $92,600 in May 2016. While entry level salaries in this role may be slightly lower, security analysts are often within the first five years of their cyber security career, signalling the ability to make a high salary fairly soon into your cyber security career. Security analyst positions are available in every branch of the armed forces, and go by the following names: - Army: cryptologic network warfare specialist and military intelligence systems maintainer - Navy: information warfrace and technology roles - Air Force: cyber systems operators - Marine Corps: information security or assurance technician As with many cyber security roles, obtaining a security clearance can be crucial if you’re looking for work in government agencies, trying to work as a government contractor, or work in an organization that works with government contractors. Security clearances can take from 3 months to one year from the start of the application process to obtain. So often the easiest way to not slow down employment prospects is to obtain a security clearance within a current cyber security role, instead of applying for jobs that prefer a security clearance when you don’t have one. One important note is that only naturalized US citizens may obtain security clearances to work with government agencies or as contractors.
<urn:uuid:f8bbcf8b-642a-4c4e-960f-a9a0b0aaa25a>
CC-MAIN-2022-40
https://www.cybersecuritydegrees.com/careers/security-analyst/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335573.50/warc/CC-MAIN-20221001070422-20221001100422-00281.warc.gz
en
0.929491
1,026
2.515625
3
Small businesses are at considerable risk of cyberattack. With fewer staff resources available to maintain and monitor networks and fewer defenses in place, small businesses are an attractive target for hackers and other criminals. The risks of a cyberattack are considerable. Companies could spend tens of thousands of dollars just to recover from an attack. There’s also the reputational and regulatory downsides if businesses allow data to be compromised, stolen or accessed by unauthorized parties. To help businesses better understand the risks and solutions, this guide serves as a comprehensive introduction to cyberattacks and how to prevent them. Defining Cybersecurity and Cyberattacks Cybersecurity is a collection of preventative measures designed to defend computers, data, mobile devices, servers, networks and users from malicious attacks. Cybersecurity generally focuses on issues that arise from internet-connected devices and systems. Cyberattacks are criminal activities designed to disrupt networks, prevent access to data, websites and systems, or extract valuable information. Some cybercrime is done solely to wreak havoc on websites and computer systems. Other activities are intentional attacks designed to steal and sell personal information, company secrets or intellectual property. Other cyberattacks are done for geopolitical reasons to disrupt governments, elections or economies. What Are the Types of Cyberthreats? Each day, there are new techniques, threats, codes, programs and approaches used to commit cybercrime. Businesses need a firm understanding of what these threat types are in order to choose the right defenses. Below are definitions of the most common threat types. - Malware. Malware is a portmanteau of “malicious” and “software.” It is an umbrella term that covers programs designed to cause harm once installed on a targeted computer or server. Malware includes: - Trojans – Programs that hide a virus or other malicious program. Many “free software” offers trick unsuspecting users into downloading and installing other apps that are actually Trojans. - Adware or Spyware – These programs can take over web browsers and redirect activity to other sites, usually e-commerce pages. These can be difficult to remove and frustrating, as they often use pop-up windows or pages that appear behind another. More nefarious adware can monitor which websites you visit or track keystrokes. - Viruses – Often spread by using removable devices on otherwise clean machines, visiting malicious pages or clicking on email attachments, viruses are designed to damage or destroy files stored on a network. - Worms – A kind of virus that replicates and propagates itself from computer to computer connected to a network. Typically, worms consume computing resources, slowing or stalling your machine. - Advanced Persistent Threats. Most hackers do not break into systems in one fell swoop. Instead, they stage their attack in phases. These advanced persistent threats (APTs), one embedded in a computer, may lie dormant for a period before they are activated. The program may also embed itself differently in different parts of the system, so if one threat is detected and eliminated, others can continue to cause harm. - DDoS. A distributed denial of service happens when outside players overload a server with requests for access or connection. Quickly a DDoS can shut down a target’s networks or websites. - Rootkits. Rootkits are a small piece of software that’s installed by another, larger software program or attacker that’s gained access. While rootkits may not be malicious, they can harbor malicious operations. Using rootkits, an attacker may be able to monitor activity, access information, change programs or use your computer to complete other actions. - Botnets. Botnets are networked, automated programs that can be controlled by one or many outside persons. Botnets are used to spread spam or viruses or to facilitate a DDoS. - Ransomware. A form of malware, ransomware infects your computer system and disables access or functionality, usually to your website. The cybercriminals usually demand a ransom in the form of bitcoin or other difficult-to-trace cyber currency in order to release control of your systems. If the ransom goes unpaid, the hackers usually threaten to release or sell information about your company, customers or employers. - Phishing. Phishing is a common way hackers steal personal information, logins and credit card information. Phishing attacks usually begin with an email urging a user to visit a known website. Once there, a user can unwittingly reveal information that compromises their identity or finances. - Fake Anti-Virus. An unsuspecting user may see an email from their anti-virus provider and install a fake update or upgrade. Instead, the installed software begins to make system modifications (including issuing false threat alerts) that make it difficult to eradicate the program. - Corrupted Files. Even common file types, such as those used for word processing, spreadsheets, images, videos and presentations, can be corrupted by malicious code that’s embedded in a seemingly legitimate file. These programs can execute malicious scripts that consume memory or bandwidth, add or delete files, allow attackers access to your computer or use your computer to attack other machines. - Zero-Day Attacks. When attackers unearth a security flaw or vulnerability in a piece of software before the developers do, it’s an incredible opportunity. Hackers exploit the vulnerability that can extract information or cause damage. The scope of a zero-day attack can be staggering as it can take months or years before the flaw is detected and a corrective patch is built and deployed. - Password Attacks. The more characters and choices your company requires of users when creating passwords, the less likely you’ll be to fall victim to a password attack. A brute-force attack occurs when a hacker uses programs or guesses until they are successful in gaining access. Other attacks include the use of dictionaries to find common words used as passwords and keylogging, which captures keystrokes to steal login IDs and passwords. - Email Spoofing. It is far too easy for hackers to create an email that looks to be from someone, even a known colleague or leader. However, these programs instead are imposters, looking to steal passwords or financial information. A recent series of spoofing emails asked employees at colleges and universities to purchase gift cards for their “colleague,” scratch off the protective coating and send the card authorization codes. If spoofing is used to contact your customers or partners, you may find them leaving to do business with someone else. - Insiders. There’s a good reason to limit the number of employees who have administrative privileges. Inside attacks are often conducted by employees with administrative rights. They use credentials to get access to confidential information and use it for their own benefit or sell it to the highest bidder. Other inside threats manifest themselves when companies fail to remove access rights or user accounts of employees who leave the organization. Cyberattacks can take many forms: targeting your users, your networks, your devices, your software or your websites. That’s one of the critical reasons why you need cybersecurity solutions that can address all of the potential threats. Who Commits Cybercrime? Many people imagine cyberattackers as hoodie-wearing, sunglasses-toting bad guys in their basement inflicting harm for personal gain or “fun.” While there is an element of truth in that stereotype, there are other players in the cyberattack realm. The criminals may be: - Single actors or groups looking to disrupt or steal - Political operatives looking to steal information for nefarious means - Nation-states authorizing attacks that cripple electronic systems and defenses, disrupt communications or infrastructure, or corrupt or steal data - Corporate actors or individuals looking to steal information that allows for an economic, strategic or military advantage While there may be many players involved in cybercrime, the common theme is that your business needs to be protected. Are These Threats Overblown? No. Consider some of these statistics from the Ponemon Institute’s global 2018 Cost of a Data Breach study: - The average data breach costs $3.86 million, a 6.4 percent increase from the previous year - The cost per stolen record is $146 - The mean time it takes to identify a data breach is 197 days - The mean time to contain a data breach is 59 days When looking at the impact of cyberattacks on small businesses, the 2018 HISCOX Small Business Cyber Risk Report notes: - 47 percent of small businesses had at least one cyberattack in the last year - 44 percent had two to four attacks - 67 percent of business owners and executives indicated they were concerned or very concerned about cyberthreats - The average cost to a small business for a cyberattack is $34,600 - Small businesses incur indirect costs from a cyberattack, including customer loss, difficulty in attracting new customers, brand damage, distraction and productivity loss and staff hours committed to a resolution Small businesses have other characteristics that contribute to their vulnerability, such as the lack of available budget and qualified personnel to manage the protective measures necessary. What Kinds of Cybersecurity Tools Are Available? The number and type of threats are constantly changing. That’s why small businesses need to be vigilant about how to address the potential attacks. The National Institute of Standards and Technology (NIST) recently released a helpful and clear framework that can help companies understand the layers and purposes of different cybersecurity solutions. The NIST framework lays out five cybersecurity functions: - Identify – Asset management, business environment, governance, risk assessment and risk management strategy - Protect – Access control, awareness and training, data security, security processes and procedures, maintenance and protective technology - Detect – Anomaly and event detection, continuous monitoring and detection processes - Respond – Response planning, communication, analysis, mitigation and response improvements - Recover – Recovery planning, improvements and communication Another way to look at the scope of solutions is to group them into categories based on the type of security they provide. These broad security categories of protective measures collectively provide your business with a comprehensive cybersecurity solution. The categories and the protections that fall within those groupings are outlined below. Operational Security. Processes and procedures for handling, storing and securing data, including user permissions, data locations and sharing guidelines. Cyberthreat Assessment. Companies looking to provide blanketed cybersecurity protection should first conduct a thorough assessment of their existing IT infrastructure. Managed IT providers usually begin their engagements with new customers with this deep dive, which provides a baseline understanding of the data, processes and protection in place and the vulnerabilities that need to be corrected. Security Policies and Practices. There needs to be a collection of well-defined and articulated policies and procedures that address what data is available, what its business function is, how and where it is stored and who can access it. These cybersecurity policies, and the consequences of not following them need to be spelled out for employees, taught and reinforced. Policies should also include the use of personal devices, peripheral devices, home computers, public WiFi and corporate credit cards or purchasing cards. Access Control. Small businesses should have clear guidelines in place for who determines access to files and servers. Administrative rights should be limited as noted above. Access should be requested and approved using a clearly delineated process that presents clear business reasons for allowing permission. Access should be provided on a need-to-know or need-to-use basis. Access policies should also consider physical access to servers, data centers, data closets, physical media and off-site locations. Regulatory Mandates. Many industries are subject to government agency or sector-based requirements for the storage and usage of data. These mandates are especially critical to those companies that collect or use personal health information, collect payment information, manage legal documents or do business with certain federal or state entities. Knowing these mandates is critical, as they inform the decisions about what solutions are used and what reporting is required to demonstrate or maintain compliance. Insurance. Purchasing cyberattack insurance helps to reduce the financial risk of a potential assault. Like with other insurance products, demonstrating the existence of protective measures may result in lower premiums. Information Security. Ensures data integrity and privacy when information is in transit or at rest. Data Backup. Backing up your data (and applications and operating systems) is essential to making sure it’s protected and accessible in the event of an attack or natural disaster. Data backups should be done for information stored in the cloud or physical servers. Backup services should use most advanced security measures and best practices, including backing up data in out-of-region locations, automating backup functions and scheduling backups at regular intervals. Encryption. Encrypting your data while it’s being backed up and when it is stored in the cloud or on-premises makes it that much more difficult for hackers to use the information, even if they can access it. Network Security. Secures your computer network from intruders, attackers and malware. Wireless Network Security. Protecting your wireless network is an absolute must. Next-generation firewalls secure your network’s perimeter, detecting, containing and destroying unwanted activity before it can cause significant harm. Network security should also include continuous monitoring with automated alerts if suspicious or unusual activity is detected. Wireless network segmentation and segregation can also prevent the co-mingling of use by employees, visitors and web visitors. Passwords and Authentication. Your small business needs password creation guidelines, multifactor authentication procedures, guest access policies that make it difficult for hackers to break in and access systems. Website Security. Maintain your website’s security with limited privileges to the web server and content management tools, multifactor authentication, aggressive password changes, application whitelisting, adequate resource availability, web firewalls and dynamic security scans. Application Security. Protects your software applications and the devices that use those apps. Hardware and Software Updating. Providers regularly distribute updates and patches that address important security issues, fix bugs and remove emergent vulnerabilities. The challenge for small businesses is making sure each connected computer, mobile device and server has the required updates installed. Automated updating is much easier with managed IT services and cloud-based applications and servers. Mobile Device Security. Mobile devices enable employees to access information and networks from anywhere and at any time. However, mobile devices are especially vulnerable, especially when employees are required to provide their own equipment. For example, employees may not install the necessary updates, access information on unsecured public WiFi networks, or have devices stolen or lost. Business Continuity and Disaster Recovery. Plans and policies to recover from a cyberattack and ensure business operations are up and running with little to no downtime. Business continuity is the overarching guidance, policies, methods and protocols that ensure that a business can continue operations during and after a natural or manmade event. Disaster recovery is a subset of business continuity, focused on the procedures used to restore systems, access data backups and deal with the incident at hand. End-User Education. Training and awareness programs for employees and other users. In many ways, education and awareness are the first lines of cyberdefense. Employees need to understand the scope of cyberthreats, how those threats appear and what they can do to defend themselves and the organization. Cybersecurity is an absolute must for small businesses today. Knowing the types of attacks and the available solutions is the best defense against criminals looking to do harm.
<urn:uuid:24487570-115c-47d4-a2a2-b8a207db09fa>
CC-MAIN-2022-40
https://integrisit.com/cybersecurity-basics-for-small-businesses/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337446.8/warc/CC-MAIN-20221003231906-20221004021906-00281.warc.gz
en
0.924641
3,189
3.203125
3
The Problem with Mobile Phones [source: ssd.eff] Mobile phones have become ubiquitous and basic communications tools—now used not only for phone calls, but also for accessing the Internet, sending text messages, and documenting the world. Unfortunately, mobile phones were not designed for privacy and security. Not only do they do a poor job of protecting your communications, they also expose you to new kinds of surveillance risks—especially location tracking. Most mobile phones give the user much less control than a personal desktop or laptop computer would; it’s harder to replace the operating system, harder to investigate malware attacks, harder to remove or replace undesirable bundled software, and harder to prevent parties like the mobile operator from monitoring how you use the device. What’s more, the device maker may declare your device obsolete and stop providing you with software updates, including security fixes; if this happens, you may not have anywhere else to turn for these fixes. Some of these problems can be addressed by using third-party privacy software—but some of them can’t. Here, we’ll describe some of the ways that phones can aid surveillance and undermine their users’ privacy. The deepest privacy threat from mobile phones—yet one that is often completely invisible—is the way that they announce your whereabouts all day (and all night) long through the signals they broadcast. There are at least four ways that an individual phone’s location can be tracked by others. For more, click here.
<urn:uuid:86624297-8bb1-4236-ad26-b37de52587c1>
CC-MAIN-2022-40
https://www.cirt.gov.bd/the-problem-with-mobile-phones-source-ssd-eff/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337446.8/warc/CC-MAIN-20221003231906-20221004021906-00281.warc.gz
en
0.926849
308
2.6875
3
Attacks on encrypted web services have become increasingly common in recent years, hence the importance of increasing the security of online data. Encryption is one of the most basic necessities in the security arsenal. It’s what makes it possible for banks to offer online banking and funds transfers, or for consumers to make purchases online using their credit or debit cards. It’s what protects the public’s online interaction with government agencies or health care providers. It should surprise no one, however, that encrypted services are prime targets of DDoS attacks. Such services enable access to a wealth of personal, confidential, and financial data. Identity thieves and cyber criminals can have a field day if they succeed in breaking web service encryption. But… how many cyber-criminals access encrypted web services? To answer this question, our partner Netscout, a leader in cyber security assurance and business intelligence solutions, conducted a study in which they explain , attacks targeting encrypted web services have become increasingly common in recent years. Among enterprise, government, and education (EGE) respondents, 53% of detected attacks targeted encrypted services at the application layer. And 42% of respondents experienced attacks targeting the TLS/SSL (Transport Layer Security/Secure Socket Layer) protocol governing client-server authentication and secure communications. Among service providers, the percentage seeing attacks targeting secure web services (HTTPS) rose significantly over the previous year, from 52% to 61%. With this data you will ask yourself: How do attacks on encrypted web services occur and how many types are there? DDoS attacks targeting encrypted services tend to fall into four categories: - Attacks that target the SSL/TLS negotiation, commonly known as the “handshake,” which determines how two parties to an internet connection will encrypt their communications. - Protocol or connection attacks against SSL service ports, which seek to exploit SSL vulnerabilities. - Volumetric attacks targeting SSL/TLS service ports, which overwhelm port capacity with high volume traffic floods. - Application-layer attacks against underlying service running over SSL/TLS. Attackers are unrelenting in their assaults on high-value encrypted targets. Given the critical nature of most encrypted applications and services, a single successful attack can have devastating consequences. The breadth, variety, and escalation of attacks on secure web services heightens the need for a multi-layered defensive posture, with capabilities to detect and mitigate the full range of today’s attack types. Up to this point they are all bad news, so it is necessary to look for solutions to prevent attacks against encrypted web services. A positive conclusion from the Netscout report is that both service providers and enterprises are recognizing that traditional firewalls and intrusion prevention systems are insufficient in confronting sophisticated DDoS attacks – particularly encrypted attacks targeting encrypted services. Encryption is essential, but cannot be relied upon on its own to thwart determined and sophisticated attackers. Operators and hosts of secure web services increasingly recognize the need for purpose-built Intelligent DDoS Mitigation Systems (IDMS) as the only effective option for mitigating DDoS attacks. Best practices call for a layered approach combining always-on, on-premise defenses with cloud-based mitigation capabilities that activate automatically based on the size and nature of the threat. At Ayscom we have the experience and support to help you defend the different areas of the network and web services from the different types of existing threats, as well as continually updating and verifying the protection against new threats that arise every day. If you would like more information about this service or others related to this subject, you can write to us at email@example.com and we will be happy to help you. Source: Netscout, Attacks on encrypted services.
<urn:uuid:344ace68-af08-4c2f-a55d-aa0d8a9b9917>
CC-MAIN-2022-40
https://www.ayscom.com/en/how-to-avoid-attacks-on-encrypted-web-services/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337853.66/warc/CC-MAIN-20221006155805-20221006185805-00281.warc.gz
en
0.934597
770
2.546875
3
Businesses are collecting more data than ever before, with IDC and Seagate projecting a 42.2% annual growth rate in enterprise data collection over the next four years. However, without a proper framework in place, a lot of this data ends up going unused or fails to reach the right destination. At a time when organizations are moving data to the cloud and experimenting with artificial intelligence and machine learning, the need for component data architecture processes is more critical than ever before. Data architecture is the framework organizations use to document all data assets and govern how these assets are stored, arranged, and integrated into other data systems. The implementation of standards, rules, and policies on data assets that flow into an organization can provide a clearer view of what is happening inside the business, enabling faster analysis and visualization of data, and permit all teams to make data-driven decisions without the need for a complete understanding of the data process. Modern data architecture practices are built for applications that require real-time processing and visualization by integrating various data pipelines into a unified platform that can be accessed by all those in the company. The use of container orchestration tools automates much of the operational effort to run workloads and services, and data fabric or mesh technologies are the newest data-as-a-platform layer that fills the gaps in unifying an organization’s data flows. Without data architecture, businesses often end up collecting a huge amount of unstructured data, which fails to achieve the goals set out by the company. This happens all too often, as organizations regularly take a ‘technology-first approach’ to data collection, according to McKinsey, and rush to implement new digital technologies without going through the necessary processes of designing data architecture and implementing infrastructure to support data and analytics at scale. This can lead to all sorts of issues further down the pipeline, including redundant and inconsistent data storage, in which an organization lacks a single source of truth and has similar datasets stored in numerous data silos. This lack of focus can also lead to development sprawl, where a single function requires multiple technologies to support it. As a business gets larger, these issues will be magnified and can lead to lower productivity, as more technologies inevitably leads to more potential blockages in the data pipelines, and new employees need to be taught how to use more technologies. Modern data architecture has many benefits to data leaders looking to implement the new wave of collection, analytics, and visualization technologies into their businesses. - Automation – The amount of data that digital organizations can collect today is far more than any manual process can handle, and many automation tools have been built with modern data architecture practices in mind to handle the workload. - Unified dashboard – All stakeholders should be able to access data regardless of the platform or system, improving collaboration and real-time access. - Performance and productivity – Modern data architecture enables the handling of data-intensive workloads, such as AI, ML, and data analytics platforms. - Scalability – Modern data architecture allows businesses to scale to meet current demands, while the infrastructure does not need to be tied to specific platforms or environments as it was in the previous generation of data collection. One of the key reasons business leaders bring up for failing to reach their businesses analytics goals is lack of data architecture. This lack of focus on the data strategy from the start is likely going to cost businesses even more than it did a decade ago because of the cost of storing more data than ever before and running it through analytics platforms. The creation of modern data architecture practices, such as data fabrics and data meshes, has provided smart organizations with the power to collect data from disparate sources and process and display it in one unified platform.
<urn:uuid:7a571779-b2cb-475b-91c3-2562ada8f559>
CC-MAIN-2022-40
https://www.clouddatainsights.com/enabling-innovation-with-the-right-cloud-data-architecture/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030333455.97/warc/CC-MAIN-20220924182740-20220924212740-00481.warc.gz
en
0.932993
752
2.875
3
Harder. Better. Faster. Stronger. Perhaps you’ve heard one of the hottest buzz words floating around cyberspace: 5G. Many can’t quite define it, but everybody’s talking about it. The notion of 5G network holds many possibilities for consumers and our new age of technological brilliance. Wireless technology began its majestic unfolding in the 1980s, when 1G wowed analog phone users with its seemingly miraculous power. That magic evolved to 2G in the 90s, when voice data became all the rage. The early 2000s saw the rise of 3G, otherwise known as high-speed data transmission. Video transmission via 4G is wireless technology’s latest phenomenon, but impatient tech lovers are already salivating at the possibility of 5G. It’s only a pipe dream so far, but experts predict that 5G will soon become a reality. What exactly is 5G, and what does it mean for consumers? Here’s a look at the benefits we’ll gain from 5G when it becomes a reality in about five years: We’re a culture that constantly feels “the need for speed.” With 5G, we’ll get it. Because mobile technology operates within the spectrum of 600 MHz and 3 GHz, however, 5G will also be confined to that limit. The way 5G uses this finite spectrum will be more efficient than ever before. Wireless science has already found a way to do more with less: Consider MiMo technology, the multiple input/multiple output phenomenon that makes use of many antennae to service multiple devices. One year ago, Samsung introduced a MiMo product that could convert one gigabyte per second. This jaw-dropping speed sets a lofty precedent for the 5G of tomorrow. Experts predict that with 5G, a phone like the HTC One M8 smartphone on networks such as T-Mobile could download a film of 800 megabytes in just one second. What’s more, this download could theoretically take place in a vehicle moving at a speed of 300 miles per hour. Mobile speed capacities are expected to double when 5G finally hits the shelves. Today’s cell phone plans are more convenient than ever, but coverage can still be an issue when you travel to areas that aren’t on your plan’s map. It’s a fact of life, at least for now, that signal strength varies depending on where you are geographically. If 5G expands MiMo function to include more than just improved download speeds, coverage areas could expand as well — an attractive feature for today’s on-the-go wireless consumer. In order for this type of expansion to happen, phone manufacturers would need to outfit their products with multiple antennae. Right now, the average phone can’t realistically house more than two antennae, according to Royal Institute of Technology professor Jens Zander. Innovators will have to think outside of the box to find ways to fit the necessary equipment into the world’s beloved palm devices. Society’s progression to 5G will be different than our previous four progressions, according to researchers from Cornell University. It won’t be a straightforward “incremental advance,” but rather a tying together of existing innovations to yield an even more powerful product. The result for consumers will be, among other things, a “seamless user experience,” according to the researchers. That’s not to say that 5G won’t be a phenomenal blessing to consumers. When this technology reaches its height, it will be due to unprecedented achievements in wireless intelligence and flexibility. Consumers will reap the benefits. Consumers in America and Japan can expect to see normalized 5G service by 2020, but consumers in South Korea and the EU are likely to get the technology even sooner. The two regions recently announced their joint effort to establish a definition and rollout timetable for 5G by 2015. Commissioner Neelie Kroes says that once South Korea and the EU get the ball rolling, 5G will become the digital economy’s “new lifeblood.” The highly anticipated 5G wireless experience is almost here. No one knows for sure what it will look like when it arrives, but when it does, one thing is for certain: People are going to fall head over heels for this next big step in the evolution of wireless technology. For more information, contact Yazmin Gray
<urn:uuid:5b524555-6dc1-4eba-941b-e39cc041d5db>
CC-MAIN-2022-40
https://www.colocationamerica.com/blog/a-look-at-5g
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334644.42/warc/CC-MAIN-20220926020051-20220926050051-00481.warc.gz
en
0.935837
929
2.546875
3
Every technology in business uses energy and has a carbon footprint — from the lights in the ceiling to the data center they illuminate. “Green IT can mean anything from energy-efficient products to easily recycled products,” John Pflueger, principal environmental strategist at Dell, told TechNewsWorld. “While data center customers have, until recently, strictly defined greenness through energy efficiency, they are now showing a lot of interest in other, related issues, such as greenhouse gas emissions and water consumption. Other green IT topics, such as materials management and waste management, are more important in the consumer and business client space.” Determining the size of this footprint, however, is not just a matter of guesswork. There’s a science to assessing the “greenness” of the technologies used in business. “We’ve come a long way over the last few years in assessing and managing IT energy efficiency,” Elisabeth Stahl, chief technical strategist for IBM’s Systems and Technology Group, told TechNewsWorld. “Tools are getting more sophisticated and integrated.” Saving energy is good for the Earth, but it’s also good for the bottom line. Green assessment is the first step toward both goals. “Nearly every day, companies grapple with new business requirements and economic pressures to glean more data, more workloads, and more capacity per dollar invested in labor, floor space, energy and technology,” said Stahl. “Should businesses assess their greenness? On one hand, now that energy is no longer just a cheap commodity, businesses can reduce expenses significantly. On another, it’s just the right thing to do.” There are multiple forces, both internal and external, pushing businesses to assess their energy usage. Customers, investors, current and potential employees, the media, and many others are all looking for detailed information about how green the businesses they deal with actually are. “Companies today are being asked — or, in some cases, pressured — to be more accountable and transparent about their environmental activities, commitments, achievements, and impacts,” Joel Makower, chairman and executive editor of GreenBiz Group, told TechNewsWorld. “Beyond that, executives are recognizing that measuring, tracking, and reducing environmental impacts leads to other sources of business value, from cost cutting to attracting and retaining talent,” he said. “Doing a green assessment is just one step in being a well-managed company.” Methods and Techniques Once a business decides to assess its energy usage, many options are available. “There are different levels of assessment,” explained Makower. “For starters, there’s the basic tracking of inputs and outputs: how much energy, water, materials, and other resources a company is using, normalized to output. These are baseline assessments that can provide an ongoing basis for making and measuring improvements. “Beyond that,” he continued, “are more sophisticated dashboards that provide real-time or near-real-time monitoring of all processes, facilities and materials flows. Clearly, these are intended for more resource- and energy-intensive industries.” How to do this assessment depends on what the business is assessing, what information it needs from the assessment, and how it’s going to use this information. The assessment can be done with software, built-in monitoring devices, or teams of specialists brought in as part of the assessment process. “There is a plethora of tools and methodologies out there now to assess your IT energy efficiency — everything from actual meters on your servers to software tools running on your systems, to thermal visualization of your entire data center,” noted Stahl. “Sometimes a business might need to consolidate and virtualize servers — sometimes just add a special pillow to redirect air flow.” Tools such as IBM Systems Director and Tivoli products can monitor, manage and optimize energy usage, and many IT vendors have a specific energy efficiency services team that can evaluate energy usage, said Stahl. IBM, for instance, offers Systems Agenda experts — systems lab engineers who work with clients to analyze IT efficiency. “There’s a large and growing list [of assessment tools], from basic spreadsheets to sophisticated and pricey software suites,” said Makower. “The products also range widely in what they do and cover — from environmental compliance solutions to life-cycle assessment tools, to more proactive software solutions.” These tools come from a wide range of companies, both large and small, including CA Technologies, IBM, Johnson Controls, SAP, AdvantageIQ, Credit360, and many others, noted Makower. In short, choosing an energy assessment tool or service is not a one-size-fits all proposition; businesses must assess their needs before they can begin to assess their energy usage. “There’s simply no way to recommend any one solution,” said Makower. “There is a broad range of solutions designed to meet a broad range of needs.”
<urn:uuid:17365334-73bc-48d9-8c52-97ce846d880e>
CC-MAIN-2022-40
https://www.linuxinsider.com/story/figuring-out-your-carbon-footprints-shoe-size-72343.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335424.32/warc/CC-MAIN-20220930020521-20220930050521-00481.warc.gz
en
0.941649
1,068
2.8125
3
As intense wildfires have become a more serious issue in the last few years worldwide – causing billions of dollars in damage in the western U.S. alone – firefighting teams need a better solution to mitigate and better understand the cause of the spreading conflagrations. NVIDIA and Lockheed Martin have teamed up with the U.S. Department of Agriculture Forest Service and Colorado Division of Fire Prevention & Control to build automated high-resolution fuel maps using artificial intelligence (AI) and digital-twin simulation as a key step in finding solutions to the problem. Previous fuel maps were found to be ineffective in mitigating the spread of wildfires. That’s because these maps were largely dependent on Landsat data which has a low resolution, but also because the maps are only updated every two to three years, don’t provide a real-time picture of the situation, and require manual updates and inputs. “We are trying to develop an automated method for producing up-to-date, high-resolution fuel maps,” said Lydia Carroll, an AI research engineer at Lockheed Martin, during NVIDIA’s GTC Conference on March 23. “The way we’re going about this is we’re classifying Sentinel-2 images into fuel classes and we’re including fuel properties by combining Sentinel-2’s hyperspectral imagery with additional satellite sources,” Carroll said. “Our short-term goal was to improve properties mappings and classification from satellite data and incorporate additional fuel properties to reduce workload,” said Jacob Sprague, an AI Software & Systems Architect at Lockheed Martin. The end goal is delivering fuel maps at a monthly rate – instead of updates taking years – and ultimately delivering them every five days, he added. The researchers also will use NVIDIA AI infrastructure and the NVIDIA Omniverse advanced visualization and virtual world simulation platform to process a fire’s magnitude and forecast its progress. By recreating the fire in a physically accurate digital twin, the system will be able to suggest the best actions to suppress the blaze. “Quality up-to-date fuel maps are critical to accurate predictions of fire behavior. Automating fuel mapping provides more up-to-date, accurate maps to firefighters,” Carroll said. And in the future, the hope is to develop a user interface that identifies areas with outdated fuel data, construct interactive fuel map layers in Omniverse, and allow users to update fuel maps from fire crew observations on the ground. “We’re trying to create a dynamic fuel environment simulation where these fire behavior analysts can do things like toggle different fuel layers,” said Sprague. “It allows users to add in manual observations and edit what perhaps satellites saw, increasing the fidelity of this product.”
<urn:uuid:c156e751-234c-4fb6-9b18-71109e3374e7>
CC-MAIN-2022-40
https://origin.meritalk.com/articles/nvidia-lockheed-martin-using-ai-digital-twins-to-combat-wildfires/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337473.26/warc/CC-MAIN-20221004023206-20221004053206-00481.warc.gz
en
0.915245
572
2.59375
3
The advent of a sharing economy has brought a sea of change to the way we commute in the city. The Lyfts of the world have made taxi riding convenient, affordable and safe. These rides have emerged as a strong alternative to public transport, clocking millions of rides per month in some cities. The emergence of hyper-local delivery models to optimize the supply chain has also led to a large number of daily trips by these vehicles. These developments have brought about the ubiquity of either standalone or smartphone app-based GPS devices to keep track of and better regulate these rides and fleets of taxis. These GPS systems spew a ton of data generating up to GBs of data per second. With automobile & technology experts predicting that self-driving cars would replace human-driven cars in no more than a decade, the volume and velocity of GPS data is only set to increase. With that context in mind, it becomes imperative to understand GPS data and the kind of insights which can be obtained by analyzing it. A GPS or a GPS-enabled device can produce all or some of the data points mentioned below at a specified frequency (generally one record per second): - Coordinates – The latitude and longitude values are the primary data points provided by GPS devices. A set of latitude and longitude values is sufficient to locate a point on the earth. For example, (51.5007° N, 0.1246° W) denotes Big Ben in London. Just to brush up, latitude is the angular separation of a point from the equatorial plane in north or south direction while longitude is the angular separation of a plane containing the point in east or west direction relative to the plane containing the prime meridian. A collection of latitude and longitude values over time can reveal the trail left by the vehicle. - Direction – This data point denotes the geographic direction in which the vehicle is moving at that instant. A direction of 450 would mean that the vehicle is headed in north-west direction while 2250 would mean that is going in south-west direction. North is taken as the reference (00) - Speed – The instantaneous rate at which the vehicle is travelling. - Timestamp – A timestamp data point can be stripped to get year, month, day, hour, minute and second information from each record - Additional data – GPS enabled devices can also send additional information like whether a taxi is carrying a passenger or not or the amount of payload a truck is carrying. These become very powerful when combined with coordinates and timestamp data. Since the size of GPS data is usually huge, it makes sense to load such data into distributed file frameworks like HDFS and then process it using tools like Hive and Spark. The processed results can be visualized in tools like R Shiny, Tableau, D3.js and Excel. If the data size is small and if one is interested in prototyping an analytics use case then Python can be used as well. With such rich data at our disposal, a variety of analytics use cases can be performed depending upon the business context. The most common of them are as follows: 1) Distance between two points – The coordinates of two points can be used to calculate the radial distance between them. Most frequently, a central point of a city is chosen as the base and the distance of the vehicle from this base is calculated at different instants of time. The distance is calculated using a Haversine formula given by following expressions. Assume there are two points P1(lat1, long1) and P2(lat2, long2). The radius of the earth is R. Then a=(sin(dlat2))2+cos (lat1) *cos (lat2) *(sin(dlong2))2 dlat and dlong should be converted to radians before calculating a. The implementation of this calculation in Python can be done as shown below: 2) Dividing a an area into square grids – If a city or town can be divided into multiple grids of a specified equal size and insights are obtained for these individual grids, it becomes much easier to implement those insights. Here is an abridged recipe for how this can be achieved (a detailed one would require a blog of its own): - Decide a center for the city along with the number and the size of the grids wanted. Suppose you want 900 1kmX1km grids. You would need a square of side 30km. - Find the line of constant longitude at a distance of 15km from the chosen center on either side (left and right) of the center. Similarly, find the line of constant latitude at a distance of 15km on top and bottom sides from the center. These lines would give the edges and their intersection would give the vertices of the overall square - Find the latitudinal and longitudinal span of the edges and divide the span into 30 equal parts. Call them latd and longd. Start from one edge to reach the other edge by incrementally increasing the latitude and longitude by these values. - Draw lines of constant longitude and latitude at those points. This would result in 30 vertical and 30 horizontal lines and their intersection would produce 900 grids with all their vertices with known latitude and longitude These grids can be visualized using leaflet library in D3.js or R Shiny. 3) Temporal averages of important metrics – The timestamp data can be used to gauge trends about the additional data across various timeframes. For example, daily averages of distances covered in each hour. These time frames can be nested as well to get a more granular picture e.g. a plot of average payload for each half hour of the day for each day of the week. The relevant time element needs to be gleaned out of timestamp followed by a grouping of the relevant metric column by the time element. An indicative temporal visualization would look as the one shown below. The horizontal axis shows the day of the week while the vertical axis shows the half hour of the day while the metric has been shown as the heat map gradient. Geospatial analytics can unravel many mysteries and can help organizations optimize routing to match supply and demand, fight theft and related frauds and minimize the chances of accidents or the damage caused by it. Like this article? Subscribe to our weekly newsletter to never miss out!
<urn:uuid:0a0a6c5a-2c72-445b-9300-3a93fc7ae697>
CC-MAIN-2022-40
https://dataconomy.com/2017/07/gps-data-geospatial-analytics/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030333541.98/warc/CC-MAIN-20220924213650-20220925003650-00681.warc.gz
en
0.915061
1,335
3.4375
3
What is the difference between 128 bit encryption and 256 bit encryption? ? 128 bits ? Without a protocol/algorithm to go with it 128 or 256 are not types of encryption, they are just bit sizes/lengths. The bit length tends to refer to the size of prime being used and prime size is normally directly related to difficulty to factor / crack. 129 bit is twice as long as 128 130 is twice as long as 129 There is a common misconception that 256 is twice as hard to crack or twice as long compared to 128. It’s not, I CBA to do the maths but it’s LOTS bigger and LOTS more secure. Could you also explain what is the difference between Hashing and Encryption? To me they pretty much sound the same. I was referring to AES earlier. Are there any other encryption protocols as well. Thanks for the reply. You should check out these guys, they really explain this stuff, aim for high entropy that makes it harder to brute force any password. Thanks very much, I will watch this soon.
<urn:uuid:2b6357c8-5718-4bb5-b193-52851cb0b741>
CC-MAIN-2022-40
https://forums.lawrencesystems.com/t/what-is-the-difference-between-128-bit-encryption-and-256-bit-encryption/7256
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334802.16/warc/CC-MAIN-20220926051040-20220926081040-00681.warc.gz
en
0.972086
225
3.078125
3
Sergey Nivens - Fotolia The National Museum of Computing (TNMOC) has started a crowd-funding initiative to enable it to host the Turing Bombe electro-mechanical computer on the same site as its Colossus rebuild. The Turing-Welchman Bombe Rebuild Trust (TWBRT) Bombe is a fully functional and accurate reconstruction of the wartime Bombe, as designed by Alan Turing and refined by Gordon Welchman. It was used to discover the daily settings of the Enigma machines used to communicate operational messages across enemy military networks. The funding will be used to provide a new gallery and move the reconstruction of the Turing-Welchman Bombe to a new location on the Bletchley Park Estate – Block H, the home of The National Museum of Computing. Speaking to Computer Weekly about the importance of the TWBRT Bombe, Andrew Herbert, chair of The National Museum of Computing, said: “It is historically very important. It was the first machine to help automate code breaking, by allowing code breakers to check possible encryption keys.” He added that the machine represents a key step in the evolution of automation and the use of mathematics calculations. The Turing-Welchman Bombe was a relatively slow electro-mechanical machine and used mechanical switches and relays to run calculations, rather than the binary circuits used in computers, which can flip-flop between the “on” state and “off” state at incredibly high speed. When the Germans switched to a stronger cipher, Lorenz, the World War II code-crackers at Bletchley Park needed a way to decipher the codes quickly, which led to the development of the world’s first programmable electronic computer, Colossus, by British Post Office electrical engineer Tommy Flowers. “While it is not a computer, it is a very important step in the evolution of computing. To house the reconstructed Bombe close to the Colossus Rebuild makes a lot of sense from many perspectives,” said Herbert. Read more about Alan Turing - One man’s mission to uncover a WWII decryption machine – Alan Turing and the Bletchley Code Breakers. - Alan Turing was the genius mathematician who cracked the German Enigma code during World War II. - Government communications intelligence agency GCHQ has released two papers by Alan Turing on the theory of code breaking. “As a pre-computing electro-mechanical device, the Bombe will help our visitors better understand the beginnings of computing and the general thought processes that led to the development of Colossus and subsequent computers. “The story of the design of the Bombe by Alan Turing, the father of computer science, leads very appropriately into the eight decades of computing that we curate. Even the manufacture of the Bombes leads directly to British computing history – the originals were built by the British Tabulating Machine (BTM) company in Letchworth, which later became part of ICT, then ICL and now Fujitsu,” he said. John Harper, chair of TWBRT, added: “Our team of volunteers is looking forward to continuing to demonstrate how the Bombes made their vital contribution to Bletchley Park’s wartime role in the new venue.” The crowd-funding campaign to create a new Bombe Gallery close to the existing Colossus gallery is now live. Together the two machines will give visitors an unparalleled insight of the wartime code-breaking genius at Bletchley Park and the beginnings of our digital world, according to TNMOC.
<urn:uuid:bc76e760-de29-4cc1-ba5f-8eedfded88ed>
CC-MAIN-2022-40
https://www.computerweekly.com/news/252434944/The-National-Museum-of-Computing-launches-crowd-funding-effort-for-Turing-Bombe-gallery
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335444.58/warc/CC-MAIN-20220930051717-20220930081717-00681.warc.gz
en
0.921887
753
2.984375
3
Recently, Enterprise Storage Forum published an article by Henry Newman in which he argued that lithography limits and disk drive density are going to keep solid state disks (SSDs) from replacing spinning disks. To help explain his argument, I’m going to outline why one of the keys to this equation, SSD density, is in grave danger of stopping next year. SSD Technology Review Between the floating gate and the substrate is the tunnel oxide – the barrier to the floating gate through which the electrons “tunnel” into the floating gate. The transistor either has electrons tunneled into the floating gate (indicating a logical 0) or does not have any electrons tunneled into the floating gate (indicating a logical 1). The process of forcing electrons into or out of the floating gate, called Fowler-Nordheim Tunneling (F-N Tunneling), is achieved by applying a voltage between the control gate and the source or drain. When the charge is removed, the floating gate either retains the electrons (if they were tunneled into the gate) or has no extra electrons if they were removed. This allows Flash memory to retain values after power is removed. To program (write) to the transistor, which creates a logical 0, a positive voltage is applied to the drain which activates the electrons underneath the floating gate in the substrate. Then a larger positive voltage is applied to the control gate forcing the electrons to tunnel into the floating gate. To erase the transistor, or to remove the extra electrons, a negative voltage is applied to the control gate and a positive voltage is applied to the source. This forces the electrons out of the floating gate and into the source. To read the transistor, a positive voltage that is much lower than the write voltage is applied to the control gate. An additional positive voltage, also lower than the write voltage, is applied to the drain. The current between the source and the drain will determine if the floating gate has extra electrons (logical 0) or does not (logical 1). This is the basic concept transistors use to store data in the SSD. Constructing SSDs from NAND is a bit more challenging and something like assembling a pyramid of pachyderms. Creating NAND Flash Units With NAND Flash, the floating-gate transistor is used as the basis for SSDs. The first step is to daisy-chain the transistors (gates) in series. Typically 32 gates are chained in series. These groups are connected in a NOR style where each line is connected directly to ground and the other is connected to a bit line. This arrangement has advantages for cost, density, and power as well as performance but it is a compromise that has some implications. A number of these groups are then combined into pages (or sub-pages). The typical page is 4KB in size. The pages are then combined to form a block. A block, illustrated below in Figure 2, is typically formed from 128 pages giving a block a size of 512KB. The blocks are combined into a plane. Typically a total of 1,024 blocks are combined into a plane, giving it a typical size of 512MB as show in Figure 3. Typically there will be multiple planes in a single NAND Flash die. Manufacturers will also put several dies into a single NAND Flash chip. Then you can put multiple chips in a single drive. Constructing drives is fairly straightforward is not dissimilar to creating memory DIMMs. But there are a number of issues that need to be addressed for SSDs to be successful and, perhaps more importantly, these issues are particularly sensitive to the lithography size. Performing I/O Operations on SSD’s One of the coolest things about SSDs is that they don’t contain spinning media, which greatly reduces latencies and increases potential throughput. Reading, writing, and erasing are all done electrically as opposed to spinning media, where a large part of the latency is due to mechanical movement of the read/write heads and waiting for the particular block to spin around to the right spot. With SSDs, voltages are applied to the chips to cause an I/O operation . However, everything is not as easy as it seems. Recall that the transistors are connected in series. To measure the state of any particular cell, all cells in the series have to be in a conductive state. This is accomplished by applying a bias voltage of +5V that is high enough to turn on all the gates in the series so that the particular cell that has the desired data can be read reliably enough, or written to (programmed), or erased. The details of the read process are a bit involved and beyond the scope of this article, but it begins by applying the +5V bias voltage. Then a 0.5V voltage is applied above the bias voltage to read a particular cell. Notice that the bias voltage is a factor of 10 greater than the voltage needed to read the cell. Writing to the cells is similar to reading, but the voltages are much higher. For writing or erasing to cells the voltages can be as high as 20V. An important observation is that to read, write, or program a particular cell, the +5V bias voltage has to be applied. This is much larger than the read voltage of 0.5V but not as large as the program voltage (20V). Even more important is that the bias voltage is applied to all cells in the line (series) even if only one cell in the series is to be used. For example, to read one cell, a voltage 10 times greater than the actual read voltage has to be applied to all gates in the line. Moreover, typically there is never just a single line that is being read so almost all cells in the block are subjected to the +5V bias voltage. Over time, the electromagnetic (EM) field created by the bias voltage affects all cells, ultimately impacting the voltage level needed to program or write to the cell. In fact, it reduces the voltage level required to program the cell. For a brand spanking new cell, the voltage required for programming is around 20V so that you can write to the cell or erase the data. But, over time, applying all the various voltages necessary for an I/O operation will impact the voltage required to program the cell. Since not every cell has the same voltage applied at the same time, each cell will have its own individual required programming voltage. However, to accommodate the worst case, a programming voltage of 20V is applied every time. This can be bad for erasing data because the erase has to happen on a block level so all cells have to be subjected to the same voltage level (20V). This can cause the cells to actually push out more electrons than required for erasing. This phenomenon is sometimes referred to as “deep-erasing.” If a cell is deep-erased then the next time it is programmed it may not have enough electrons to be measured correctly. This is compensated by a “verify” step of the write cycle that detects the problem and repeats the write cycle until the cell achieves the correct charge (number of electrons). But, having to do this reduces the performance of the SSD because it is just this one cell that does not have the correct number of electrons. Good SSD controllers can recognize this problem and deactivate the particular line on the page or deactivate the page and use a page from the “pool” of backup pages. There has even been discussion of using a lower program voltage as the block ages. One can see that there are many aspects to making SSDs work correctly and repeatedly. One thing to remember is that without changes to the materials or the techniques for storing and retrieving data, the voltages and associated problems stay the same. Data Corruption and Retention Issues Potential SSD issues go beyond deep-erase problems. As a result of the voltages being applied to the cells, EM fields are created within the SSD. While designers have gone to great lengths to isolate cells from one another, the fields from one cell can extend to another. The effects of a field from a neighboring cell or line may have some impact on a particular cell. A fresh SSD will have cells that need approximately +5V for the bias voltage, 0.5V for a read voltage, and 20V for programming/erasing. But over time, the required voltage for programming a cell actually decreases. During programming, a cell will create a field large enough to perhaps change the properties of a neighboring cell. While designers go to great lengths to make sure that this doesn’t happen or that the SSD can compensate for this happening, this phenomenon still occurs and can lead to silent data corruption. The silent data corruption scenario is fairly simple – a cell has some voltage applied to it, a bias voltage or a program/erase voltage. The resulting field can disturb neighboring cells changing their properties. For example, the number of electrons in the cell can decrease or electrons can be tunneled into the cell. In either case, it’s possible to change the value in the cell silently. Designers are aware of this issue and have created techniques to prevent data corruption. However, these techniques cost money and no one wants to make SSDs more expensive. As a result, most SSDs are designed to meet JEDEC’s data retention requirements meaning that a brand new SSD should be able to store data at least 10 years (when the SSD has more write/erase usage, the data retention time decreases). These requirements define how long data is to be retained without loss as a function of how many write/erase cycles have occurred (on average). Therefore SSD designers will include techniques designed to prevent data loss just beyond these data retention time spans. I can’t criticize them for doing this – adding more measures would only increase cost and customers are already pretty sensitive to SSD pricing. However, it is important to realize that you can get data corruption on the SSD at some point. One of the important questions to ask is what happens to the SSD as lithography reductions are used to build NAND devices? Reducing Lithography Size Hurts Reducing lithography size brings the cells closer together, reducing the distance between the source and the drain. This allows more cells in a given space, hopefully reducing costs and allowing SSDs to have larger capacities. However, the one fundamental aspect that does not change with lithography size is the voltages that are applied to the cells. The 5V bias voltage has to be applied to bring the cells to a conductive state, you still need 0.5V to read a cell, and 20V to program/erase a cell. However, existing data corruption problems may actually get much worse because the EM fields are the same size and will be stronger in neighboring cells as they are closer together. This only makes the problem of possible data corruption worse. Getting to larger densities may not be easily achieved because of this possible data corruption. Some sources indicate that the lower limit may be 20nm. With increases in density, the probability of read disturbance increases as well. Overlaying this are the JEDEC requirements for data retention. The combination imposes some severe limits on the probability of data corruption, limiting today’s SSD designs to approximately 20 nm. Getting Out of the Lithography Pickle While the title of this article is foreboding, and we are in trouble with respect to current SSD designs, developers are actively developing new techniques. In the immediate future all SSD’s are likely to stick to the basic physics of floating-gate transistors. However, there are some additional techniques that can be used to reduce data corruption probabilities and improve data retention. For example, engineers in Japan have found a way to isolate the particular cell in a line by applying 1V to the bit line connected to the cells that are not targeted. This should reduce write disturb problems (where the 20V program voltage disturbs neighboring cells even to the point of data corruption). This allows designers to make more dense chips and to also improve throughput. Other options include the use of new materials, but these are so proprietary that they are almost non-existent. A third option is to use a new physical phenomenon. For example, there is a new type of memory called carbon-resistive memory, but switching to a new storage technology has its own set of pitfalls. Another way to improve density is to go beyond the MLC (Multi-Level Cell), which can store 2 bits per cell. This article details 3-bit and 4-bit cells being developed. These cells pose a challenge because voltage levels must be adjusted to read a particular bit in the cell without disturbing any of the others. But if this technology is successful, you can get more data with a given number of cells, increasing density and reducing costs. However, 3-bit cells have a much smaller number of program/erase cycles than others – around 1,000 cycles (MLC is typically around 10,000 and SLC is around 100,000). It’s fairly evident that SSDs using today’s technology and techniques are pretty much stuck at 20-25nm. Anything smaller and the data protection and data corruption issues become so great that either the performance is abysmal, the data retention period doesn’t meet JEDEC standards, or the cost increases. There are efforts underway to develop new technologies that improve SSD performance and density and decrease costs (a big goal). However, these are fairly new techniques and are not likely to make it into devices for quite some time. Be forewarned – SSD development could easily stagnate at 20-25nm mark. Jeff Layton is the Enterprise Technologist for HPC at Dell, Inc., and a regular writer of all things HPC and storage. Follow Enterprise Storage Forum on Twitter.
<urn:uuid:1948efec-6a07-4679-80c2-0806163dc229>
CC-MAIN-2022-40
https://www.enterprisestorageforum.com/hardware/why-flash-drive-density-will-stop-growing-next-year/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337480.10/warc/CC-MAIN-20221004054641-20221004084641-00681.warc.gz
en
0.947229
2,876
2.96875
3
User interfaces will not necessarily change the way we live — the technology will. However, for technology to work and be embraced by new consumers and emerging populations around the world, user interface design is crucial; it enables people to access and use the advancements in technology. Without simple and intuitive interfaces, end-users are not able to consume the myriad of new technologies that come out every day. Leading companies increasingly recognize that their success and the adoption of their breakthroughs are tied to a product’s ease-of-use. User interface is all about convergence, which is an old word with a new meaning. Today, “convergence” is defined as the seamless intersection and interaction of technology and life. Think of “The Jetsons,” where even life’s most ordinary undertakings — like showering and cooking — are suddenly automated through technology. In the Jetsons’ world of TV-Land comedy, a series of hyper-responsive and mischievous robots became the ultimate cliche for a culture infused with technology. Yet, the vision it encapsulates — technology embedded in all undertakings — is not simply a futuristic series of errors. Looking forward, the real land of “The Jetsons” will be represented by intuitive and easy-to-use technologies that are responsive in managing and increasing the efficiency of all kinds of things — from complex business analytics to surfing the Internet. Imagine a world where mobile phones exchange information with one’s television. The tablet supplements daily experiences and is a multifaceted device for on-demand entertainment, information exchange and work. Appliances are connected, communicating and able to be controlled remotely — all to create an interactive and technologically responsive environment that can be easily manipulated by the end-user. Imagine a world without paper currency where a flick of the wrist runs a transaction. Large consumer product companies have started looking at true convergence among their own products. In the future, however, the drive for intra-device communications will lead to the establishment of protocols across all consumer electronics. For this to truly work, all touchpoints will have to be aligned with a meaningful “life flow” that is designed to be responsive in human interaction. As in the home, everything in the work environment will be able to interact seamlessly. Accessing work files on the go, running software on any operating system or browser, and communicating globally in real-time through video conferencing will soon become simple and cost pennies. Operating systems will automatically align and sync; the mind-numbing lack of incapability will fade from memory. While convergence is on the horizon, end-users are still in for much chaos. Today, processes are being translated differently. One may not realize it, but end-users are increasingly being overwhelmed with the number of user interfaces they have to deal with and figure out. For example, look at the enormous number of mobile phone handsets in the world market. There are countless iterations and modifications between handsets and often between handsets from the same manufacturer. This forces users to constantly relearn interface conventions. What’s more, the cognitive load of today’s world on device users is poised to increase exponentially. Relief will not come in the near term. Today, companies are working in their individual silos to carve-out a market leading position. However, over the next decade or more, the disparate conventions and approaches will force consolidation and a standard set of protocols. When a standard set of protocols for all types of user interfaces emerges, consumers will benefit from a new era of simplicity in technology that will dramatically improve the totality of the experience. Consumer electronics continue to lead in the advancement of technology. The new tablet has ushered in a paradigm shift forcing companies to rethink their product offerings. It has disruptively shone a light on the power of seemingly simplistic technology design and accessibility. From home to work, people have a new idea of how technology should respond to their needs. With Baby Boomers trickling out of the workforce in larger and larger numbers, a new generation of leaders — Gen-X — is adopting and adapting to new technologies. Yet, for all the tablets Gen-X is bringing into the home and workplace, it is nothing compared to how Gen-Y will interact with technology. The next big leap for user interfaces remains the seismic shift of the tablet paradigm. Many industries want to leverage the platform. Because the tablet has its limits from a computing standpoint, for the first time, companies are leading with user-centric design to improve functionality in order to create a touch-enabled environment for their applications. Not all industries will be able to do that with ease. Companies themselves need to stay on their toes and understand that success increasingly relies on their ability to create solutions and products within the context of the new tablet environment. The consumer electronic nirvana is ease of use and engagement for end-users. The way companies will get there is by understanding their users and what motivates them. For now, user interface design is driving innovation — to both the benefit and the burden of the consumer. The road ahead will be filled with chaos in design as companies, entire industries and consumers try to untangle and leverage new concepts in convergence.
<urn:uuid:2858c333-dfec-4614-90ff-e6e05d79d46b>
CC-MAIN-2022-40
https://www.linuxinsider.com/story/designing-technologys-future-72757.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337480.10/warc/CC-MAIN-20221004054641-20221004084641-00681.warc.gz
en
0.937379
1,101
2.796875
3
A study in Biological Psychiatry has identified unique genetic changes in the brain’s reward circuitry that are associated with cocaine use, including first-time use, withdrawal, and re-exposure to the drug after prolonged withdrawal. The findings reveal important information on how cocaine addiction reprograms gene expression and provide insight into the molecular basis of cocaine addiction in unprecedented detail. In the study, mice were allowed to self-administer cocaine as a model of human addiction, and the gene expression changes were associated with their addiction-like behavior. “This study elegantly highlights the complexity of the brain’s molecular response to self-administered cocaine, pointing to mechanisms that might be targeted by treatments,” said John Krystal, Editor of Biological Psychiatry. Previous studies have been limited, focusing either on specific genes, a particular brain region, or one aspect of cocaine addiction. But molecular studies aimed at improving addiction treatment have been complicated by alterations in genes that differ throughout the brain–increasing in some regions and decreasing in others. “This study is the first of its kind to characterize the global transcriptome in brain during the life-cycle of cocaine self-administration,” said senior author Eric Nestler, MD, PhD, of Icahn School of Medicine at Mount Sinai, New York. The researchers examined six regions composing the brain’s reward circuitry, providing an enormous resource of information for studying the biological basis of cocaine addiction. To characterize the entire life-cycle, Dr. Nestler and colleagues looked for differences in gene expression when mice were first exposed to cocaine; in cocaine-addicted mice after a short (24 hours) or long (30 days) period of withdrawal from the drug; and when addicted mice were re-exposed to cocaine after the 30-day withdrawal. “The experimental design thus allowed us to study how gene expression across brain reward regions changes over time as a result of volitional cocaine intake,” said Dr. Nestler. The analysis revealed changes in many transcripts involved in key biological processes, providing clues into the brain functions that might lead to cocaine addiction. Many changes were in the same direction (increased or decreased) throughout the reward circuitry, suggesting they may be good targets for new treatments. Interestingly, the size of the changes depended on the condition–where the mice were in the life-cycle of cocaine self-administration–highlighting unique gene changes associated with the different stages of drug taking. The study also identified several molecules responsible for regulating the expression of the genes associated with addiction-like behavior. Source: Rhiannon Bugno – Elsevier Original Research: Abstract for “Cocaine Self-administration Alters Transcriptome-wide Responses in the Brain’s Reward Circuitry” by Deena M. Walker, Hannah M. Cates, Yong-Hwee E. Loh, Immanuel Purushothaman, Aarthi Ramakrishnan, Kelly M. Cahill, Casey K. Lardner, Arthur Godino, Hope G. Kronman, Jacqui Rabkin, Zachary S. Lorsch, Philipp Mews, Marie A. Doyle, Jian Feng, Benoit Labonté, Ja Wook Koo, Rosemary C. Bagot, Ryan W. Logan, Marianne L. Seney, Erin S. Calipari’Correspondence information about the author Erin S. Calipari, Li Shen, and Eric J. Nestler in Biological Psychiatry. Published April 25 2018.
<urn:uuid:de6e69e2-5417-43b6-9838-4c6e276266dc>
CC-MAIN-2022-40
https://debuglies.com/2018/06/01/a-new-study-reveals-how-cocaine-use-alters-gene-expression-and-provides-new-insight-into-the-molecular-basis-of-cocaine-addiction/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334871.54/warc/CC-MAIN-20220926113251-20220926143251-00081.warc.gz
en
0.911663
752
2.75
3
FTC Releases 10 Data Security Guidelines The Federal Trade Commission (FTC) has released a list of data security guidelines gleaned from the agency’s 50+ data security settlements, Start with Security: A Guide for Business. Their main goal is to protect consumers privacy and data, and so their guidelines reflect as such: 1. Start with security. That means, factoring security into every dept. of your business, including personnel, sales, accounting, IT, etc. Start with scope - reduce your risk of losing data by securely disposing of personal information your company doesn’t need. There’s no need to store data after a sale is complete. Hold onto info only as long as a legitimate business need requires you to, and don’t use personal info when unnecessary. 2. Control access to data sensibly. Restrict access to sensitive data to only employees that need to use the data to do their jobs. Consider separating user accounts to limit access to personal data, or control who can use particular databases. Administrative access should be limited to employees tasked with the job. Not everyone needs super-administrative rights to make system-wide changes to your system - examine job roles and restrict their permissions accordingly. 3. Require secure passwords and authentication. According to the FTC, that includes insisting on complex and unique passwords that aren’t the same or similar to passwords on other accounts. They also recommend storing passwords securely - and not in plaintext in email accounts or on their computers. The guidelines fall short of saying so, but encrypting passwords is best practice. To protect against password compromises, the FTC also recommends that businesses consider using two-factor authentication, which can stop an attacker from accessing a company network with only a password. Mobile-based two factor requires a physical device to log in, such as a smartphone. Brute-force attacks can also crack your password using automated programs - by suspending or disabling user credentials after a certain number of unsuccessful login attempts, you can guard against a potential breach. Testing for web application security flaws can also protect against an authentication bypass. 4. Store sensitive personal information securely and protect it during transmission. The FTC recommends encrypting confidential data during storage and transmission, and using industry-tested and approved methods. Some possibilities include TLS/SSL (Transport Layer Security/Secure Sockets Layer) encryption, data-at-rest encryption or an iterative cryptographic hash. Map your encryption strategy to secure data at all stages, during transmission between all servers and locations. And, ensure proper configuration of encryption, which, if done improperly, can make apps vulnerable to attacks. 5. Segment your network and monitor who’s trying to get in and out. Protect sensitive data by housing it separately on your network, and don’t allow other computers to communicate with it. The FTC also recommends using firewalls, intrusion detection and prevention tools to monitor activity on your network. 6. Secure remote access to your network. Ensure endpoint security for any computers with remote access to company networks, and limit third-party access. In one case cited by the FTC, attackers stole remote login credentials to access consumer data. While they encourage the use of firewalls and updated antivirus, two-factor authentication can also effectively stop remote attackers, as it also requires the use of a physical device to log in. Some two-factor solutions also allow you to generate temporary bypass codes for vendors that expire after one use or a set time. 7. Apply sound security practices when developing new products. Start by training engineers in secure coding practices, and follow platform guidelines for security. Verifying your third-party software’s privacy or security features is also key to ensuring they work. Testing for common, well-known vulnerabilities is also often overlooked - the FTC recommends testing for vulns identified by the OWASP (Open Web Application Security Project). 8. Make sure your service providers implement reasonable security measures. By insisting that security standards are part of your third-party vendor contracts, you can ensure reasonable security precautions like encryption are part of the deal before you sign. Verifying their compliance with security policies and standards can also help protect your company and consumer data. 9. Put procedures in place to keep your security current and address vulnerabilities that may arise. Updating and patching third-party software can reduce the risk of a compromise by ensuring protection against known vulnerabilities. Establishing a process for quickly addressing security vulnerabilities can also prevent an attack - consider an effective process for clearly publicizing security alerts to your security staff. 10. Secure paper, physical media, and devices. When it comes to paper and physical data, they should be physically secured behind lock and key, and disposed of properly by shredding documents or wiping devices. Protecting point-of-sale (POS) devices is also recommended, to prevent tampering and potential data-skimming attacks. Establishing and teaching employees about security practices while data is en route can stop physical data theft - unencrypted data left inside an employee’s car, for example, can leave a company prone to a breach.
<urn:uuid:2a870158-db9b-4963-9787-57a9c2e28781>
CC-MAIN-2022-40
https://duo.com/blog/ftc-releases-10-data-security-guidelines
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335058.80/warc/CC-MAIN-20220927194248-20220927224248-00081.warc.gz
en
0.92866
1,070
2.5625
3
Change is always difficult. Especially when presented to a species advanced enough to recognize it as it’s happening. To the human mind, the concept of Artificial Intelligence swings a pendulum from fascination to fear. Will the inevitable, apocalyptic prophecy of machines taking over the world lead to the end of civilization as we know it, as depicted in countless science fiction scenarios to date? Not exactly. Artificial Intelligence is here, happening now, intertwined with our daily lives, unbeknownst to most of us who recognize it as a convenience. AI is actively changing civilization as we know it, but don’t panic; (most of) it is for the better, and still requires the human element of actual intelligence throughout the process. The fascination with Artificial Intelligence comes from its ability to redefine the human experience and make life easier. The fear stems from the possibility it makes life so easy we become extinct, in one form or another. Both arguments hold water in theory, but currently in practice, AI is improving the quality of human life in more ways than it causes harm. Using Deep Data Science, Machine Learning and Natural Language Processing (NLP), Artificial Intelligence continues to enhance every industrial sector. Healthcare, security, travel, real estate, digital search, education and research have all seen increased efficiency, with more growth and progress expected in future trends. However, we can’t ignore the virtual elephant in the room and the prospect of lost jobs as automation from chatbots and computers continue to out-produce humans. Nothing frightens civilization more than a looming threat to survival. The real future of AI depends on how well humans will be able to adapt to an evolving world. Adaptation starts with understanding what Artificial Intelligence is, identifying its future trends and preparing for the cons that come with it. Deep Data Science, Machine Learning and Natural Language Processing (NLP) With each keystroke entered in the web, an element of data is born. Vast amounts of data processed on the web between all machines and devices require the development of systems that can automatically process large data sets and perform automated transactions. Artificial Intelligence relies on applications to help computer systems learn and predict user behavior without being specifically programmed to do so. The most visible face of AI technology is Machine Learning, particularly deep learning (based in neural networks) that learn to recognize objects in images, drive a car, or defeat the best human player in challenging games like chess. Using algorithms that work with structured and unstructured data, Machine Learning captures and understands data transferred across trillions of points per second, from Internet traffic, web search queries, content, ad-clicks, and more. Machine learning allows for the understanding, interpretation and prediction of behavior as a way to take specific action. Data Science is the art of interpreting this data to glean valuable information and insight from a user base. Deep Data Science represents the point where deep Machine Learning and Data Science intersect. Through the combination of Machine Learning and Deep Data Science, AI is able to understand human behavioral patterns on every level, then provide solutions according to those patterns. As a result, businesses across all industries, at every size, are able to see what information individuals want, compare it to mass amounts of data and deliver appropriate, relevant results. The current and future impact of AI on five key industries Healthcare – The World Health Organization (WHO) estimates the global needs-based shortage of health-care workers could exceed 14 million by 2030. Artificial Intelligence applications can drastically improve many areas of healthcare, freeing up doctors and medical personnel for more critical cases. AI Health Assistants use NLP and Machine Learning algorithms to help identify symptoms, suggest basic treatments for non-threatening illnesses and provide a personalized tracker that sense pattern deviations and warn to take further action. At Stanford University, a team of researchers created an automated classification of skin lesions using deep convolutional neural networks (CNNs) to overcome the challenge of pixel and grain issues that come when viewing these images with the naked eye. Using datasets containing thousands of images and diseases, using pixels and disease labels as inputs, the team compared the automated results to those from 21 board-certified dermatologists who diagnosed similar images. Results showed the CNN in accordance with the doctors, indicating an AI capable of identifying skin cancer could potentially reach mobile devices, extending far beyond the waiting room door, helping millions of patients in the process. The opportunities for AI in healthcare are endless, from predicting heart disease to analyzing data sets of molecular structures that aid in pharmaceutical drug development. But such sensitive information walks a fine line between ethics and privacy, not to mention the potential loss of administrative jobs, consumed by automation. Security – Artificial Intelligence is being used as a form of protection, from evils in every direction. From hackers to home burglars, security requires extensive visibility into a given situation at any moment in time. This results in exorbitant amounts of data incapable for humans to process efficiently. Fraud detection technology, facial and vocal recognition arm the cyber space and Internet of Things, while physical security measures used by law enforcement such as drones, ground-based robots and body cams contain smart software with deep learning-based video analytics solutions. Legal – In an industry reliant on paperwork, Artificial Intelligence is a gift. Natural Language Processing can sift through thousands of documents at a fraction of the pace of its human counterpart, with less error. AI tools can help lawyers scan historical content for precedent or evidential testimony, and analyze internal data to help define billing patterns, case wins and streamline operations. Automotive – 2016 accounted for the highest number of motor vehicle deaths in the United States since 2007, according to the National Safety Council. At 40,200 deaths, the total cost including injuries and property damage reached over $432 billion. One of the most prominent uses of AI technology is for the development of self-driving vehicles, with its most valuable benefit being the estimated increased level of safety. AI is used to learn human behavior and react to driving conditions the way a human would, if a human had superhero-like powers to sense terrain changes, weather factors and remain free from distraction. Manufacturers can use AI data to monitor vehicle performance, identify faulty parts and gain actionable information for improvement. Connected consumers benefit from a range of perks such as local information on nearby stores, eateries and fuel stations, or additional safety features such as software vehicle updates, biometric capabilities or applications that can detect inattentive or tired driving. Search -The world of search is a wide one that connects to each industry. Generally, the first step to consumer action in any of the industries covered above starts with a quest for information. Search providers are using Artificial Intelligence to monitor the infinite bytes of data collected all over the world to help refine the results users see. Perhaps you need to search for a doctor nearby or a lawyer who specializes in estates. Maybe you’re interested in a smart home security system or you want to know how soon you can read a book while your car drives you to work. The combination of AI technology such as Machine Learning, Deep Data Science and NLP works together to interpret a query and provide accurate answers obtained from recognized patterns pulled from the data abyss. Users breed data. AI learns from that data, which leads to more insightful, powerful search applications. Preparation and agility are the antidote to obsolescence Artificial Intelligence bridges the gap that once existed between time and efficiency. Tasks that seemed impossible to financially account for in man-hours can now be done in seconds for a fraction of the cost. No human could ever collect, categorize and analyze data at the pace or with the precision of AI. This opens a world of opportunity for anyone willing to embrace and invest in an AI future, leaving all others caught in the dust of a changing world. In order to set themselves apart from other retailers, brands are using AI to generate highly personalized experiences, dependent on demographic data to help with decisions ranging from inventory assortment to campaign decisions and loyalty programs. This may demonstrate an optimistic attitude towards Artificial Intelligence, but the reality is that in order for companies to compete, they must manage to adjust for the AI disruption.
<urn:uuid:2d4efca1-ec6e-4d64-9fc6-3a29fb577cd7>
CC-MAIN-2022-40
https://bdtechtalks.com/2018/02/22/nlp-deep-learning-data-science-human-experience/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335469.40/warc/CC-MAIN-20220930113830-20220930143830-00081.warc.gz
en
0.922377
1,664
3.203125
3
[German]There is a previously unpatched vulnerability in the implementation of the NTFS file system used by Windows 10. Via this vulnerability, it is possible for attackers to destroy the contents of an NTFS volume used under Windows 10. It is enough to place an appropriately crafted file on an NTFS volume to trigger the flaw. A security researcher has now pointed this out for the umpteenth time. I've pulled out the latest disclosure by @jonasLyk from January 9, 2021 on Twitter once – see the following tweet. Accessing a crafted folder is enough to exploit the vulnerability. The whole thing can be triggered remotely (e.g., downloading a crafted shortcut file or ZIP archive). onas L. had already publicly disclosed this NTFS vulnerability in August 2020 as well as in October 2020, without anything happening. So he reached out to Bleeping Computer, who tested it and then disclosed it in this article. A short command is sufficien For an attacker exploiting this vulnerability, a one-line command is enough to corrupt an NTFS-formatted hard drive. This can be done by placing a crafted file (even remotely) in a folder on the affected drive. Once this file is read, Windows prompts the user to restart the computer to repair the corrupted disk entries. Bleeping Computer shows a command here as an example of such a trigger, but warns against testing that on a system. This is because the NTFS drive may be unreadable and lost afterwards. After running the command in the Windows 10 command prompt, the error: "The file or directory is corrupted and unreadable." is immediately displayed. For testing, one should use a virtual machine. The command shown accesses the $130 attribute of an NTFS volume. This $I30 attribute is used by the NTFS file system to maintain an index of all files/subdirectories that belong to a directory. Currently, it is unclear why accessing this attribute corrupts the drive. Speaking to Bleeping Computer, Jonas L. said: 'I have no idea why accessing the attribute corrupts NTFS volumes. It would be a lot of work to figure that out because the reg key that should trigger a BSOD when corrupted doesn't work. So I'll leave that to the people who have the source code.' It is enough to put a prepared file (e.g. .lnk file) on an NTFS drive (the data does not need to be opened) to trigger the error. Also conceivable would be prepared ZIP archives that trigger the error when unpacked. After the drives have been corrupted, Windows 10 generates errors in the event log stating that the Master File Table (MFT) for the respective drive contains a corrupted record. Bug present from Windows 10 version 1803 As Jonas L. stated to Bleeping Computer, he was able to exploit the bug in Windows 10 version 1803. The bug is said to remain exploitable up to the current version 20H2. Bleeping Computer writes in the article that sources from security research circles say that serious vulnerabilities like this have been known for years. The bugs were reported to Microsoft earlier, but were not fixed. Bleeping Computer checked with Microsoft to find out if they already knew about the bug and if they would fix it. The response, "Microsoft has made a commitment to its customers to investigate reported security issues and we will provide updates to affected devices as soon as possible." Addendum: There is an inoffizial fix, see Windows 10 NTFS bug gets unofficial fix from OSR Cookies helps to fund this blog: Cookie settings
<urn:uuid:0e6c5b1f-bac2-4f15-a3fa-b5c0e18e608f>
CC-MAIN-2022-40
https://borncity.com/win/2021/01/15/windows-10-schwachstelle-ermglicht-ntfs-medieninhalte-zu-zerstren/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337398.52/warc/CC-MAIN-20221003035124-20221003065124-00081.warc.gz
en
0.946037
760
2.546875
3
In this column, we briefly introduce the concepts of virtual reality (VR), mediated reality (MR), mixed reality (MR), augmented reality (AR), diminished reality (DR), augmented virtuality (AV), and hyper reality (HR). There are so many new technologies bouncing around these days that it’s difficult to wrap one’s brain around all the terminology and abbreviations, so I thought it might be a good idea to provide some easy-to-read overviews. In this column, we are going to briefly introduce the concepts of alternative realities, virtual reality (VR), mediated reality (MR), mixed reality (MR), augmented reality (AR), diminished reality (DR), augmented virtuality (AV), and hyper reality (HR). Since time immemorial (see Is Time Truly an Illusion?), humans have been entranced by the thought of realities other than our own. Imagine the scene thousands of years ago in which a group of people are huddled around an open fire at night, listening to their storytellers weave tales involving gods and supernatural entities, daring deeds and fabulous failures, heroes and heroines, winners and losers, and monsters and mythical creatures. Every time we open a book, watch a movie, or see a play, we are transported to an alternate reality. Occasionally, some amongst our number decide to help things along, as did the anthropologist Carlos Castaneda with psychotropic plants like peyote and interesting mushrooms of the genus Psilocybe. Amazingly enough, Carlos somehow managed to document his experiences in his book A Separate Reality. Excluding things like movies, one of the earliest examples of using technology to create an alternative reality was called the Sensorama. This little beauty was created in 1962 by Morton Heilig, who was a producer, director, writer, cinematographer, and editor of films and TV programs. The Sensorama offered an immersive, multi-sensory (now known as multimodal) sensory experience. Primarily a mechanical device, the Sensorama boasted a stereoscopic color display, fans, odor emitters, a stereo‐sound system, and a motional chair. It simulated a motorcycle ride through New York and created the experience by having the spectator sit on an imaginary motorcycle while experiencing the street through the screen, fan-generated wind, and the simulated noises and smells of the city. All of these elements were triggered at the appropriate times, such as the smell of the exhaust when the rider approached a bus. The petrol fumes and the aroma of pizza snack bars were recreated using chemicals. I don’t know about you, but I would love to take a ride on the Sensorama. If I can ever get my time machine working (again), I would also love to bring Morton Heilig forward to the present day and let him play with my Oculus Rift virtual reality system, but we digress… The Future is Closer Than We Think Arthur C Clarke was a prolific writer and a brilliant futurist, and he is widely known for the third of his famous three laws, which states: “Any sufficiently advanced technology is indistinguishable from magic.” He was right; these days we have access to technologies that would have seemed the stuff of magic not so long ago. Let’s take a brief look at a few of these little rascals… Physical Reality (PR): This refers to the real world we see around us, although the more I read, the harder it becomes to determine what is real and what isn’t. In his book, Reality is Not What It Seems: The Journey to Quantum Gravity, Carlo Rovelli works his way from Sir Isaac Newton with his space, time, and particles to Faraday and Maxwell and their fields to Einstein and his spacetime, ending up with the current theory of Covariant Quantum Fields. In a nutshell, it now appears that we are all swimming in a sea of quantum foam — there really isn’t such a thing as space that “contains” things and there isn’t really such a thing as time during the course of which events occur. I know this sounds silly when you say it out loud, but it makes a lot more sense when you read the book. Virtual Reality (VR): In this case, the reality is completely generated by a computer. VR has tremendous application in industry for training, such as teaching crane operators how to manipulate heavy loads in adverse conditions. As one example, consider the task of loading cargo onto a boat in high winds with a heavy swell in limited visibility. Preferably, you would wait for conditions to improve, but you might not have this luxury in the case of a mission-critical situation like an emergency aid deployment. It’s much better to practice this stuff in the virtual world than to take chances with real boats, cranes, and people. VR is also of interest for education and entertainment. I personally enjoy a quiet afternoon cracking puzzles (see Solving Multifaceted Mysteries in VR), surviving a zombie apocalypse (e.g. Arizona Sunshine) or exploring alien worlds (e.g., Obduction). Of course, this is going to be so much better when we are no longer tethered to a honking big host computer (see Time for an Oculus Quest?). Also, we are going to see increasing use of VR in conjunction with artificial intelligence (AI) (see AI, ANNs, ML, DL, and DNNs). Games like Star Trek: Bridge Crew already make use of AI technology. In this case, the idea is that you are part of the bridge crew flying a starship on various missions. This crew comprises four positions: Captain, Helm, Tactical and Engineer. You can either play with your friends or — on the off-chance you don’t have any — you can populate the other positions with AI-enabled characters. Augmented Reality (AR): This refers to an interactive experience of a real-world environment in which the objects that reside in the real world are enhanced by computer-generated perceptual information, sometimes across multiple sensory modalities, including visual, auditory, haptic, somatosensory, and olfactory. Diminished Reality (DR): Also known as deleted or deletive reality, DR is the complement of AR. As opposed to adding information or stimuli to a real-world scene, DR involves removing or diminishing information or stimuli from the real world. Examples would be to fade down (or out) extraneous voices or other sounds when you are conversing with someone in a noisy environment, fading or blurring portions of the scene you are viewing, or completely removing objects or people from the reality with which you are engaging. Mediated Reality (MR): This refers to adding information to (augmenting), subtracting information from (diminishing or deleting), or otherwise manipulating one’s perception of reality. This can be achieved by means of a wearable computer (e.g., a headset) or a hand-held device (e.g., a smartphone). I personally believe that, in the not-so-distant future, the combination of MR (i.e., PR + AR + DR) with AI is going to dramatically change the way in which we interact with the world, our electronic systems, and each other. Augmented Virtuality (AV): As opposed to AR, in which objects and scenes in the real world are augmented with computer-generated information, augmented virtuality refers to augmenting virtual environments with real-world objects or people. Mixed Reality (MR): This refers to merging of real and virtual worlds to produce new environments and visualizations, where physical and digital objects co-exist and interact in real time. Mixed reality does not exclusively take place in either the physical or virtual world, but is a hybrid of reality and virtual reality. Reality-Virtuality Continuum: As illustrated in the diagram to the right, this refers to a continuous scale between 100% physical reality (PR), the real world, at one end of the spectrum, and 100% virtual reality (VR) at the other end. Is mediated reality (MR) really a superset of mixed reality (MR)? To be honest, I would personally say they are one and the same thing, but those people who are fond of presenting things in the form of Venn diagrams usually depict mediated reality as encompassing mixed reality. It’s also probably worth noting that different people may have slightly different interpretations of a lot of this, plus folks are constantly adding new terms that muddy the waters (not that I’m bitter, you understand). Hyper Reality (HR): First, we have to distinguish hyperreality (one word) from hyper reality (two words). The former, hyperreality, is an inability of consciousness to distinguish reality from a simulation of reality, especially in technologically advanced postmodern societies. Hyperreality is seen as a condition in which what is real and what is fiction are seamlessly blended together so that there is no clear distinction between where one ends and the other begins. Hyperreality also allows the co-mingling of physical reality (PR) with virtual reality (VR) and human intelligence (HI) with artificial intelligence (AI). By comparison, hyper reality (HR) refers to a total information overload, which we might think of a mediated reality (MR) on steroids. Perhaps the best way to visualize this is to look at the Hyper Reality concept film by Keiichi Matsuda (it’s only ~6 minutes long). This amazing video depicts a kaleidoscopic vision of the future in which physical and virtual realities have become totally entwined. On the one hand, it looks exhilarating; on the other hand, I don’t think my poor old noggin would be able to take the strain. As always, what’s your take on all of this?
<urn:uuid:3c356e88-1972-4259-8991-02520fce25bc>
CC-MAIN-2022-40
https://resources.experfy.com/bigdata-cloud/what-the-faq-are-vr-mr-ar-dr-av-and-hr/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337398.52/warc/CC-MAIN-20221003035124-20221003065124-00081.warc.gz
en
0.949344
2,045
2.96875
3
Real-time data mining — powered by neural-network technology — has begun to remake the way large corporations manage customer accounts. The technology has been helping companies gain deep insight into customer purchasing patterns. While the technology is just now beginning to gain appeal, research to develop neural networks was started years ago by the Pentagon’s Defense Advanced Research Projects Agency (DARPA) and a firm founded by technology visionary Robert Hecht-Nielsen, whose company, HNC Software, was acquired last fall by Fair Isaac, a developer of statistics-based consumer and corporate credit score software. “Traditional fraud detection operates with a delay of months or years,” Tammy Delatorre, a spokesperson for Fair Isaac, told TechNewsWorld. “While useful, this approach does little to prevent fraudsters from committing costly fraud schemes and then disappearing with the money.” To solve this fraud problem and to help predict customer trends, large corporations have begun to deploy neural-network technology — so-called because it is patterned after the human brain’s own synapses. Predictive Neural Networks The brain learns from experience, and that is the general goal of neural-network technology, which is generally thought to be the next major advancement in the computing industry. Even simple animal brains are capable of functions that are currently impossible for computers. Computers do certain things well, like keeping ledgers or performing complex math. But computers have trouble recognizing even simple patterns, much less generalizing those patterns of the past into actions for the future. However, advances in biological research have begun to enable an understanding of how natural-thinking mechanisms store information as patterns. Some of these patterns are very complicated — even the seemingly simple brain patterns that give us the ability to recognize individual faces from many different angles. This process of storing information as patterns, using those patterns and then solving problems encompasses the entire neural-networking focus in computing. The field does not use traditional programming, but rather focuses on the creation of massively parallel networks and the training of those networks to solve specific problems. This field also uses words that are very different from traditional computing terminology — words like “behave,” “react,” “self-organize,” “learn,” “generalize” and “forget.” From Lab to App The neural-networking technology at work today so far has taken the form of fraud detection and customer-behavior prediction. The fraud-detection software works by identifying patterns of suspicious behavior and providing a score of a transaction based on that claim’s degree of risk. “The score allows claims professionals to determine which claims must be taken out of the payment stream for further investigation, and allows the rest of the claims to be fast-tracked for payment,” said Fair Isaac’s Delatorre. “The system also provides reason codes to help investigators determine the most appropriate action on each high-risk claim.” Many banking clients try to keep their use of such fraud-fighting technologies quiet. But this spring, Fair Isaac announced that London-based HSBC Holdings PLC, one of the largest banking and financial services organizations in the world, with 9,500 offices in 80 countries and territories, is using Fair Isaac’s predictive software to fight fraud. Other technology developers are taking a similar approach, saying their technologies provide “predictive analytics” for use in enterprise environments. By Any Other Name To be sure, neural-networking technology falls under the umbrella of artificial intelligence. “AI is known by different names — data mining, statistics, machine learning — and it has fallen in and out of fashion over the last few decades, suffering from overhype and an inability to deliver,” Faye Merrideth, a spokesperson for developer SAS, told TechNewsWorld. In recent years, companies have taken batches of data and extracted information from that data. This process generally has been conducted in the back office. Now, however, the technologies are moving toward the front office — to sales representatives and others who need the information for immediate decision-making to act upon business events at the front lines of an organization. Using relational databases and Predictive Model Markup Language (PMML) — two technologies at the heart of the neural-network strategy — developers are creating data-mining technology and providing scores for all kinds of information, such as when to offer a customer a new product and which customers are most likely to respond to an offer. The Next Generation “A new generation of data mining is upon us,” Stephen Brobst, CTO of Teradata, told TechNewsWorld. “This means that data-mining algorithms must be executed by engines capable of real-time scoring rather than by traditional batch-file processing techniques.” To predict which customers are likely to accept an offer, the algorithms are created on the basis of which customers have accepted similar offers in the past and which customers have rejected offers. “After selecting an appropriate discovery algorithm, the data-mining tool is applied to the training set to uncover the patterns that can be used for predicting the targeted behavior,” Brobst explained. Mole Hills to Mountains In the most advanced neural-network and predictive-analysis systems currently deployed, only a few dozen transactions by any given customer are needed to determine likely future behavior. Data used for analysis can come from all kinds of transactions, including call-center logs, warranty logs, e-mail, faxes, Web-site contacts and even surveys, Sue MacDonald, a spokesperson for developer Intelliseek, told TechNewsWorld. “Not only do they derive factual information from all that data, they’re using language-processing algorithms and structures to detect other facets, such as underlying sentiments,” said MacDonald, noting that the software can help determine if customers are pleased, angry or indifferent. “One of the goals is for companies to better segment and target their customers.” Intelliseek operates an Advanced Research Center in Pittsburgh, site of the former WizBang Labs, which Intelliseek acquired. Other players in this emerging technology field — including Knowledge Extraction Engines and CombineNet — use advances in statistics and game theory to optimize business processes for companies. DARPA is still active in this area, too, recently investing US$29 million in a research project called Perceptive Assistant that Learns (PAL). The idea is to learn the habits and thinking styles of military commanders in the field so they can automate some of their routine functions and concentrate on winning a particular conflict. The growth of client-server computing during the 1990s gave predictive analytics the boost it needed in the corporate marketplace. Developers created huge relational databases using client-server technologies — as well as advances in networking protocols — to bring neural-network applications to life. These applications today are not HAL from Stanley Kubrick’s movie “2001: A Space Odyssey,” but thinking machines are starting to offer companies significant strategic and tactical advantages in the marketplace. Building on these developments, and further refining neural-net systems so they can incorporate all kinds of data types — such as images — is what many companies working in this area hope to accomplish. “This is the most important scientific challenge of our time,” said Hecht-Nielsen. “And finding the answers will be the challenge of the millennium.”
<urn:uuid:16ba5959-d969-4107-9be2-d9fd83409ed0>
CC-MAIN-2022-40
https://www.crmbuyer.com/story/neural-network-technology-moves-into-the-mainstream-31280.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337504.21/warc/CC-MAIN-20221004121345-20221004151345-00081.warc.gz
en
0.9376
1,555
2.640625
3
Welcome to the next tutorial of SAP PP training! Before reading this tutorial it is mandatory to learn about SAP bills of material because this tutorial will build upon this concept. SAP phantom assembly is a special non-stock material that has its own components (i.e., a product structure). SAP phantom assemblies are frequently used in automobile industry where several items (like nuts, bolts or accessories) are required at different levels of BOM. Instead of repeating these items again and again in BOM, they are attached to a phantom item and this phantom item will represent those items. Advantages of SAP Phantom Assembly SAP phantom assemblies reduce master data maintenance. Since several items are grouped by a phantom assembly, we do not need to maintain all these items again and again. Also, changes in items can be done inside BOM of the phantom assembly, so that changes are applicable to the high-level BOMs where this phantom assembly is used. Master Data Settings for SAP Phantom Assembly Phantom materials are identified by “Special procurement” key in material master (MRP 2 View). The special procurement key “50” represents a phantom item. SAP Phantom Assembly in Planning SAP phantom assembly is not actually produced or purchased. Therefore, phantom assemblies pass the dependent requirements directly to the lower BOM level. Dependent requirements will be generated for components of the phantom assembly, and planned orders or purchase requisitions are created only for these components. Let’s say a wheel assembly is a phantom material which groups together rear wheels, front wheels and their associated components. This wheel assembly is assigned to a final product “truck”. Here when truck production is planned for future, SAP system also plans for cabin, chassis and engine sub-assemblies. The requirement or demand for the chassis assembly will be passed to the wheel assembly which is a phantom assembly. During MRP (Material Requirement Planning) system will pass the chassis requirement directly to front wheels and rear wheels. SAP Phantom Assembly in Costing During costing, SAP system explodes BOM and calculates the planned costs for all the material components in the phantom assembly. Phantom assembly’s material components are displayed and updated in the cost estimate of the higher-level material. The results of the cost estimate will be updated with a reference to the phantom assembly. Explosion Control for SAP Phantom Assembly We can stop the phantom assembly behavior at the BOM item level by using phantom assembly off indicator. This will override the special procurement type 50 setting and will make it behave as a normal assembly. This business function can be used in cases where the same assembly acts as a phantom item in some of the final assemblies and as a normal assembly in some other final assemblies. Let’s consider an example. A truck with 12 Ton capacity wheel assembly where it acts as phantom item (i.e., directly assembled into the truck). But for a truck with 20 Ton capacity, wheel assemblies are assembled separately and are attached to the truck due to some technical design aspect. In former case, the wheel assembly is defined as “Phantom Item” in material master by setting the special procurement key 50. On the other hand, in the BOM of 20 T truck, the phantom assembly off Indicator will be set. So that a separate planned order will be generated for 20 T truck requirements alone but not for 12 T truck. In the material master of the wheel assembly, phantom assembly is set in the special procurement field. The wheel assembly is a material component in a higher-level BOM and has item number 0010. It also has a BOM of its own, containing components rear wheels and front wheels. In the standard system, dependent requirements are generated for components when the BOM is exploded in MRP. The following graphic shows how a BOM that contains a phantom assembly is exploded for production. The phantom assembly (wheel assembly) is ignored and only the components of the phantom assembly (components rear wheels and back wheels) are read. The BOM of 20 T truck that contains a phantom assembly (wheel assembly) is also exploded. However, the explosion type is set to off for phantom assembly in BOM, so the material wheel assembly is also included in the planning run. Did you like this tutorial? Have any questions or comments? We would love to hear your feedback in the comments section below. It’d be a big help for us, and hopefully it’s something we can address for you in improvement of our free SAP PP tutorials.
<urn:uuid:72de283f-3233-413c-98dd-95ebd11ef3ea>
CC-MAIN-2022-40
https://erproof.com/pp/sap-pp-training/sap-phantom-assembly/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337668.62/warc/CC-MAIN-20221005203530-20221005233530-00081.warc.gz
en
0.905957
934
2.671875
3
Follow @ASM_Educational Using DHCP Server with Inter-Vlan Routing (Router on stick) On Last Video we talked about the router on stick that is we made sure that client from Network 10.10.10.0 can communicate with Network 184.108.40.206 Monthly Archives:: April 2016 Follow @ASM_Educational Setting up VLAN’s On a new switch, all the ports are in VLAN 1 by default. We just plug in the Ethernet cables and the devices can communicate. Furthermore, all the ports are in the up/up (administratively up) mode. Alternatively referred to as a bus slot or expansion port, an expansion slot is connection or port located inside a computer on the motherboard or riser board that allows a computer hardware expansion card to be connected. Power Supplies Okay, now that we’ve tested our AC outlet and put some protective power devices into play, let’s go ahead and talk power supplies. You may not typically handle the insides of your computers, but if your business requires knowledge of Central Processing Units (CPU), then you’ll also need to learn about CPU sockets Computer port In computer hardware, a port serves as an interface between the computer and other computers or peripheral devices. What does a hard drive do? Storage is the hard drive’s responsibility. Everything you keep on your computer is on a hard drive. What Is Memory, or RAM? System memory, frequently called main memory or RAM (Random Access Memory), is a type of computer memory that can be accessed randomly. Oversight Committee Representation Organizational or corporate governance has existed since time immemorial to ensure the efficient running via control structures. Wireless Communications WAP (Wireless Applications Protocol) –Developed for wireless (PDA’s, mobile phones, pagers, etc. devices to communicate) Tunneling is a method of transferring data from one network to another by encapsulating the packets in an additional header. The additional header provides routing information so that the encapsulating payload traverse the intermediate networks. IPSec IPSec is an architecture or framework for security services for IP networks. It works at the Network Layer of the OSI Model. It is actually a standard for secure data transmission. It provides mechanisms for authentication and encryption. Defined by RFC 4301 and carries a set of functions, it is mandatory in IPv6. IPSec… Read more » Open Systems Interconnect (OSI) Model WANS and Their Components Wide area networks (WANs) are considerably different than LANs. Organizations usually own their own LANs, but WAN services are typically leased; it’s not feasible to have your network guy run a cable from New York to Dallas. Source mc mcse Certification Resources Cable Testers – Cable testers are electronic devices used to test a cable’s integrity by checking for opens and shorts which can cause connectivity problems. Source mc mcse Certification Resources Crosstalk Symptoms: Slow network performance and/or an excess of dropped or unintelligible packets. In telephony applications, users hear pieces of voice or conversations from a separate line. Source mc mcse Certification Resources Gather Information on the Problem In a contact center network, problems are typically discovered and reported by one of the following types of users: Source mc mcse Certification Resources Bandwidth – is the average number of bits that can be transmitted from the source to a destination over the network in one second. Follow @ASM_Educational Cisco CCNA-Router on Stick Inter VLAN Communication (Router on Stick) Earlier we built and tested the following topology: The Types of Wireless Attacks Part 2 CompTIA Security+ Objective 1.2 WPS Attacks Wi-Fi Protected Setup (WPS) allows users to configure a wireless network without typing in the passphrase. Instead, users can configure devices by pressing buttons or by entering a short personal identification number (PIN). For example, a user can configure a new wireless… Read more » The Types of Wireless Attacks Part 1 CompTIA Security+ Objective 1.2 Replay Attacks A replay attack occurs when an attacker copies a stream of messages between two parties and replays the stream to one or more of the parties. Unless mitigated, the computers subject to the attack process the stream as legitimate messages, resulting in a range… Read more » CompTIA and Immersive Labs are challenging cybersecurity professionals in the United States and United Kingdom to test their penetration testing skills this month. The two organizations launched the pen test challenge on the opening day of RSA® Conference 2019. “Penetration testing, if done right, is a proven and valuable activity that all organizations should engage… Read more » The nature of cyberattacks is constantly in flux, always evolving to keep pace with the times. Hacking of websites, theft of credit card information and other personal information has become an almost daily occurrence, along with illegal remittances via Internet banking. Recent years have witnessed the accelerating dissemination of new technologies such as IoT… Read more » “Richard Clarke, a former counter-terrorism expert for the United Government, once said, “If you spend more on coffee than on IT security, you will be hacked. What’s more, you deserve to be hacked”. While the latter is a tad harsh (we wouldn’t wish a cyber attack on anyone!), the former is certainly true.” If you… Read more » 1. Your manager asks you to implement a system that can filter out unwanted content, such as viruses and unproductive Internet content. The best way to accomplish this would be through a system that implements a: In this cram guide, we would like to give you some short tips on what to study for the Security+ exam to ensure a great score. One of the most essential portions of information security is the design and topology of secure networks. What exactly do we mean by “topology?” Usually, a geographic diagram of a network comes to mind. Definition – What does Malicious Software (Malware) mean? Malicious software, commonly known as malware, is any software that brings harm to a computer system. Malware can be in the form of worms, viruses, trojans, spyware, adware and rootkits, etc., which steal protected data, delete documents or add software not approved by a user.
<urn:uuid:22ca3872-7ca8-4efb-8c1c-3fef2988864f>
CC-MAIN-2022-40
https://asmed.com/2016/04/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337971.74/warc/CC-MAIN-20221007045521-20221007075521-00081.warc.gz
en
0.902575
1,336
2.734375
3
Earlier this month, the US Government announced the availability of nearly $1 billion in U.S. Department of Commerce’s National Telecommunications and Information Administration (NTIA) grants to expand broadband access and adoption on Tribal land. Vice President Kamala Harris, Commerce Secretary Gina Raimondo and Interior Secretary Deb Haaland announced the funding on 3 June 2021 at the White House. The funding is part of the Consolidated Appropriations Act 2021 which in turn established the Tribal Broadband Connectivity Program. Grants will be made available to eligible Native American, Alaska Native and Native Hawaiian entities for broadband deployment as well as for digital inclusion, workforce development, telehealth and distance learning. Why is this program needed and who will it help? According to an American Indian Policy Institute analysis of federal data, just 67 percent of tribal lands in the continental U.S. have access to broadband internet, with the majority only having access to broadband speeds considered by the Federal Communications Commission (FCC) to be less than “minimally acceptable.” “Broadband is basic infrastructure. It’s as necessary now as electricity once was,” Rep. Teresa Leger Fernandez, D-N.M., the chair of the House Subcommittee for Indigenous Peoples of the U.S., said in an interview. “Everybody knows now, viscerally, what it means to depend on the internet.” “This kind of package would not only provide resources to actually put lines in the ground or overhead, but it would also end up saving lives, when you think about how it can help people get vaccinations, get educations,” she added. “We can’t leave anyone behind.” “For generations, a lack of infrastructure investment in Indian Country has left Tribes further behind in the digital divide than most areas of the country,” said Secretary of the Interior Deb Haaland, in a prepared statement. “We have a responsibility as a country to build infrastructure that will fuel economic development, keep communities safe, and ensure everyone has opportunities to succeed.” The American Jobs Plan also proposes a $5 billion fund called the Rural Partnership Program that would specifically help rural regions and tribal nations with economic development, which could include additional funding for broadband infrastructure. The plan ‘prioritizes’ support for networks owned and operated by local and tribal governments and ‘providers with less pressure to turn profits and with a commitment to serving entire communities.’ The plan would put money into infrastructure that must be built on tribal lands, including for extending fiberoptic cables over long distances to remote homes. The plan also seeks to reduce the cost of broadband service — which for many on tribal lands who have it already has been prohibitively expensive — and increase internet usage. Any proposed resolution will need to address problems creatively and with targeted solutions While the pandemic blew the digital divide wide open, says Traci Morris, executive director of the American Indian Policy Institute, tribal lands for years have suffered from additional unique circumstances far beyond a lack of funding that made building broadband internet especially challenging. They include a complex permit process because hundreds of Native American reservations are sovereign lands; the cost of building broadband infrastructure in sparsely populated areas with rugged terrain is not profitable for private communications companies; and many reservation residents still need running water, electricity and a cellphone signal. Any investment for fixing the digital divide needs to target these problems through solutions that are specifically designed to address such barriers. Fixed wireless access (FWA) which is a way of providing wireless internet access through radio links, without the need for cables can be an ideal solution. Deploying BLiNQ solutions, like BLiNQ’s FW-600 for example, allow wireless service providers to offer their customers in remote, rural areas high-speed, reliable, affordable internet access. With superior non-line-of-sight performance where there is challenging terrain, filling in pockets with robust internet connectivity can be more achievable. Moreover, the BLiNQ units are mountable on towers which require no additional building or construction of additional infrastructure. Details on who can apply and how NTIA is seeking infrastructure projects that expand the availability of broadband services on Tribal lands and prioritize deploying broadband infrastructure to unserved households, as required by the Act. More information about the program, including requirements for grant applications, can be found in the Notice of Funding Opportunity published today on grants.gov. NTIA will be holding webinars to inform the public about the grants. The next Tribal Broadband Connectivity webinars will be held on June 16th and 17th. “These investments will help many Native American, Alaskan Native and Native Hawaiian communities gain long overdue access to life-saving technologies, economic opportunities, remote learning and countless other benefits,” said Secretary Raimondo. “This funding is an important step forward, but we cannot stop here. Access to broadband is as essential as electricity to our everyday life. Under President Biden’s American Jobs Plan, we will have the resources to deploy high-quality broadband infrastructure to every Tribal community.”
<urn:uuid:04812c7e-e259-411e-8389-e7d74dc3755a>
CC-MAIN-2022-40
https://blinqnetworks.com/what-you-need-to-know-about-the-tribal-broadband-connectivity-program/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337971.74/warc/CC-MAIN-20221007045521-20221007075521-00081.warc.gz
en
0.943382
1,066
2.53125
3
In 2011, Google added public key pinning to Chrome. They white-listed the certification authority public keys that could be used to secure Google domains. The intent was to mitigate man-in-the-middle attacks that were performed by using a fraudulent SSL certificate. The approach was not scalable for all websites, but Google did offer to pin other large, high-security websites. The approach was useful as it helped to identify the DigiNotar attack and the recent CA certificate mis-issuance by TURKTRUST. To address the scalability issues, a new approach to public key pinning is being proposed and is documented in a yet-to-be-published RFC. The goal will have website operators define their public key pins through an HTTP header. The browsers would respect these headers and produce an error when the pin has been violated. The public key pin will be identified by the SHA-1 or SHA-256 algorithm of the key. The public key can be the website certificate, an intermediate CA or the root CA. The methodology pins public keys, rather than entire certificates, to enable operators to generate new certificates containing old public keys. The public key pin will also have directives to define the maximum age of the pin (in seconds), whether it supports subdomains, the URI where to report errors, and the strictness of the pinning. An example pin would look as follows: Public-Key-Pins: pin-sha1=”qvTGHdzF6KLavt4PO0gs2a6pQ00=”; pin-sha256=”LPJNul+wow4m6DsqxbninhsWHlwfp0JecwQzYpOLmCQ=”; max-age=2592000; includeSubDomains Based on the effectiveness of the current public key pins, I believe the RFC will be finalized and will be supported by all the mainstream browsers. Will provide more updates as the RFC gets finalized. Update May 1 2015: Public Key Pinning Extension for HTTP is RFC 7469.
<urn:uuid:e5cf9281-d6c2-4f4d-8004-d33db99aa33c>
CC-MAIN-2022-40
https://www.entrust.com/de/blog/2013/01/public-key-pinning-extension-for-http/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337971.74/warc/CC-MAIN-20221007045521-20221007075521-00081.warc.gz
en
0.923177
440
2.609375
3
The Zero Trust Network, or Zero Trust Architecture, model was created in 2010 by Forrester Research principal analyst John Kindervag. Now, more than a decade later, CIOs, CISOs, and other corporate executives are implementing Zero Trust as the technologies that support it move into the mainstream and cyber attacks become more sophisticated. Zero Trust is a complete departure from the castle-and-moat concept, where implicit trust was the norm and networks were protected by firewalls, VPNs, and web gateways. A Zero Trust model trusts no one and nothing. What is Zero Trust? Zero Trust is an IT security model that requires strict identity verification for every person and device attempting to access resources on a private network. Zero Trust gained relevance in security circles due to the sudden rush to remote work in 2020, which exposed the flaws of the implicit trust model. It became clear that hijacking users via employee VPNs was the key to entering the firewall. Zero Trust helps with compliance auditing and offers better insight into networks. The model also utilizes microsegmentation — the practice of breaking up security perimeters into small zones to maintain separate access for separate parts of the network. This prevents an intruder from gaining access to everything. The biggest challenge to adopting this security model is legacy. Older authentication protocols, tools, apps, and other resources may be difficult to integrate into a Zero Trust architecture. Organizations no longer have their data in just one place; information is often spread across multiple locations and devices using the cloud. That’s why they’re being pushed to replace their legacy systems with more robust and comprehensive security. New Thinking With Today’s Tools Zero Trust isn’t a product. It’s a holistic approach to network security that incorporates several principles and technologies. The main principles behind Zero Trust security include continuous monitoring and validation, least-privilege access, device access control, microsegmentation, and multi-factor authentication. Zero Trust security means that no one is trusted by default from inside or outside the organization, and verification is required from everyone trying to gain access to the network. We need a cultural pivot — a paradigm shift in how we think about cyber security. A Zero Trust model is dynamic and constantly changing. After your system verifies the user and device and assures minimum access, it’s vital to monitor, learn, and adapt. That means Zero Trust is a growing, adaptable process. Trust No One When it comes to Zero Trust, employees, user devices, data sources, and services all have the same status — the system does not let them in by default. Instead of a set-it-and-forget-it method, authentication should be applied each time a new access request is made. Real-time visibility into user IDs, device behavior, device credential privileges, device location, app update status, and other attributes is key. Additionally, the Zero Trust mindset reduces the role of the perimeter. In the old castle-and-moat model, organizations focused on defending their perimeters while assuming everything already inside didn’t pose a threat and therefore was cleared for access. Today’s IT departments require a new way of thinking because, for the most part, the castle no longer exists in isolation as it once did. Although it may seem like a big change, it’s just a matter of using today’s tools differently. Implement Zero Trust Security with Mindcore Mindcore provides expert cyber security services for companies in New Jersey and Florida, including the implementation of Zero Trust security. After a thorough evaluation, we will work with you to find the most cost-effective and reliable solutions for your IT infrastructure. Keep your data safe and schedule a consultation with our team today! Learn More About Matt Matt Rosenthal is a technology and business strategist as well as the President of Mindcore, the leading IT solutions provider in New Jersey. Mindcore offers a broad portfolio of IT services and solutions tailored to help businesses take back control of their technology, streamline their business and outperform their competition.Follow Matt on Social Media
<urn:uuid:77cdb97b-2e56-47ef-995d-4247f2d6512d>
CC-MAIN-2022-40
https://mind-core.com/blogs/cybersecurity/zero-trust-follow-a-model-not-a-tool/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334515.14/warc/CC-MAIN-20220925070216-20220925100216-00281.warc.gz
en
0.934489
843
2.734375
3
Why Cybersecurity Awareness is Critical Phishing is the most common cyberattack because it’s easy to perform and it works. If a cyber threat actor sends out enough emails or makes one look sufficiently realistic, then someone within the organization will likely fall for it. Whether clicking a link, opening a malicious attachment, or handing over sensitive information, this accomplishes the attacker’s goals with minimal effort. Phishing and other human-focused attacks are going to continue to target employees within an organization. To protect itself against these threats, organizations need cybersecurity awareness training that educates its employees about these threats and how to respond to suspected attacks. The Common Cyber Threats Employees Need to Know Cyber threat actors can use various techniques to target an organization’s employees and achieve their goals. Some of the leading threats that employees may face include: - Phishing: Phishing is the most common human-focused attack where cyber threat actors attempt to trick, coerce, or bribe the target into taking some action that benefits the attacker. As the targets of these attacks, employees need to know how to identify a phishing attack and how best to respond to it. - Social Engineering: Social engineering attacks go beyond phishing, including shoulder surfing, tailgating, impersonation, and other attacks. Employees need to know why it’s considered bad security to be polite and hold the door open for someone they don’t know. - Ransomware: Ransomware has emerged as a major cyber threat to organizations as cyber threat actors capitalize on the success of attack campaigns and the large profits that ransomware can bring. Ransomware and other malware may be distributed via phishing messages, compromised accounts, and other attacks targeting employees, and an employee’s ability to recognize an attack and respond properly is essential to minimizing its impact on the business. - Account Takeover: With the rise of remote work, cyber threat actors are increasingly taking advantage of compromised accounts and remote access solutions to gain access to corporate networks. Often, these attacks are made possible by employees’ use of weak, reused, or breached passwords, making account security a vital topic for cybersecurity awareness training. - Mobile Devices: Bring-your-own-device (BYOD) policies and remote work mean that devices with access to sensitive corporate data and resources may be used outside of the organization or not controlled by it at all. Employees should be trained on device security best practices to protect against malware infections and other endpoint security threats. Why Implement Cybersecurity Awareness Training for Employees? By implementing a cybersecurity awareness program, an organization informs and educates employees about the cyber threats that they will face. Companies may start cybersecurity awareness training for various reasons, including: - Improved Security: If employees know how to identify and respond to a phishing email or other attack, this reduces the probability that an organization will fall victim to a damaging and expensive attack. - Cyber Risk Visibility: Cybersecurity awareness training provides an organization with a means of measuring its vulnerability to cyberattacks based on how employees respond to training. This risk visibility can help to inform strategic planning and security investments. - Regulatory Compliance: Companies are subject to a growing number of regulations, and security awareness training is a common requirement. Implementing a cybersecurity awareness training program may be essential for compliance with regulatory requirements. Types of Cybersecurity Awareness Programs Organizations can train their employees by implementing different types of cybersecurity awareness programs, including: - Security Awareness Training: General security awareness training provides employees with information about the threats that they might face and how to address them. For example, awareness training may explain best practices for password and mobile device security or how to identify and respond to a suspected ransomware infection. - Phishing Simulations: Hands-on experience is the best way for employees to learn how to identify and respond to certain threats, such as phishing attacks. Phishing simulations emulate the latest phishing threats, giving employees experience in identifying these attacks and providing organizations with the ability to perform more targeted training based on employees’ needs. Cyber Security Awareness Training with Check Point Check Point has deep visibility into the leading attacks and techniques that cyber threat actors use in their attack campaigns. Check Point SmartAwareness uses this insight to develop targeted, personalized anti-phishing and security awareness training that engages employees and delivers relevant training to the employees that need it the most. Learn more about making your security awareness timely, engaging, and relevant with a free SmartAwareness demo.
<urn:uuid:08526456-e30e-405b-ad59-df60abf18209>
CC-MAIN-2022-40
https://www.checkpoint.com/cyber-hub/cyber-security/what-is-cybersecurity/cybersecurity-awareness-training/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334912.28/warc/CC-MAIN-20220926144455-20220926174455-00281.warc.gz
en
0.940289
922
3.046875
3
The rise of mobile and social has seen many benefits in our lives. The ability to send messages quickly to our friends by SMS or Facebook has taken the world by storm. Along with it we’ve seen a rise in shorthand acronyms and abbreviations used get our messages across with the least amount of characters possible. While these acronyms may be accepted (somewhat) when texting or using social media, this language is slowly making its way into our everyday lives—and used everywhere. One issue with that this language is worming its way into the classroom. Kids have become so used to shorthand, that they are forgetting to use proper grammar at school. This habit often carries over to post-secondary, and often their professional lives. Student’s are having a hard time with their English skills in university, and are using “chat speak” when filling out applications or writing resumes. Teachers are having an increasingly hard time trying to instill good writing skills, though thankfully, they remain persistent. Nobody wants to send their kids into the working world with poor writing skills. Perhaps a secondary issue is when “texting talk” becomes a part of our verbal language. In all seriousness, how many times have you let a “LOL” or “WTF” slip in real life? Be honest. One of the joys of texting and social media is that we don’t have to worry about our grammar or spelling, as long as our message is (semi) coherent, but let’s not forget to turn off our “text talk” button once we put our phones away.
<urn:uuid:d0bdcf8d-7421-4efd-946c-ded013a53683>
CC-MAIN-2022-40
https://www.faronics.com/news/blog/text-talk-takes-over
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334912.28/warc/CC-MAIN-20220926144455-20220926174455-00281.warc.gz
en
0.959429
341
2.953125
3
Inflammation, long considered a dangerous contributor to atherosclerosis, actually plays an important role in preventing heart attacks and strokes, new research from the University of Virginia School of Medicine reveals. The work also raises an important caution about a high-profile drug, canakinumab, being tested to treat advanced atherosclerosis, which is the narrowing of arteries due to plaque buildup on the artery walls. The drug would need to be prescribed only to a select group of patients, UVA’s research suggests. Inflammation and Atherosclerosis In medicine, inflammation is often viewed as harmful, something best suppressed with drugs. Health care company Novartis, for example, has recently tested a potent anti-inflammatory drug, canakinumab, in hopes of benefiting patients with advanced atherosclerosis who have previously suffered a heart attack. The new study from the laboratory of UVA’s Gary Owens reveals, however, that inflammation plays a key role in maintaining the stability and strength of atherosclerotic lesions inside the blood vessels. Blocking too much of a key inflammatory mediator resulted in weaker lesions that are more likely to rupture – possibly leading to heart attacks or strokes – because the absence of inflammation leads the body to conclude, wrongly, that the repair has been completed. “We believe that globally suppressing inflammation gives the tissue a false sense of inflammation resolution,” said Ricky Baylis, an M.D./Ph.D. student in the Owens lab and major contributor to the study. “And by removing this key danger signal, you appear to be taking away the ‘good guys’ prematurely.” That means that doctors would need to be highly selective in prescribing canakinumab for atherosclerosis, should it be approved for that purpose by the federal Food and Drug Administration, said Owens, director of UVA’s Robert M. Berne Cardiovascular Research Center. “What our data suggests is that you need to be extremely cautious in starting to give this drug more broadly to lower-risk patients,” he said. “This is not a drug that should be prescribed broadly like statins, because we believe our data suggests that if you suppress inflammatory response without first removing or reducing the cause of the inflammation, which is lipids, necrotic tissue debris and other plaque components, that this could become dangerous and have unintended consequences. … If you give it to the wrong person, it could do the opposite of what you intended.” That’s an important caveat because many doctors are eager for a drug to help patients with advanced atherosclerosis who are at high risk. Novartis’ drug testing, for example, seeks to benefit “a group of patients that, despite our best therapies, still have an elevated risk for major cardiovascular events,” Owens said. “There is really no drug for them to take.” Creating Safer Drugs Owens has an ongoing partnership with Novartis, and he noted that the collaboration of companies such as Novartis with scientists at academic universities helps make for safer, more effective medicines. “I hope people will look at the Canakinumab Anti-Inflammatory Thrombosis Outcomes Study] clinical trial and our study and find encouragement in the fact that there are scientists in the lab working day and night for years trying to better understand the effects of this intervention strategy,” he said. “Our mouse studies allow us to generate hypotheses that can then be tested in humans, to gain a better understanding of what happens when we give these types of drugs. And it’s likely that with that knowledge, we’ll be able to better design drugs that are more effective and safer by targeting the bad parts of inflammation, but retaining or enhancing the good parts of inflammation that increase the stability of atherosclerotic lesions.” Rapidly Changing Lesions Suggests Both Promise and Peril In addition to a more nuanced understanding of inflammation in atherosclerosis, the study suggests that our lesions are more susceptible to change – both for better or for worse. The traditional view of atherosclerotic lesions is that they’re dormant – that they remain largely unchanged after the body creates them to seal off accumulations of dangerous material inside the blood vessels. In that sense, the fibrous caps covering the lesions have been viewed like patches on a tire. But UVA’s study finds that the caps change significantly over time and can change quite quickly. This became apparent when treatment with the anti-inflammatory therapy rapidly reduced fibrous cap integrity. The finding suggests that doctors and patients have a much greater opportunity to strengthen the caps to prevent heart attacks and strokes. “This study seems to indicate that the fibrous cap, as a structure, is actually much more plastic than previously thought,” Baylis said. “This can be seen as an obstacle, but also an opportunity, suggesting that proper treatment and lifestyle changes can rapidly stabilize risky lesions, but also that poor management – even for short periods – may have the opposite effect.” The researchers have published their findings in the scientific journal Nature Medicine. More information: Delphine Gomez et al. Interleukin-1β has atheroprotective effects in advanced atherosclerotic lesions of mice, Nature Medicine (2018). DOI: 10.1038/s41591-018-0124-5 Journal reference: Nature Medicine Provided by: University of Virginia
<urn:uuid:50973aa4-7db3-4882-b2d7-aa2787498c97>
CC-MAIN-2022-40
https://debuglies.com/2018/09/19/inflammation-plays-an-important-role-in-preventing-heart-attacks-and-strokes/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335059.31/warc/CC-MAIN-20220927225413-20220928015413-00281.warc.gz
en
0.94103
1,152
3.140625
3
Why is DRAM confined in a 10nm semiconductor process prison when microprocessors and the like are being built using 7nm processes, with 5nm on the horizon? If DRAM could be fabricated with a 7nm process, costs per GB would go down. However, for the next few years 7nm DRAM is fantasy, due to capacitor and other electrical limitations at the sub-10nm level. DRAM process shrink progression DRAM is more expensive and more tricky to manufacture than processor silicon, due to its critical nature. It has to hold data over many logic clock cycles. So it will lag in fabrication processes. Process size shrinkage is the key to lower DRAM costs. Broadly speaking, the industry standard 300mm semiconductor wafer has a fixed cost. Therefore the more chips you can fit on the wafer the lower the cost per chip. DRAM process sizes shrank significantly in recent years, until 2016. - 2008 – 40nm-class – meaning 49nm to 40nm and classed as 4x - 2010 – 30nm-class – 39nm – 30nm – or 3x - 2011 – 20nm-class – 29nm – 20nm – or 2x - 2016 – 10nm-class – 19nm – 10nm – or 1x Today, vendors are still shipping at the 1xnm node level. There are three sub-levels and the industry refers to them as: - 1xnm – 19nm – 17nm (Gen1) - 1ynm – 16nm – 14nm (Gen 2) - 1znm – 13nm – 11nm (Gen 3) Next, and in R&D, vendors have three more scaled generations of DRAM on the roadmap, all still at the 1xnm node level. Those are called: - 1anm (Gen 4) - 1bnm (Gen 5) - 1cnm (Gen 6) Each progression should involve some decrease in DRAM cell size in some dimension to increase density, lower power, etc. The 1a DRAM chips are slated to ship in 2021 or possibly sooner. DRAM cell sizes are measured using an nF² formula where n is a constant, derived from the cell design, typically between 6 and 8, and F is the feature size of the process technology. Thus, with a 16nm feature size and n=8, then the area is 8 x (16 x 16) = 2,048 square nanometres. Use a 14nm feature size instead and the area is 1,568 square nanometres, 23 per cent less. This is why it is worth shrinking the feature size. You get more GB of DRAM on a wafer that way. Billions of cells can fit on a single chip, with thousands of cells in a row. Smaller cells can mean billions more in a chip. Overall we have six 1x-class generations and cannot breach the 10nm process barrier. Why? Capacitor aspect ratio The capacitor in a DRAM cell needs to be large enough to store a measurable charge. Shrinking DRAM cell size laterally, by its length and width, decreases the capacitor’s volume, which is, to be blunt, bad: it reduces its effectiveness. This reduction in volume can be compensated for by increasing the capacitor depth or height. Debra Bell, senior director of DRAM Product engineering at Micron, wrote in an article: “In cell capacitor scaling, the aspect ratio is a challenge.” That’s the ratio between the height and lateral size of the capacitor structure. Make one or both too small and the capacitor cannot do its job effectively. Increasing the capacitor’s depth has its own problems, as an IBM Zurich research blog stated: “In the long term, this represents a bottleneck – not only due to geometrical constraints, but also because charge accumulation at the top of the ‘well’ makes it more challenging to use the entire storage capacity.” Another problem is that capacitors leak charge, which is why they have to be refreshed periodically: this is why it is called dynamic random address memory or DRAM, as opposed to static memory, which doesn’t need constant charge refreshing, as data is stored in logic gates but is more expensive. DRAM researchers are looking at getting rid of the capacitor, we hear, by storing the charge in the transistor body using different transistor materials. Another difficulty is that as cell size decreases and fabricators cram more cells into an array, the relative length of the word and bit lines increases. This affects the time it takes to put a charge in a capacitor and move the charge along the lines. The net result is that decreasing DRAM cell size beyond the limits met at the 1x nanometer processes is impractical to impossible in the short-term, certainly out to 2025 at earliest. It looks increasingly as if there will have to be some breakthrough in materials to enable us to make the jump into sub-10nm DRAM. To understand why this is the case let’s take a brief tour of kinds of semiconductor chip manufacturing and DRAM design, where we’ll see the key role played by capacitors. Under the hood Random-access memory cells are each made up of tiny circuits, typically involving one transistor that acts as an access gateway, and one capacitor that functions as a charge store. In semiconductor jargon, this design is termed ‘1T1C’ and shown below. Each circuit is addressable using so-called address strobes. This makes them usable in rows and columns of RAM, called arrays. And each 1T1C circuit in this array is a cell. The cells line up in rows with a bit line connector. Columns of these cells are connected by word lines. A cell’s address is defined by its location at the intersection of these two orthogonal bit lines and word lines. The address strobes do their work along the word and bit lines as shown in the diagram below. That’s for memory cells. But logic chips, built out of high-density logic gates, are entirely different. They are super-high density. They are incomparable to 1T1C repetitive building blocks, with relatively large capacitors in them, which is why they can scale down past the 10nm barrier.
<urn:uuid:89162173-791b-47a6-b716-8c81aa48b5cb>
CC-MAIN-2022-40
https://blocksandfiles.com/2020/04/13/dram-is-stuck-in-a-10nm-process-trap/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335326.48/warc/CC-MAIN-20220929065206-20220929095206-00281.warc.gz
en
0.937924
1,332
2.765625
3
Amazon made news recently when details of its proposed emotion-sensing wearable device came to light. Interestingly, the device is being developed by the Alexa team who is using voice-recognition technology to detect the emotional state of the consumer. This is not the first-time voice-recognition is being used to diagnose a medical condition. Studies have established the use of voice in detecting early stages of Parkinson’s disease and it is logical for the same technologies to be used to detect and treat a range of conditions. Voice is potentially the next wave in the UI and technology platform shift, after the web and mobile waves, in the past two decades. Indeed, voice may be the next battleground for big tech firms as consumers increasingly use one of the dominant voice interfaces to access information, much like how Google’s text-based search engines dominated the consumer’s attention for years. Amazon, with reportedly over 10,000 associates on its Alexa team, is not alone in turning voice-recognition technology towards solving healthcare problems. Besides Amazon, Google (Google Assistant), Apple (Siri), and Microsoft (Cortana) are also investing billions in voice-based personal assistants to gain a dominant position in the market (or at least not be left behind). However, consumer-grade voice-recognition services such as asking for a restaurant recommendation does not translate to deployment of the technology in a healthcare context. Healthcare is bound by HIPAA data privacy rules which govern what information can be shared, with whom, and how. Amazon recently released a set of Alexa “skills” that transmit and receive protected health information and is running pilots with a handful of health systems. While many of the skills are relatively mundane services, such as status updates for prescription refills,these pilot projects set up a test bed for secure communication for a range of complex care management protocols down the road. Dwight Raum, CTO of Johns Hopkins Medicine, says that voice is the future in healthcare delivery, and believes that “Voice is the most obvious next step of user interface that is going to radically change the way we interact with technology.” Let us look at a few scenarios of how this can play out in healthcare: Access to healthcare services We are fast approaching a time when consumers will want their healthcare delivered at a time, place, and manner of their choosing (in some cases, even before they realize they need it, if AI algorithms have their way). North Carolina based Atrium Health’s Alexa pilot is using voice recognition to help customers identify a nearby urgent care center and get a same day appointment. This will significantly improve access and drive consumer satisfaction, not to mention the bottom-line impact from increased revenues for the health system. There have been ample studies about how the advent of electronic health record (EHR) systems and technology in general have increased the workload for physicians and other caregivers. We are at a point where the role of the next wave of technologies is expected to reduce the burden of caregivers, and voice fits neatly into the picture. John Kravitz, CIO of Geisinger Health System, believes that “it is one of those areas where, in our hospital setting, we want to have the ability and hopefully take some of the stress off our nurses, where a patient can speak to an Alexa or Google Home type voice- enabled device. Nurses should and be able to interact with the devices and hopefully be able to serve our customer population more effectively.” Delivering superior experiences is one of the primary goals of any digital transformation program. As we enter the era of zero-UI technology where a touchscreen is replaced by a natural language interface such as voice, technology starts to become less front and center and dissolves more into the background of our everyday experience. The Alexa skills described earlier will achieve that in the near term for a range of simple tasks. However, over time the software will significantly improve customer experiences. As the voice-recognition software becomes more sophisticated, with the help of AI and machine learning, it will adjust to accents and a broader range of terminologies. The latter aspect becomes critical in a medical context as clinical terminology is progressively built into the lexicon of the interface. If there is a major roadblock to increase the use of voice-enabled services, it would be concerns around privacy, especially if the data on voice-based interactions is stored in a cloud by one of the big tech firms. Concerns about Amazon’s Alexa “snooping” on consumers is also likely to give pause to efforts to expand on the use of the technology in general until sufficient privacy safeguards are established. Over time these concerns will be addressed, much as concerns about patient data storage in cloud infrastructure is no longer an issue today. However, the key is that voice is already a part of consumers’ lives for a variety of other services, such as banking and retailing. If there is one thing to say about the future of voice-enabled services in healthcare, it is this: you ain’t heard nothin’ yet.
<urn:uuid:5f13e12c-a553-43e4-aa97-2766cd8f60b0>
CC-MAIN-2022-40
https://www.cio.com/article/220167/voice-enablement-the-next-big-wave-in-healthcares-digital-transformation-journey.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335326.48/warc/CC-MAIN-20220929065206-20220929095206-00281.warc.gz
en
0.950782
1,049
2.53125
3
Have you ever come across an Excel file with a bunch of formulas you can’t make any sense of? I do all the time. Not only do I come across these files often, I’m also the one that makes them! Lots of them. For anyone who has ever programmed before, the ability to add comments is vital. There is no easy way to comment a formula and Excel doesn’t have a way to do this. Let’s look at a number of ways we can bring clarity to the formulas we use in Excel. 1. One of the most obvious ways to comment anything in excel (including formulas) is to just add a comment. You can do this from the review tab, or by pressing SHIFT+F2. Once the comment window pops up, you can write a novel about the formula. 2. You can get away with no comments at all if your formulas are easier to read. It helps if you break your formula line by line, which you can do by pressing ALT+Enter on Windows or Control + Option + Enter if you are on a Mac. 3. If the end result of your formula is a numeric value, you can use a nifty hack with the N function. The N function converts any non-numeric value to a number so we can use this formula to embed a comment. Whatever the comment is, it will turn into a zero value, thus not affecting any calcs. To add a comment, just add +N(“You comment goes here”) to your formula wherever you please. However, if your formula returns a string, this won’t work and the end result will be an error with #VALUE!.4. It also helps to have the full formula visible in a cell. We can get it done by adding ‘ to the formula. I usually have two columns running side by side, the first with a formula that calculates the values, and the second with the ‘ character, so the entire formula is shown as text. I hope this helps, happy excelling!
<urn:uuid:d15b2f2c-2fb7-4c3e-b0ed-bf9023580e47>
CC-MAIN-2022-40
https://www.faronics.com/news/blog/how-to-add-comments-to-a-formula-in-excel
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335326.48/warc/CC-MAIN-20220929065206-20220929095206-00281.warc.gz
en
0.910223
432
2.953125
3
The most significant change to U.S. patent law since 1836 — or perhaps 1790 — is being implemented on March 16, 2013. Part of the America Invents Act of 2011, it concerns the doctrine of first-to-invent, laws concerning the protection of original inventors, regardless of whether they were the first to apply for a patent. The new law awards inventorship under a first-inventor-to-file standard. The emphasis is now placed on inventors seeking immediate patent protections, preferably before any disclosures to any third parties and patent filings of others. Under the soon-to-be old law, such disclosures were somewhat protected and true inventorship ascertained and rewarded; however, those protections have almost all been eliminated. Now, there are greater hazards to inventors — particularly small inventors — which will possibly preempt them from obtaining a patent entirely. Philosophical Break From Past The so-called grace period of one year from a public disclosure of an invention to patent filing under the old law is technically still there. However, third-party patent filings during that grace period now trump the earlier inventor merely by reaching the Patent Office first. This change in the law tries to objectify determination of a true inventor instead of engaging in procedural challenges called “interferences,” which it eliminates. Many view this substantive change as contrary to the philosophy of the U.S. patent system, which rewards true and original inventors, i.e., the first to invent. The new law is patterned on that of foreign patent systems, which place much less emphasis on the individual in favor of corporations, which want more certainty in the patenting process. Since some third-party patentees may derive their “inventions” from primary inventors who are second to file for patent, new proceedings determine inventorship under these circumstances. These derivation proceedings are mini-trials, with evidence, to ascertain whether the earlier inventor was primary and whether the later filer gleaned the invention by theft or other illicit means. The U.S. Patent & Trademark Office over the past year or so has been hiring and training hundreds of Patent Office judges for the purpose of hearing these proceedings and administrating many other new proceedings already implemented under the Act, e.g., various post-grant opposition actions to challenge newly issued and all other patents. First to File For inventors, although there are ambiguities in the new law, there are some clear lessons. First, “first to file” means just that, with the consequence of possible automatic curtailment of patent rights for many second filers. For example, an individual inventor is typically unaware of the extremities of the patent laws with regard to deadlines. Publication or dissemination of the invention prior to filing for patent already kills virtually all patent rights outside the United States. Often, these inventors are well into the grace period when they consult a patent attorney. Now, those same inventors may have unwittingly destroyed their opportunity for patenting by delaying the filing. Fortunately, provisional or informal patent filing is available, which can preserve patent rights, but the first-to-file rule still applies. Unfortunately, those rushing to file informal and often sketchy documents may find their patent descriptions challenged for insufficiency, i.e., the details may be so bare as to raise a question of whether the inventor was in possession of the invention claimed. The most dire potential consequence of the new law is no patent — no protection. The countermeasure, of course, is education and consultation with a patent attorney early in the innovation process — and definitely prior to third-party disclosures. Great minds often think alike, and others are also addressing the same technological problems. However, putting inventive thoughts to paper is often a difficult thing for some inventors. Since a patent is a property right, its contours must be set forth in detail. For example, in land purchases, the delineations of the property line are critical. So too with patenting — the patent claims carve out a portion of new technology, separate from what is known. A patent attorney can help navigate this terrain and stake out a claim for particular technological knowledge. For the general public out there inventing, this article and other measures hopefully will make this change known. It is very important to be first to file a patent application under a post-March 16 first-inventor-to-file standard. Sadly, the actual contours of this change in the law will not be known for many years — that is, after the courts make pronouncements — but it is the law. Since the new standard is draconian, it behooves us all to educate ourselves as to this imminent change in the patent laws.
<urn:uuid:95e0dd32-99ff-489d-a4e8-aed030dcb9a4>
CC-MAIN-2022-40
https://www.ecommercetimes.com/story/caveat-inventor-the-new-patent-paradigm-77541.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335491.4/warc/CC-MAIN-20220930145518-20220930175518-00281.warc.gz
en
0.946719
984
2.78125
3
Artificial intelligence and machine learning get lumped together so often these days that it’d be easy for people to mistake them as synonymous. That’s not quite accurate, though: They’re most certainly connected but not actually interchangeable. “Artificial intelligence and machine learning are closely related, so it’s no surprise that the terms are used loosely and interchangeably,” says Bill Brock, VP of engineering at Very. If you’re not using AI or ML yet, you soon will be evaluating its potential for your organization. “AI as a workload is going to become the primary driver for IT strategy,” Daniel Riek, senior director, AI, Office of the CTO, Red Hat, recently told us. “Artificial intelligence represents a transformational development for the IT industry: Customers across all verticals are increasingly focusing on intelligent applications to enable their business with AI. This applies to any workflow implemented in software – not only across the traditional business side of enterprises, but also in research, production processes, and increasingly the products themselves. The improved scale of automation achievable with AI will quickly become critical for a company’s competitiveness building and will make AI a strategy-defining technology.” Advancements in natural language processing and other AI-enabled capabilities help organizations rethink customer service chat and analyze large pools of unstructured data. That will enable more predictive analytics, drive increased efficiency, and enhance decision-making. [ Is RPA a form of AI? Learn the differences: How to explain Robotic Process Automation (RPA) in plain English. ] So what’s the difference between AI and ML? Let’s start by defining the terms. What does AI mean? “AI, simply stated, is the concept of machines being able to perform tasks that seemingly require human intelligence,” Brock says. “This involves giving computers access to a trove of data and letting them learn for themselves.” Machine learning is a specific application or discipline of AI – but not the only one. In machine learning, Brock explains, “algorithms are fed data and asked to process it without specific programming. Machine learning algorithms, like humans, learn from their errors to improve performance.” As a starting point for distinguishing AI and machine learning, it’s helpful to think of AI as the higher-level or umbrella category that encompasses multiple specific technologies or disciplines, and machine learning is one of them. “AI includes various fields of study including ML, NLP (natural language processing), voice/audio recognition, computer vision/image recognition, search, routing, autonomous robots, autonomous transport, [and other disciplines,]” says Mahi de Silva, CEO and co-founder of Amplify.ai. Speaking of umbrellas, Michael McCourt, research engineer at SigOpt, offers a distinction-by-comparison for a rainy day: “Machine learning is like a spoke running out of the artificial intelligence umbrella, with a much more specific definition.” Let’s back up for a second: McCourt notes that AI by definition is very broad – it’s the umbrella – so much so that if you ask a group of ten people to give their definition, you’ll likely get ten different answers. “Artificial intelligence is an umbrella term without a concrete definition, as it encompasses all mechanical, robotic, and automotive tasks that emulate human capabilities,” McCourt says. Moreover, AI’s definition has changed, and it will continue to change over time: “Twenty years ago, tools like spellcheck were considered artificial intelligence,” McCourt notes. “Ten years ago, artificial intelligence meant being able to classify images.” What does machine learning mean? While machine learning technologies and uses might evolve, the core definition is much more concrete and specific. “Machine learning models generate findings based on stored data sets and queries for the purpose of learning a specific pattern,” McCourt says. “If the answer is not previously stored, machine learning analyzes the environment to present its best guess as to what the correct response might be.” Tom Wilde, CEO at Indico Data Solutions, points out that there’s a very current reason that AI and machine learning get used and confused in tandem. “The reason for confusion is understandable: ML can be considered as the current ‘state of the art’ of AI,” Wilde says. Spell-check aside, he adds, machine learning is one of the oldest and best-established AI disciplines. It’s also the one bearing the most current fruit in terms of enterprise use cases. Understanding the difference between AI and ML isn’t just a matter of clarifying terms or relieving annoyance with non-technical folks who just don’t get it. Rather, it’s table-stakes for success with AI projects. “It’s important to distinguish between AI and machine learning, as this is critical to successfully designing, building, developing, and maintaining an application or platform,” Brock says. That’s true for your in-house knowledge and AI skills development; it’s also true for evaluating and selecting the right vendors. Remember when every product suddenly had the word “cloud” added to its name? You may see some of that with AI and ML, too. Subscribe to our weekly newsletter. Keep up with the latest advice and insights from CIOs and IT leaders. While I can appreciate some of the benefits of AI and ML? I feel we have a long way to go towards ensuring the safety of humans before this is adopted in a wide-spread, dare I say, global scale. And don't get me wrong, I'm all FOR AI and ML when it comes to things like logistics for companies and the transport of their goods, and I think it would do well as a security device/appliance that just sits on the edge of the network and LITERALLY asks people for ID before allowing passage onto the domain. But there are things I'm never going to be comfortable with AI/ML handling....things like my personal safety when it comes to travel. Sorry, I know there are a lot of driver-less car enthusiasts out there, but I prefer the fallible and most loved version...of driving myself and my son around wherever we need to go. Yeah...its safe, and there are Uber and Lyft cars that are driverless (at least there are for Uber...not sure about Lyft!) But there was a report of someone getting run over by a driver-less car with the driver IN the vehicle. Imagine what would be the fallout had there been no one in it! And to magnify this scenario?...what happens when an 18 wheeled- truck goes "rogue" and stops responding to its program?...or what about pilot-less planes?.....Yeah I know it sounds science fiction and all, but we already HAVE pilotless drones, so whats to stop some odd ball company from being the first to market there as well? (Just know that I will NEVER get on one of those!.....Ever.) But the progression to more intelligent AI is definitely a wanted thing where it needs to be (cancer research, massive data collection analysis, space exploration...and the more mundane things like monitoring traffic, or keeping a constant eye on schoolgroud perimeters for "abnormal" behavior from either student or staff. (And why AREN'T we working on that kind of tech?....aren't the children our most IMPORTANT resource?) Anyway, just wanted to vent a bit!
<urn:uuid:3073b9ab-df36-4d5f-9203-9c3b445f52b9>
CC-MAIN-2022-40
https://enterprisersproject.com/article/2019/3/ai-vs-machine-learning-what-s-difference
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337404.30/warc/CC-MAIN-20221003070342-20221003100342-00281.warc.gz
en
0.942782
1,611
2.828125
3
Purchasing a new computer is a major commitment and you should do your research, but the same is true when you decide to custom build a PC. Whether you decide to buy a pre-built computer or buy components and build one yourself, you are making a serious investment. Pre-built computers are great because they’re ready to be used right away, however, they usually have baseline components and flaws that might limit their performance later. If you want to ensure the quality of your PC, you should custom-build one. There are downsides to building your own computer though. Some computer components are costly and factoring in your time and human error can definitely increase your expenses. So how can you build yourself an affordable and dependable PC? Let’s explore. Carefully Consider Your Storage Size Storage size is the maximum number of programs, files, and other types of media you can have on your computer. It’s commonly expressed in gigabytes or terabytes. If you have a greater number of GBs and TBs, you will have more storage space available on your computer. The hard drive (HDD or SSD) is the component in charge of storing files. This component can be easily replaced and upgraded, so it’s not a major concern when buying your device. There are also other innovative options now available, such as the cloud or external hard drives. Look into the CPU Power Your PC relies on the power of your CPU or processor. You can evaluate processor power by checking the clock speed, which is expressed in Gigahertz. In simple terms, the greater this number is, the more powerful the processor. When you check the market, the majority of current CPUs run at 3-4GHz. Keep in mind that it’s not always easy to upgrade your CPU. For this reason, it’s necessary to stick to a high-performance product that should last you for years before getting a new upgrade. Determine the GPU Power The GPU is the graphics card. It is just as important as the CPU in determining how powerful your PC will be. Clock speed is usually used to determine the GPU performance. This is why it’s vital to complement your CPU and GPU to make the most of each other’s functions. Another important measure is the VRAM. This is the amount of memory available to the GPU. As the number of VRAM increases, so does the GPU’s power. However, you must be cautious about pushing your GPU or CPU to its limits. This can lead to overheating and power issues. Make Sure Your RAM Is Enough for Your Needs Your PC’s memory, often known as RAM, is the component that stores all of the data that your computer is actively using. It’s also responsible for the number of tasks your computer can perform simultaneously. Often enough, you can easily increase the RAM in your computer, but more and more some computers do not have an option to swap out RAM. When you finish building a new PC, it can be one of the most satisfying emotions in the world. After hours of research and development, you will have a powerful computer that will function for a long time. While it can be somewhat stressful, there are numerous tools available that make building a PC possible for everybody. Get the best custom PC builder with J-BIT TECH. We are a PC and MAC Repair service company in Albuquerque, New Mexico. We provide expert assistance with all your custom PC needs. Book an appointment today to find out how we can help you.
<urn:uuid:24733736-c548-46d9-b344-d9acbf91393b>
CC-MAIN-2022-40
https://www.jbitabq.com/single-post/considerations-when-custom-building-a-pc
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337404.30/warc/CC-MAIN-20221003070342-20221003100342-00281.warc.gz
en
0.941952
742
2.515625
3
The Ultimate Guide to Endpoint detection and response Also known as EDR. End points are the pointy end of the network and are often quite literally found in the user’s hands – as a laptop, smartphone, or other access device. Endpoints can also include IoT sensors or any other connected item, including printers and other peripherals. Endpoint Detection and Response is the cybersecurity technology responsible for monitoring, detecting, and addressing malware at device level.
<urn:uuid:0cb6a547-89da-4895-b537-76254c82b335>
CC-MAIN-2022-40
https://securitybrief.asia/tag/endpoint-detection-and-response
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334528.24/warc/CC-MAIN-20220925101046-20220925131046-00481.warc.gz
en
0.95017
94
2.53125
3
What Is a Public Safety Alert? A public safety alert is a cautionary signal from the state to the general public about potential danger. These alerts are issued when a public safety concern might result in severe or fatal harm without prompt action. Typically, a public safety alert is broadcasted to raise the degree of protection in hazardous situations, and interested parties must take action within a reasonable timeframe. The government leverages commercial or private infrastructure designed to rapidly or continuously disseminate alerts to the public. The simplest of these systems were sirens that would ring in the event of danger and alert the public to prepare for any risky scenario. As technology evolved, those sirens were replaced with notifications disseminated electronically via social media, mobile devices, and other means. What Is a Public Alert System? Major Public Alert Systems Wireless Emergency Alerts Wireless Emergency Alerts (WEAs) are transmitted to wireless devices based on location and relevance. These warnings are issued by the Federal Emergency Management Agency (FEMA). Examples of wireless emergency alerts: - Amber Alerts (for missing children) - Silver Alerts (for missing adults) Emergency Alert Systems Opt-In Alert Systems What is a Public Safety Alert on the iPhone? How to Choose an Emergency Notification System The public safety management team must ensure everyone has the most up-to-date information during an emergency. Depending on an emergency’s level of urgency and potential danger, ENSs must provide information quickly. Here are some essential criteria to consider when choosing an ENS for your organization. Speed and System Capacity Public Contact Management Broadcast Messaging Option Emergencies and business disruptions happen. It’s how you prepare and respond to them that determines their impact. Our critical event management solutions combine a secure emergency notification system with incident response tools and capabilities—so you can quickly deploy your response teams and enable them to better prepare for, respond to, and recover from critical events faster. Critical Event and Emergency Planning Resources
<urn:uuid:2523250e-2d01-4212-957a-5d812d40b17f>
CC-MAIN-2022-40
https://www.blackberry.com/us/en/solutions/critical-event-management/public-safety-alerts
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334915.59/warc/CC-MAIN-20220926175816-20220926205816-00481.warc.gz
en
0.913466
418
3.546875
4
Security experts warn us not to use the same password for multiple accounts, since if one gets hacked, they’re all hacked. But we do because it’s easier to remember one than many. They also warn us to make them more complex than 123456 (the most commonly used password, by the way) so hackers can’t crack them in a second. Yet it’s amazing how many people use “password” or their username as their password. If you do a Google search on “password cracker”, it comes up with almost 10,000,000 hits; Bing has about 1.5 million. There are thousands of tools and scripts out there to automate cracking a password that run on Windows, Linux, MacOS and even DOS. Many of these are billed as tools to help you recover a forgotten password, but many are just hacking tools in disguise. So why do people crack and steal passwords? Simple. It’s makes them money. People do it, because it’s big business. There are three common ways for someone to crack a password. - Dictionary attacks – an automated tool will search through common words or phrases that are likely to be used as passwords. Looking for words like “password” and “123456” are simple examples. - Social engineering – using something about you to guess your password. If someone finds information about you on Facebook or another site, they will try your dog’s name, birthdate or other personal information, since those are common passwords. - Brute force attack – trying every possible combination of letters, numbers and symbols until the password is found. Automated tools can do this faster than any person typing on a keyboard. According to research studies done on bad password practices, about 42% of passwords only use lowercase letters. 37% use numbers and letters and about 4% had special characters in them. So what do you think the hackers will try? Steve Gibson, noted security expert and co-host on the Security Now podcast, has come up with a great tool and webpage to explain how to develop and remember a better password. The Password Haystacks tool shows how you and I can create passwords that are easier to remember and harder to crack. Steve recommends using what he calls padding to turn an easy password into one that’s hard to crack. The calculator shows how much more time a brute force attack will take as you make your password longer and a bit more complex. As the page notes through an example, the calculator is not a password strength meter, because even though “123456” may have over 1,000,000 possible guesses, a dictionary attack would crack it almost instantly. Brute force attacks search through the letters of an alphabet, numbers and special characters or symbols that one could type. If you used the roman alphabet or ASCII printable characters, you would have 95 choices of things to type. 26 lowercase letters + 26 uppercase letters + 10 numbers + 33 symbols = 95 characters If you added Arabic, Chinese, Russian, Greek and other character sets, the attack space would grow very large, but many sites only using ASCII characters. If a website does not use Unicode as its coding system, there may be difficulties dealing with non-ASCII characters. The best passwords should have a combination of lowercase letters, uppercase letters, numbers and symbols. According to Gibson, once you satisfy this rule, password length is what matters. This makes sense, since each time you add a character, the complexity increases by an order of magnitude. Here’s an example: If I use “password”, it would take 2 seconds to crack, assuming 100 billion guesses per second; the latest Intel Core i7 chip can achieve over 100 billion operations per second (GFLOPS). If I capitalize the “P”, then guessing time goes up to 9 minutes. Here’s what happens if I continue to add characters based on the rules above: - Password1 – 1.59 days - Password1$ – 19 years - Password1$] – 18 centuries - Password1$]] – 1,740 centuries - Password1$]]] – 165,000 centuries Now this is solely based on a brute force attack with no dictionary or other logic. This exercise shows that just adding different combinations of characters and making a password longer, makes it that much harder to crack. The important thing to take away is that you can create a password with something easy to remember and then pad it with characters that make it hard to guess with a password cracker. Try something as simple as iLovemywife<2> or ilovemy3Kids;;. According to the calculator, it would take 15,670 centuries to crack using a system that had one hundred trillion guesses per second. That might be possible using a large array of cloud-based computing systems or a super computer. Even with more hackers using cloud-based systems to crack passwords, that is still a pretty strong password. I look at password creation the same way I look at locking my front door. If I don’t lock the door, anyone can break into my house. If I lock it, I’ll deter the large majority of people. If I use some of these simple ideas to create a password that I can remember with a little padding, I will deter most of the hackers out there. Photo credit Sketcher1Jedi
<urn:uuid:60bee1cc-96e9-4e4a-ad29-ba1b002de268>
CC-MAIN-2022-40
https://en.fasoo.com/blog/finding-a-password-in-a-haystack/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335350.36/warc/CC-MAIN-20220929100506-20220929130506-00481.warc.gz
en
0.927179
1,141
3.0625
3
Americans widely acknowledge that the amount of food wasted in the country is a problem. Researchers from the University of Missouri, however, suggested that consumers should pay particular attention to the amount of meat that they throw away. Bioengineering professor Christine Costello and a research team evaluated food waste from four all-you-can-eat restaurants and analyzed the greenhouse gas emissions for each food category. Although fruits and vegetables comprise a larger percentage of the country's wasted food, meat required more emissions when researchers accounted for fertilizer use, utility demands and transportation needs. "If consumers choose to prepare extra food 'just in case,' they should use plant-based foods," said engineering professor Ronald McGarvey, a study co-author. McGarvey hopes to develop future techniques that improve both food production and consumers' ordering decisions in an effort to curb food waste. A June study from Johns Hopkins University valued the amount of food wasted by Americans at $161.6 billion annually.
<urn:uuid:e941104d-18ed-4e06-a111-022dd25154ee>
CC-MAIN-2022-40
https://www.mbtmag.com/global/news/13213406/study-wasted-meat-accounts-for-more-emissions-than-vegetable-waste
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335504.37/warc/CC-MAIN-20220930212504-20221001002504-00681.warc.gz
en
0.938032
198
3.15625
3
Head transplant is a surgical operation which involves the point of insertion of a one organism’s head onto the body of another, including brain. In 1970, a first head transplant successfully moves out on a monkey, but unfortunately it diesafter the head reject by the body’s immune system, and another reason is technology. Lately, this transplant seemed totally doubtful. While, testing on nine rats with polyethylene glycol (PEG), eight alive a month after the operation, and they had regained the ability to walk. Recently, a Chinese doctor conducted head transplants on more than 1,000 mice. However, those mice have lived longer than a few minutes. Earlier this year, Italian neurosurgeon Sergio Canaverobelieve it’s possible and announced his plans to perform the first human head transplant. He declares the schedule of first head transplant on 2017. He estimates the procedure of transplant to take 100 surgeons nearby 36 hours long journey to complete the process. The estimation of operation cost is $10million. Challenges of head transplant However, the procedure will involve spinal cord fusion (SCF).The surgery requires the brainfirst cooled down to 12-15˚C to ensure that the cells last longer than a few minutes without oxygen,and not all brains can survive. Combining a spinal cord has never done before, and may not possible. This is probably the main objection for people. Using a sharp blade to cut the tissue around the neck to limit the amount of damage the spinal cord sustains. Canavero suggests that, the spinal cord of the donor body will be fused using a chemical called polyethylene glycol. The chemical promotes nerve growth in the spinal cord, and it could be used as a sort of glue between the body and head. Also,the immune system not occur when the body sees a new part as foreign and attacks. However, all the animals in these experiments died so quickly, and many places the blame at organ rejection. In head, there are more organs to be rejected. Although the technology has come on a long way. Some believe the quantity of anti-rejection medication required would poison the body. Side effects of a human head transplant If the surgery goes ahead as planned. But, the main risk is that if the body will reject the head and the person will die. While, transplanting on animals the longest any has lived is eight days. However, Canavero argues that his methodology is different, but he’s pretty vague on the details. Canavero explains, the patient will be kept in a coma for a month, during the time spinal cord will be subject to electrical stimulation via implanted electrodes in order to boost the new nerve connections. While, a volunteer Valery Spiridonov, a 30-year-old Russian man faces a rare genetic disorder called Werdnig-Hoffmann disease, in which motor neurons in the spinal cord and the brainstem gradually die, resulting paralysis. He was a first person to volunteer for the procedure. He said, “If I don’t try this chance, my fate will be very sad. With every year my state is getting worse.” “I can hardly control my body now. I need help every day, every minute. I am now 30 years old, although people rarely live to more than 20 with this disease.” In 2015, Spiridonov confirmed that private donors had approached him to ensure the surgery goes ahead. Canavero also received several offers. The ethical issues of a human head transplant have taken something of a back seat, but there some definite concerns. The most aspect of this transplanting a whole body to save a single life at great expense is the best use of a cadaver full of smaller organs ripe for transplant. First head transplant As the first head transplant now looks set to take place in China, many have raised concerns that the donor bodies will be provided by recently executed criminals. In the past, “China criticized for using the organs of executed prisoners without their consent.” Some experts have pointed out the proposed surgery, Dr. Christopher Winfree, assistant professor of neurological surgery at Columbia University said, A fancy glue is not going to fix that, even if you could get them physically, nerves themselves don’t grow too much in the spinal cord. Gordon R. Tobin, from University of Louisville said, nerves did not function as we planned, and that was just a few of them. Neurosurgeons just don’t know enough about the brain yet to even know what they need to connect. It’s hard to solve the problem if you don’t even know what that problem is. So, the transplant will take place by the end of 2017. The surgery runs at Harbin Medical University in china. More information: [Alphr]
<urn:uuid:7ef020e4-e916-4de4-ab7a-08429651a080>
CC-MAIN-2022-40
https://areflect.com/2017/11/04/facts-about-the-human-head-transplant-sunday/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337421.33/warc/CC-MAIN-20221003133425-20221003163425-00681.warc.gz
en
0.953513
1,019
3.578125
4
What is a Server Rack Blanking Panel A server rack blanking panel is a piece of plastic or metal that fits into the empty spaces in your rack. Typically they come in 1U or 2U sizes, and they close off that U, only being removed when a new piece of equipment is to be installed. The fact that they will oftentimes end up being removed means that many people question their need as they will be surplus to requirements at a later date. The fact is though, they are an essential piece of equipment, and once the benefits are understood the need for server rack blanking panels will become apparent. Proper Server Rack Airflow Server racks help to direct airflow through the cabinet as the equipment mounted inside draws the air through. When the cold air passes over the hot components in the servers it carries with it the heat which is the expelled through a performed screen door at the rear, or sometimes chimneyed through the top of the rack. The rack is designed to ensure an even flow of air to remove heat as efficiently as possible. However, if there are open bays in the rack, it will disrupt this even distribution of airflow. When this occurs insufficient cold air may pass through the servers. Other results of this would be :- - Dead Spots – When there is an open U in the rack, airflow will pass through the easiest path, the large openings provide an easier route than through the constricted pathway through a server. This results in dead spots where the air is not moving or moves slowly. These are also known as rack hot spots. - Incorrect Airflow Direction – in a containment setup where hot and cold air is segregated into hot and cold aisles, any openings between these aisle can allow the mixing of air. Even without containment, the opening can allow hot air being expelled at the rear of the rack to re-circulate to the front and be sucked back into the server degrading cooling efficiency. Proper pressure differential must be maintained and can be checked with AKCP differential air pressure sensors. - Tornado Effect – When air hits an opening it can spiral and blow in the wrong direction. This spinning air disrupts the cooling effects of the airflow. - Reduction in Efficiency – Having areas of stagnant air and hotspots in your rack decreases the efficiency from the planned capacity. This results in extra loads being placed on cooling equipment and increased power consumption with the associated energy costs. Server Rack Cooling The above points illustrate the importance of utilizing blanking panels as part of your data center plan. Ensuring proper airflow management and cooling of your servers. Correctly utilized, blanking panels mean the air goes exactly where it is supposed to, aiding in efficiencies of cooling and power usage, improving your PUE numbers. The cost of blanking panels is minimal in comparison to the improved efficiency and lower energy costs than going without. Another byproduct is a cleaner rack. The blanking panels prevent dust related issues. Although data center air is filtered, it is impossible to remove every bit of dust. Dust that finds its way into servers can be expelled at the rear of the rack. Blanking panels prevent this dust from being pulled back through the rack to the front and recirculated. Finally, using blanking panels improves the aesthetic look of the data center, making your racks look professional and organized. Blanking Panel Types Blanking Panels can come in 1U or 2U sizes, and are also available with brushes to allow cables to pass through while maintaining the integrity of the blanking panels.
<urn:uuid:b506d03e-8436-409b-b2d7-f831b9a76c27>
CC-MAIN-2022-40
https://www.akcp.com/blog/server-rack-blanking-panels-why-you-need-them/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337529.69/warc/CC-MAIN-20221004215917-20221005005917-00681.warc.gz
en
0.948595
723
2.625
3
A topology is a layout of a network, which determines how the network communicates with different devices. Both the physical and logical structure of nodes and connections in a network are shown in network topology diagram. Network topology software is best used by IT administrators to determine the right layout for each node and aid the traffic flow. A well-designed network topology enables an organization to double down on issues quickly, resolve them, and have the network functioning at the most efficient data transfer rate. Network topologies are classed into physical and logical topologies. Physical topologies showcase the actual physical wired layout, showing where and how exactly the network is connected. Logical topologies show the logical network path along which data travels from one end to the other. Some of the primary network topologies are bus, ring, star and mesh topologies. Star topology: Star topology is the most commonly used topology system. Every node connects to a central network device in this layout, like a hub, switch or computer. Star topology is centralized in nature, making it user-friendly, reliable, and easy to manage. However, star topology involves high costs and requires continuous maintenance. Bus topology: In bus topology, every workstation is connected in a series to the main central cable. The simple linear layout and cost efficiency makes it ideal for small networks. However, bus topology tends to be slow for larger networks, and problem identification is difficult in this topology in the event of a network failure. Ring topology: In a ring topology, network devices are cabled together so that the last network device will be connected to the first. Each device connects to exactly two devices, forming a continuous ring. Ring topology is cost-efficient, with minimum chances of packet collision. However, ring topology is dependent on one cable, difficult to troubleshoot, and expensive to maintain. Mesh topology: In Mesh topology, nodes are linked with connections in such a way that paths between at least some points of the network are available. A "fully meshed" network mapping is one that has all nodes connected to all other nodes, whereas a "partially meshed" network will have only some nodes with multiple connections with others. Meshing multiple paths improves network resiliency. However, more space is needed for dedicated links and involves high costs. Hybrid topology: When a network uses any combination of two or more topologies, it is called a hybrid network topology. Hybrid topology offers flexibility to different departments of an organization, within a network. A department can opt to implement a customized network topology mapping that is more suitable to its needs. Using business-level mapping, also called network dependency mapping, IT admins can obtain a graphical representation of devices according to business services that they serve. IT admins can create network-level dependency by adding links between devices. These links can then be configured based on user or organization requirements. You can also add maps to represent geographies where data centers are located. Business-level maps ensure that business-critical applications are always available and help with quicker troubleshooting. Layer 2 maps simplify complex and cluttered network interconnections. This is accomplished by first discovering a set of devices or subnet within a network. Layer 2 maps then presents the data-link layer of the network topography, enabling IT admins to monitor network connections, and each of its nodes and links. For instance, thanks to Layer 2 mapping, IT admins can instantly ascertain the device interconnections from a seed router. In doing so, an IT admin can monitor the seed router and all connected devices closely and in real time while not getting overwhelmed by alerts or alarms from all the devices. Layer 2 maps can be discovered using various protocols, including: Address Resolution Protocol (ARP): ARP can be used if you want to map an IP address to a physical or media access control (MAC) address that is recognized on the local device. Link Layer Discovery Protocol (LLDP): LLDP is used to discover wired LAN ethernets. Within a wired LAN ethernet, LLDP can be used to transmit the identity, capabilities, and neighbors. Cisco Discovery Protocol (CDP): CDP is used to transmit information about directly connected Cisco devices. If Cisco devices are used in your network, CDP is recommended to use for discovery. Network topology mapper enables an IT admin to segment or group a specific set of devices or interfaces within a network. The grouping can be by department name, floor number, location, criticality, interface type, or any other defined group. Grouping allows an IT admin to apply bulk configuration changes to multiple devices in an instant easily. An IT admin can also pinpoint faulty devices or interfaces in each branch, group, or subgroup. Device grouping promptly highlights trends and patterns to an IT admin, which in turn complements the forecasting and capacity planning of the IT infrastructure. ManageEngine OpManager provides comprehensive real-time network topology mapping capabilities. Network topology services combined with OpManager's wide array of features have made it a dependable solution for organizations and IT administrators around the world. OpManager's prominent real time network topology mapping features include: The purpose of network topology is to give you an overview of the network, by visualizing the physical connection of devices and the data flow within the network. It draws a distinction between logical and physical connection. Network topology helps an IT admin understand different elements of the network and where each of them connects. It also shows how each component interacts and aids in performance management. Network topology works by providing an advanced representation and overview of your network using topological diagrams, explaining how nodes are connected and how data transfer is happening in the network. The right network topology varies according to the size, scale and needs of the network of an organization. The right decision can be made after evaluating your network requirements, budget, scalability of topology, reliability and ease of implementation.
<urn:uuid:83919367-7420-499b-af4c-f79445a9f5f4>
CC-MAIN-2022-40
https://www.manageengine.com/za/network-monitoring/tech-topics/network-topology.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337529.69/warc/CC-MAIN-20221004215917-20221005005917-00681.warc.gz
en
0.916688
1,232
3.921875
4
During the first 40 years of the Internet, security has been about patching mistakes; the future will be about creating an environment that is secure by design and protected by big data. During the first 40 years of the Internet, security has been about patching mistakes that have been made in its architecture and software and defending these vulnerabilities against attacks from the outside. The future will be about creating an environment that is secure by design and has the ability to anticipate rather than merely respond to threats, government and industry experts say. This will not be a single secure infrastructure, but an ecosystem of technologies designed to better handle the basic chores of systems development; identity management and access control; and monitoring, analysis and response. The Internet was not designed to be secure. When originally launched by the Defense Department’s Advanced Research Projects Agency, it was not intended for public use. As it grew, development has always been more about what could be done with it rather than what should or should not be done. Today the Internet has become integral to the economy and security, while cybersecurity has become a front-page issue and observers see the need for fundamental shifts in the way we secure it. “It is clear that old, reactive, perimeter-based models of security are inadequate.” -- Art Coviello, RSA. “This does not mean perfect security, but instead a model that evolves and learns from change, whether process, technology or threat related,” said Art Coviello, executive chairman of RSA. “We have a tendency to talk about the threat, and there’s not a lot we can do about that,” said Ron Ross of the National Institute of Standards Technology, who heads the implementation program for the Federal Information Security Management Act. Good cybersecurity hygiene is important, Ross said. That means covering the basics of knowing your systems, understanding and managing their configurations, and ensuring that the proper defenses are in place. “But we need to go beyond that and make a difference in architecture and engineering,” he said. NIST is contributing to this shift with its catalog of FISMA security controls, Special Publication 800-53 Rev. 4, which contains guidelines for agencies to specify trustworthy design and operation of systems being procured. NIST also is working with an interagency working group to develop guidelines for engineering and supply chain security, which are intended to create a foundation for a more secure infrastructure in the future. And the agency is taking the lead in the creation of an Identity Ecosystem under the National Strategy for Trusted Identities in Cyberspace, which Jeremy Grant, head of the NSTIC National Program Office, calls a marketplace that will offer a variety of interoperable credentialing solutions. NSTIC is intended to address the failings of current identity management and access control schemes. Although technologies to securely authenticate remote users exist, scaling them across large user bases and multiple applications is cumbersome. As a result, many users and applications default to less-than-secure schemes such as simple user names and passwords, which are vulnerable to a wide variety of attacks. The marketplace already is responding to the problem, and companies including Google, Microsoft, Amazon and Apple have started offering customers multifactor authentication, Grant said. “NSTIC will provide a framework to facilitate interoperability between these solutions and others, and ensure that they are privacy-enhancing, secure, cost-effective and easy to use.” “NIST is leading a public-private partnership to develop the Identity Ecosystem by funding pilot projects that are enabling consumers and service providers to obtain and make use of trusted credentialing solutions,” Grant said. “It is supporting creation of the Federal Cloud Credential Exchange to help federal agencies more easily accept trusted, FICAM-approved credentials for access to government applications.” The NSTIC program office is working with the Identity Ecosystem Steering Group, a private sector-led organization formed to craft the legal, policy and standards framework to support the Identity Ecosystem. Trustworthy systems and trusted identities will not eliminate threats, and administrators still will have to defend their systems from attacks. “We are at the critical crossroads in the next phase of the evolution of the Information Age,” Coviello said. “As we face an evolving and escalating threat landscape, it is clear that old, reactive, perimeter-based models of security are inadequate.” Coviello is a proponent of using big data for this, harnessing data analytics for what he calls intelligence-driven security. “An intelligence-driven security model consists of a thorough understanding of risk, the use of agile controls based on pattern recognition and predictive analytics to replace outdated controls, together with the ability to analyze vast streams of data to produce actionable information,” he said. “In an age of open, hyperconnected enterprises this is the only model that will allow us to handle known and even unknown threats, and to help reduce risk to acceptable levels. A model that allows us to detect attacks quickly and respond quickly, a model based on big data.” None of these schemes to improve cybersecurity require creation of new technologies from scratch. But widely implementing and integrating them will be the work of the next five to 10 years.
<urn:uuid:e4ab19aa-82d6-442c-973b-87784f0ddfa6>
CC-MAIN-2022-40
https://gcn.com/cybersecurity/2013/05/cybersecuritys-best-bet-an-internet-that-learns-to-defend-itself/281612/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335365.63/warc/CC-MAIN-20220929194230-20220929224230-00081.warc.gz
en
0.945675
1,091
2.671875
3
Enterprise Resource Planning (ERP) is a system that integrates all the departments and functions of a company into one cohesive unit. It allows for the flow of data between all business areas, which helps to improve efficiency and decision-making. It is implemented in many ways, including on-premise, cloud-based, or hybrid systems. This blog post will explore the definition of ERP, how to implement it, the different types of systems available, and more! Enterprise Resource Planning (ERP) software is a comprehensive business management solution that enables companies to manage their core business processes in one system. ERP integrates all aspects of a company’s operations into a cohesive system, from accounting and financials to sales and marketing. This can be a huge advantage for businesses that need to manage complex operations or multiple locations. The best way to explain it in layman’s terms is to give you an example. Imagine you are the owner of a small business. You have a shop and sell clothes. You also have a website where people can buy your clothes online. To run your business, you need to manage two different types of resources: physical resources (e.g., the stock of clothes in your shop) and digital resources (e.g., the website where people can buy your clothes). Enterprise resource planning systems allow you to manage physical and digital resources using a single system. This is done by integrating different applications into a single system. For example, it might include an application for managing your shop’s stock, an application for managing your website’s content, an application for managing your human resources (e.g., employees), etc. A History of ERP Gartner initially used the term ERP in 1990. However, the business management software and applications used in the manufacturing sector have evolved in more recent decades as industry demand varies. ERP had its roots in the early 1960s when large American corporations began using specialized software to manage their complex businesses. The first generation of enterprise resource planning software was designed for manufacturing companies and focused on streamlining manufacturing resource planning and improving inventory management. The pace of the digitization of businesses will accelerate. As a result, companies using digital technologies in all aspects of their business will fundamentally change their operation. As a result, companies also demand robust ERP systems. Global ERP software will surpass the US$77.40bn mark and grow 10.2% between 2019 and 2024. Benefits of ERP Systems Today ERP solutions provide rich features for a business. However, what each firm considers to be the best value of these systems can be varied according to the company’s needs. Following are many benefits to implementing it in a business, Streamlined Business Processes It integrates all core business processes into a single system, which eliminates data silos and allows companies to make connections across different departments. This streamlined workflow results in faster and more efficient operations. Improved Decision Making The Enterprise resource planning system provides real-time data analytics so that business owners and managers can make informed decisions based on accurate information. Increased Efficiency and Productivity It automates routine tasks, saving employees time and focusing on more challenging projects with more significant potential for growth. ERP systems are often more affordable than implementing and maintaining individual software applications for each department. In addition, it can help businesses reduce waste and optimize their inventory levels, which leads to increased profits. Disadvantages of ERP systems Despite the great benefits of implementing ERP systems, many business problems exist in the marketplace. Unfortunately, companies could have avoided many by selecting a suitable supplier partner. Enterprise resource planning systems can often be complex and challenging to implement, especially for large companies that rely on them across multiple departments. Therefore, it is essential to consider its benefits and potential challenges before deciding whether or not it is suitable for your business. Here are some common challenges, The cost of ERP software varies depending on the solution you choose (cloud or on-premise) and how many modules you need, and if any customization is required, it can be expensive. Therefore, you should also factor in the costs of implementation, which may include training employees or data migration services, etc., into your budget when deciding whether or not it is suitable for your business. The solution implementation process can often be lengthy, especially if you integrate it with existing systems or make significant changes to your business. Therefore, it is essential to set realistic timeframes and goals for implementation and ensure everyone involved is aware of their expectations so that the process goes as smoothly as possible. ERP systems are complex and challenging, especially for employees unfamiliar with them. Therefore, you should provide adequate training and support to help employees make the most of its features and functionality and ensure they understand how it can benefit their role within the organization. Otherwise, employees may struggle to use it, resulting in employees doing their work manually or using a separate system, etc., defeating the purpose. It can be customized to meet your unique requirements, but that generally comes at an additional cost. Therefore, you should think carefully about what software would need to do to benefit your business and whether or not the company can customize the software accordingly before deciding whether it is right for you. It can be scalable, but that doesn’t necessarily mean it will be a good fit for your business if it expects to grow soon. Therefore, you should consider its scalability before deciding and ensure it can accommodate your future growth plans. Features of ERP systems Some essential characteristics distinguish an ERP solution from any other software type. This includes, - The advantage resulting if an ERP system combines multiple data sets from different databases and is centralized with other applications. A single source of instantaneous data removes manual merging between separate data sources based on their business operations. In addition, a shared database provides a consistent, cross-functional perspective on companies. - Across department roles, all people use the same User Interface (UI) and have similar user experiences when working with it. Integrated ERP includes several different components based upon specific features adapted to the various aspects of the organization, including front-office tasks. This quick overview shows the most common modules. - Accounting and Finance. - Customer Relationship Management (CRM). - Supply chain inventory management. - Human resources. - Business intelligence. - Purchase management - Sales management What is new with enterprise systems? The following are a few high-impact innovations. - Artificial intelligence and big data analytics - Integration with the Internet of Things (IoT) - Applications over smartphones - Blockchain integration - Cloud enterprise resource planning - SAAS model Find details of high-impact innovations below. Artificial intelligence and big data analytics Artificial intelligence (AI) with machine learning can help make better business choices. It will allow businesses to optimize their all operations, including, - Business operational processes. - Software systems. - Management structures. - Hardware and technology infrastructure. Enterprise software systems gather many enterprise data from day-to-day business processes and generate big data. Extensive data analysis can predict demand and help make future business decisions better and more efficient. Integration with the Internet of Things (IoT) The Internet of Things (IoT) is a connected physical object accessible. Automated, machine to machine, meaningful communication is established with it. Following are the advantages of integrating IoT with enterprise software systems, - It has enhanced data availability and accuracy. - Accurate and efficient communication. - Greater business intelligence. Applications over smartphones Mobile applications help in accessing information on the go. They also help in collecting job site information accurately. The most popular mobile operating systems on which client applications are built are, Enterprise systems are adopting blockchain technology. It helps businesses in achieving, - Enhanced transparency. - Greater security. - Increased traceability. - Improved efficiency. It helps in achieving greater control over supply chain management. Vendors host their software on the cloud computing system instead of customers’ data centers. It helps in faster upgrading and reduces maintenance efforts. Software-as-a-service models allow small and medium-scale businesses to use software systems without substantial initial investments. This model does not demand higher installation costs or IT people. Instead, it is pay-as-you-go based on how much you use. How to select an ERP system? ERP software installation can be a daunting process with numerous software options. The best software must meet the requirements and the business needs while providing the necessary support for successful implementation. The following are key factors to consider when selecting an ERP system: - The size of the company and its industry - Company’s business processes and needs - Implementation timeline and resources available - Cost of software and implementation services - Maintenance and upgrade requirements Once you have identified your needs, you can use the information to decide by comparing systems with similar features and functions. The following are some of the critical factors that will help you select ERP software for your business: - Product Functionality - Customization and Integration Capabilities - Vendor Support Services (implementation, training, maintenance) - Cost of the System and Implementation Services. It can be expensive and requires careful selection to ensure that it will meet your business needs now and in the future. When selecting software, consider all costs, such as software licenses, implementation services, hardware and software, upgrades, training, and support. Enterprise Resource Planning Software Deployment Options Enterprise Resource Plan systems vary in scope depending on the size and functions of a company. Three main types of application deployment models are, - Deploying on company premises is the traditional on-premises installation, where software and hardware are installed at a customer’s site. The company controls these resources and manages updates, security, and other maintenance. - Hosting – this model involves the provider hosting the application in their data center(s) and delivering it to users via a web browser. Customers have no hardware to procure or software to install but depend on their internet connection for access. - Deploying in the cloud is a SaaS model where applications are installed at the vendor’s data center and delivered via a web browser or other thin client. Customers have minimal control over infrastructure, upgrades, and security issues; they use the software. - Hybrid (some on the cloud or others offline) – companies are increasingly taking a hybrid approach to deploying software. Some system modules may be deployed in the cloud while others remain on-premises. Depending on the solution, the system may support different parts of a business, meet business needs, or have other deployment methods. The current ERP system provides many business functions. It needs connectivity and integration with other applications and data sources, including CRM/CRO and HCM software, e-commerce platforms, industry-specific products, and ERP software. The modern ERP System provides openness and flexibility and can easily integrate with various product suites using connectors or custom adaptors like the application programming interface or API. Ten things to look for in an ERP system A sound ERP software system may have several functions mainly determined by the industry and module they offer. Ten basic and essential characteristics are the following: 1. Comprehensive Functionality – The best packages have a comprehensive range of essential functions. These include inventory management, warehousing, distribution, and accounting (purchasing, sales, and financial) modules. Make sure the solution you choose has all of these functions. 2. Scalability – A sound system must grow with your business. Therefore, it should have the ability to accommodate increasing numbers of users and transactions as your business grows. 3. Customizability – The system must be customizable to meet the specific needs of your business. The best systems offer a high degree of flexibility to configure them to work the way you want them. 4. User-Friendliness – The solution must be easy to learn, use and maintain. A single database should contain all the information you need to run your business. In addition, the system should offer a single user interface for viewing inventory levels or processing sales transactions. 5. Robust Reporting and Analytics – Reporting and analytics capabilities should be a standard feature of any product. The system must provide the ability to track and analyze all of the vital information to your business, including financial reporting. 6. Seamless Integration with Other Applications and Systems – They must integrate seamlessly with other applications and systems that are important to your business. This includes both back-office and front-office systems. 7. Support for Multiple Operating Systems and Databases – They must run on multiple operating systems and support multiple databases. This allows you to choose the best operating system and database for your business. 8. Cost-Effectiveness – They can be costly, especially if you purchase a customized package. You should choose a system that offers your needed features at a cost within your budget. 9. World-Class Customer Service and Technical Support – They are complex and challenging to use. Therefore, you need to rely on the vendor for world-class customer service and technical support. 10. Timely Updates and Upgrades – They are complex and ever-changing. The vendor must provide timely updates and upgrades to ensure that your ERP system continues to meet your needs. ERP buyer types ERP vendor selection and deployment type depend on the needs of a buyer. Here are the common types of buyers. - Small businesses – Small businesses are an important market for enterprise software systems. They must manage their finances, operations, and human resources effectively to grow. - Medium-sized businesses – Medium-sized businesses need software to manage their more extensive and complex operations. In addition, they need a system that can scale and integrate with other business systems. - Large enterprises – Large enterprises have the most complex operations and need software to manage their resources effectively. They also require a system that can be customized to meet their specific needs. - Government organizations – Government organizations need enterprise software to support their specific needs. In addition, they require a system that is secure, scalable, and compliant with regulations. - Non-profit organizations – Non-profit organizations need software to manage their donations, volunteers, and finances. They also require a system that is easy to use and can be accessed by all employees. - Educational institutions need software to manage students, faculty, and finances. - Healthcare institutions need software to manage patients, staff, and finances. - Retail businesses need software to manage inventory, sales, and employees. - Manufacturing companies need software to manage production, inventory, and finances. - Technology companies need software to manage employees, projects, and finances. ERP Best Practices Enterprise resource planning systems usually follow established best practices. The software vendor has designed processes based on the experiences that have worked for hundreds of thousands of clients. The approach is often flexible for the software provider to adapt strategies. Following industry standards has significant economic benefits as a result. Companies find it easier to modernize and improve processes and thus increase productivity. Therefore, firms must follow best practices to meet financial standards. Following are the 11 best practices: - Identify your requirements - Design the project management team with a clear vision and strong leadership - Define key business processes that need to be delivered. - Involve customers in the decision-making process. It is crucial to get feedback from end-users to know what is happening and be more willing to use the software in the future. - Understand the change management process because ERP implementation is a significant change for your company, and you will be forced to change your organization and use new technology. - Ensure all costs are part of the total cost of ownership (TCO) – hardware, software, implementation, training, and support. - Select the right partner with a strong track record in your industry - Leverage the Partner’s Implementation Methodology (PIM) to shorten your project timeline - Use pre-configured modules whenever possible to speed up deployment and get up and running quickly - Train end users thoroughly and ensure that support is available when needed - Perform a post-implementation review to identify areas for improvement. When you implement it, you will go through a specific life cycle. This life cycle will help ensure that your implementation is successful. Here is a look at the different phases of the ERP Implementation Life Cycle. Cost of ERP The cost of the system can vary widely depending on the size and complexity of the business. The cost of implementation and customization can also be high, so it’s essential to factor these expenses into the overall cost of the system. Enterprise Resource Planning solutions track and integrate all aspects of a company’s business, from accounting and inventory management to human resources and customer relations. As a result, these systems are critical tools for businesses of all sizes, and companies can use them to streamline operations, improve decision-making, and boost profits. If you’re looking for a comprehensive overview of Enterprise Resource Planning systems or want to learn about their benefits to businesses, keep reading. We have discussed everything you need to know about it in this post. [ ERP filetype PPT presentation file ]
<urn:uuid:dcefdcb3-b0cb-48e1-8fd9-10cefecd684d>
CC-MAIN-2022-40
https://www.erp-information.com/what-is-erp.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337322.29/warc/CC-MAIN-20221002115028-20221002145028-00081.warc.gz
en
0.921474
3,669
2.984375
3
Representative means a natural or legal person established in the Union who, designated by the controller or processor in writing pursuant to Article 27, represents the controller or processor concerning their respective obligations under this Regulation. The use of representative applies to companies that lack an establishment in the European Union but still process personal data of residents in the European Union. These companies need a representative to comply with GDPR. An example could be an American company without an establishment in the European Union that sells products online to citizens within the European Union. This company need to have a representative within the European Union to comply with the GDPR.
<urn:uuid:8b6ee2d5-ab74-4d0f-9b62-da777320d902>
CC-MAIN-2022-40
https://www.gdprsummary.com/gdpr-definitions/representative/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337322.29/warc/CC-MAIN-20221002115028-20221002145028-00081.warc.gz
en
0.944243
123
3
3
Cybersecurity & The Human Element. The human element plays a significant role in terms of cybersecurity. It is usually divided into two cybersecurity groups, the “Bad Actor” and the “Good Actor”. The “bad actors” are cyber criminals and hackers that deliberately compromise organizations with malicious software to gain a profit. “Good actors” are the executives who promote security plans and programs to defend organizations against attacks. According to a 2022 Verizon data breach investigation report, 85% of cybersecurity breaches involved the human element. Unfortunately, this confirms that good actors are still the weakest part of most organizations’ cybersecurity defense strategy. Innocence doesn’t exempt you from responsibility. Unfortunately, good actors are still making mistakes that enable bad actors to succeed. The Good Actors Cyber Security Mistakes Good actors usually fall victim to social engineering attacks like phishing. Phishing consists of receiving legitimate-looking emails from known enterprises. This email asks the recipient to perform a specific action like accessing credentials or redeeming a coupon code. By clicking the link, good actors are unsuspectingly downloading malware. Another mistake made by good actors is modifying permissions on a network to allow easy access for other people. They believe that they are trying to help a fellow co-worker, but changes in security settings can accelerate a bad actor’s work. Data mishandling is a common mistake among good actors. Saving information on the hard drive instead of the network cloud may seem like a small mistake, but in reality, the desktop does not have the same security measures as that the cloud server does. By doing this, good actors are making bad actors’ jobs easier. What happens when a corporate email is unavailable? Good actors usually share business information through their personal email accounts. Personal email accounts are not as protected as professional accounts. Thus, sharing personal information through email leaves an open vulnerability for hackers to take advantage of. IT Teams Can’t Escape Human Element Mistakes IT Teams professionals that are tasked with deploying new hardware, systems, and applications, can easily forget to configure important cybersecurity settings. The technology can still work, but it is not properly protected. Neglecting system maintenance is another mistake made by IT professionals. Not patching or updating software applications is a colossal cybersecurity error that leaves the whole IT infrastructure vulnerable to attack. Don’t forget that software is created by humans that fit into the category of “Good Actors”. Most of the time the development of new software tends to be rushed, resulting in software with security flaws. The Reality of the Human Element Good actors that make mistakes are contributing to the failure of their cybersecurity measures. But how can we reduce the number of mistakes that us humans make? How can we confidently respond to cybersecurity incidents that arise due to our own error? The Key to Reducing Human Element Mistakes Create effective cyber security policies: Good policies will provide a set of rules and guidelines that people are required to follow to maintain high-security standards. For example, in an office environment, there is a dress code that people must abide by. Employees know what they can and cannot wear for proper dress attire. Cybersecurity policies will provide the same direction in terms of safe security practices. Mandate cybersecurity awareness training: Educate people on how to deal with and respond to common social engineering attacks. Just as people know how to detect any dangerous situation, employees need to be taught how to detect the signs of malicious software in their inboxes. Develop good emergency response plans: Good cybersecurity policies and awareness training will reduce the possibility of human error but not eliminate it entirely. A good emergency response plan will provide the specific procedures that need to be followed to successfully recover from a cyberattack. Wrapping it Up To provide structure for people, we must prepare them to deal with failure. This is the correct way to address human errors associated with cybersecurity. BTI has more than 35 years of experience in IT, security, and communications. If you would like to implement high-security measures or need assistance with your security-related issues, contact us now to schedule a free consultation!
<urn:uuid:12ea4161-b6cd-4716-b18b-3c563312d1a8>
CC-MAIN-2022-40
https://www.btigroup.com/post/human-error-and-cyber-security
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337432.78/warc/CC-MAIN-20221003200326-20221003230326-00081.warc.gz
en
0.943484
855
2.84375
3
As technology takes over every aspect of day to day life, criminal activities evolve and adapt. Social media, mobile apps, and the dark web have created new platforms and channels for criminals, terrorists, and hackers to operate their illegal activities and communicate online with minimal visibility. As a result, threats today are more sophisticated, leveraging technology to evade detection while posing risks to corporations, governments, and civilians. Security methods have advanced to prevent data leads, hacking, phishing attacks, and more, but it is a constant challenge in a never-ending game of cat and mouse. While automation and increased connectivity have led to significant advancements in many industries, they are also used more often in large-scale cyber attacks.
<urn:uuid:becbdd36-a6cf-4869-b100-932f920b5f50>
CC-MAIN-2022-40
https://cobwebs.com/resource/white-paper/the-rise-and-future-of-cyber-attacks/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337537.25/warc/CC-MAIN-20221005042446-20221005072446-00081.warc.gz
en
0.960575
141
2.78125
3
According to the most recent annual Cost of Cyber Crime Study by the Ponemon Institute, the average cost of detecting and recovering from cyber crime for organizations in the United States is $5.4 million. Median costs have risen by almost 50 percent since the inaugural study in 2010. The finding masks the enormous variation of data breach costs which can range from several hundred thousand to several hundred million dollars, depending on the severity of the breach. A growing number of insurance companies are offering cyber protection to enable organizations to manage such costs. This includes traditional carriers in centers such as London, New York, Zurich and elsewhere, as well as new entrants targeting the cybersecurity insurance market. Carriers in the latter category should be carefully veted since some new entrants have been known to offer fraudulent policies in order to exploit the growth in demand for cyber insurance. Cybersecurity insurance has been commercially available since the late 1970s but was limited to banking and other financial services until 1999-2001. It became more widespread after Y2K and 9/11. Premiums also increased after these events and carriers began to exclude cyber risks from general policies. More recently, the dramatic rise in the threat and incidence of data breaches has propelled cybersecurity into a boardroom issue and led to a growing interest in cyber policies from organizations looking to limit their exposure. A 2011 study performed by PriceWaterhouseCoopers revealed that approximately 46% of companies possess insurance policies to protect against the theft or misuse of electronic data, consumer records, etc. However, this is contradicted by the findings of 2012 survey by Chubb Group of Insurance Companies which revealed that 65 percent of public companies forego cyber insurance. The confusion may be due to a general lack of awareness among survey responders of the exact nature of insurance coverage. Many responders appear to be under the impression that cyber risks are covered by general insurance policies even though this is no longer the norm. The cybersecurity insurance industry is highly diverse with carriers employing a plurality of approaches. Some offer standardized insurance products with typically low coverage limits. Others provide customized policies tailored for the specific needs of each client. Furthermore, the industry is evolving rapidly to keep pace with evolving threats and trends in cybersecurity. Policy premiums are driven primarily by industry factors. E-commerce companies performing online transactions while storing sensitive information such as credit card data are generally considered high risk and are therefore subject to higher premiums. Health institutions hosting data such as social security numbers and medical records are also deemed high risk. Premiums typically range between $10,000 to $40,000 per $1 million and provide up to $50 million in coverage. However, most standard policies only provide coverage for specific third-party costs to cover losses incurred by a company’s customers or partners. This includes risks related to unauthorized access and the disclosure of private information, as well as so-called conduit injuries that cause harm to third party systems. Polices that provide coverage for first-party areas such as crisis management, business interruption, intellectual property theft, extortion and e-vandalism carry far higher premiums and are therefore relatively rare. This limits the appeal of cybersecurity insurance and ensures organizations need to self-insure for such risks for the foreseeable future. The situation is unlikely to improve until actuarial data is more widely available and shared between carriers for cybersecurity risks. This may require the establishment of a federal reinsurance agency and legislative standards for cybersecurity. Carriers are unlikely to offer full cover for all first and third party costs arising from security breaches. This is due to the moral hazard associated with such coverage. Organizations that completely transfer cyber risk have no incentive to invest in preventative and monitoring controls to manage security risks. However, most carriers have exclusions for breaches caused by negligence. Other exclusions include coverage for fines and penalties, often due to regulatory reasons. Aside from industry considerations, other factors that drive premiums for cybersecurity insurance are risk management cultures and practices in organizations. Carriers often assess cybersecurity policies and procedures before deciding premiums. Organizations that adopt best practices or industry standards for system security are generally offered lower premiums than those that do not. Therefore, insurers work closely with clients during the underwriting process to measure the likelihood and impact of relevant cyber risks. This includes consideration for management controls. Carriers that decide not to assess the cybersecurity practices of prospective clients tend to compensate by including requirements for minimal acceptable standards within policies. These clauses ensure that carriers do not reimburse organizations that failed to follow generally-accepted standards for cybersecurity before a security breach. Cybersecurity standards for SAP systems are embodied in benchmarks that are aligned to security recommendations issued by SAP. This includes the SAP Cybersecurity Framework outlined in the white paper, Protecting SAP Systems from Cyber Attack. Cybersecurity insurance is most valuable for organizations with mature cyber risk cultures including effective standards and procedures for preventing, detecting and responding to cyber attacks. It enables such organizations to transfer the risk of specific costs arising from security breaches that are more cost-effectively covered by third-party coverage rather than self-insurance. Cybersecurity insurance is not a viable option for companies with weak risk management practices. Even if carriers were willing to insure such high-risk organizations, the premiums are likely to outweigh the cost of self-insurance. Furthermore, the likelihood that organizations would be able to collect upon such policies is low.
<urn:uuid:09c39320-1f13-40c5-be53-ee96871045a6>
CC-MAIN-2022-40
https://layersevensecurity.com/cybersecurity-insurance-is-it-worth-the-cost/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337537.25/warc/CC-MAIN-20221005042446-20221005072446-00081.warc.gz
en
0.961549
1,074
2.5625
3
Optical Wavelength Bands Evolution In the optical fiber communications system, several transmission bands have been defined and standardized, from the original O-band to the U/XL-band. The E- and U/XL-bands have typically been avoided because they have high transmission loss regions. The E-band represents the water peak region, while the U/XL-band resides at the very end of the transmission window for silica glass. Intercity and metro ring fiber already carry signals on multiple wavelengths to increase bandwidth. Fibers entering the home will soon do the same. Now several types of optical telecom systems have been developed, some based on time division multiplexing (TDM) and others on wavelength division multiplexing (WDM). This article may represent the evolution of optical wavelength bands mainly by describing these two high-performance systems. Wavelength Division Multiplexing (WDM) Normally, the WDM system could be divided into coarse wavelength division multiplexing (CWDM) and dense wavelength division multiplexing (DWDM). Dense Wavelength Division Multiplexing DWDM systems were developed to deal with the rising bandwidth needs of backbone optical networks. The narrow spacing (usually 0.2 nm) between wavelength bands increases the number of wavelengths and enables data rates of several Terabits per second (Tbps) in a single fiber. These systems were first developed for laser-light wavelengths in the optical C-band, and later in the L-band, leveraging the wavelengths with the lowest attenuation rates in glass fiber as well as the possibility of optical amplification. Erbium-doped fiber amplifiers (EDFAs, which work at these wavelengths) are a key enabling technology for these systems. Because WDM systems use many wavelengths at the same time, which may lead to much attenuation. Therefore optical amplification technology is introduced. Raman amplification and erbium-doped fiber amplifiers are two common types used in WDM systems. In order to meet the demand for “unlimited bandwidth,” it was believed that DWDM would have to be extended to more bands. In the future, however, the L-band will also prove to be useful. Because EDFAs are less efficient in the L-band, the use of Raman amplification technology will be re-addressed, with related pumping wavelengths close to 1485 nm. Coarse Wave Division Multiplexing CWDM is the low-cost version of WDM. Generally these systems are not amplified and therefore have limited range. They typically use less expensive light sources that are not temperature stabilized. Larger gaps between wavelengths are necessary, usually 20 nm. Of course, this reduces the number of wavelengths that can be used and thus also reduces the total available bandwidth. Current systems use the S-, C- and L-bands because these bands inhabit the natural region for low optical losses in glass fiber. Although extension into the O and E-band (1310 nm to 1450 nm) is possible, system reach (the distance the light can travel in fiber and still provide good signal without amplification) will suffer as a result of losses incurred by use of the 1310 nm region in modern fibers. Time Division Multiplexing TDM systems use either one wavelength band or two (with one wavelength band allocated to each direction). TDM solutions are currently in the spotlight with the deployment of fiber-to-the-home (FTTH) technologies. Both EPON and GPON are TDM systems. The standard bandwidth allocation for GPON requires between 1260 and 1360 nm upstream, 1440 to 1500 nm downstream, and 1550 to 1560 nm for cable-TV video. To meet the rise in bandwidth demand, these systems will require upgrading. Some predict that TDM and CWDM (or even DWDM) will have to coexist in the same installed network fibers. To achieve this, work is underway within the standardization bodies to define filters that block non-GPON wavelengths to currently installed customers. This will require the CWDM portion to use wavelength bands far away from those reserved for GPON. Consequently, they will have to use the L-band or the C- and L-bands and provided video is not used. In each case, sufficient performance has been demonstrated to ensure high performance for today’s and tomorrow’s systems. From this article, we know that the original O-band hasn’t satisfied the rapid development of high bandwidth anymore. And the evolution of optical bands just means more and more wavelength bands will be called for. In the future, with the growth of FTTH applications, there is no doubt that C- and L-bands will play more and more important roles in optical transmission systems.
<urn:uuid:7ebcef1f-ffea-4841-82f2-541ff8b8e768>
CC-MAIN-2022-40
https://community.fs.com/blog/from-o-to-l-the-evolution-of-optical-wavelength-bands.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030338280.51/warc/CC-MAIN-20221007210452-20221008000452-00081.warc.gz
en
0.940257
987
3.484375
3
I believe virtual reality (VR) and augmented reality (AR) are on the cusp of mass success, and will dramatically change the way we use technology. However, with new technology comes new attack surfaces. After watching security researchers and attackers pick apart the Internet of Things (IoT) so easily, it would be foolish not to consider the security of new technologies. So that begs the question – what are the risks of virtual and augmented reality? To answer simply, AR and VR pose little risk today, and likely will post little risk for the next five years. However, these technologies could become seriously dangerous in the next decade. Let’s explore why. Little new attack surface Today’s VR and AR technology offer little new attack surface since they’re built upon existing platforms. At the highest level, AR and VR are mostly just new display and input mechanisms added to pre-existing devices. The underlying computers that power the technology—whether it’s a PC, console, or mobile device—haven’t really changed much. AR/VR tracking software doesn’t really need to connect to the Internet and doesn’t pose any greater risk than any other software you might add to your computer; perhaps even less risk than your average network connected game. No valuable data Another way VR/AR could increase your risk profile is by collecting new data that makes you a more valuable target. VR/AR technology does track the motion of your head (and sometimes your hands), but this data is of little use to a criminal attacker. Other data that VR/AR systems track, like voice and video, have risks of their own, but these aren’t unique to VR/AR – people have shared voice and video data online for years. Current motion tracking data is relatively low fidelity and doesn’t provide much value to an attacker or criminal. For instance, VR motion tracking is pretty accurate, but the criminal value of knowing where your head or hands are positioned is negligible. Currently no way to monetize a VR attack Even with these drawbacks, attacks would still target VR/AR if they could find some way to monetize their attacks, like through social engineering for example. But fortunately for end-users, there’s little criminal money to be made with how most early adopters use AR/VR. Today, VR is more commonly used for gaming. But since gamers already expect to be in a fantasy world disconnected from reality, there’s little opportunity for attackers to alter VR for social engineering either. On the other hand, real AR currently exists in the form of experimental, non-consumer products like Google Glass and Hololens, and novelty uses like Pokemon Go. Until AR becomes such a ubiquitous tool that it’s used in everyday life, it won’t be trusted enough by the public, which means attackers won’t have the opportunity to trick us. In short, AR/VR doesn’t provide a rich enough target for attackers… yet. Safe present, potentially grim future So far, I’ve painted a relatively safe picture for AR/VR that will likely hold true for the next five years or so. However, as these technologies improve and become more commonplace, they will pose a bigger danger—especially AR. Here are some examples of what VR/AR hacks of the future might look like: 1. Finer tracking data allows for more dangerous hacks Imagine the future of online shopping. This could become an entirely VR experience, where users literally browse a virtual store front, interact with items, and perhaps even try them on an avatar. Of course, the program used knows a person’s credit or debit card, so when an item is purchased for shipment, it can be processed and sent. However, for added security, online shops could make a user avatar virtually enter some sort of pin or code to verify the debit or credit used. In the virtual world, a user might do this like they would in the real world, by using fingers to type the code on a virtual keypad (in the air from the user’s perspective). However, doing this means the system must record and transmit the fine finger tracking data showing fingers type a pin. If an attacker can capture that data, they have all they need to recreate a user pin (and they would presumably have some way to capture a card’s digital data too). The future of AR/VR headsets may also include eye tracking, for a variety of reasons (one being to render a virtual world from the proper perspective). This eye tracking data could provide additional value to malicious actors. Knowing exactly what a user is looking at could reveal valuable information to an attacker. For instance, if a user was in an online shop, the eye tracking data would show what the person is most interested in. Marketers are already using web-coding tricks to monitor mouse movements and clicks, to help figure out buying habits and interests. Attackers that capture that data could recreate user actions, in the same way the manual pin entry could be mimicked in the above example. 2. Warping augmented reality In the future, users may wear some sort of AR device that overlays a digital heads up display (HUD) over their real-life field of vision. Hands could be used to gesture and pin computer screens and browser windows to various walls in the real world. Virtual speed limit signs could pop up to warn drivers of excessive speeds. Contact information could remind users about acquaintances they meet (or even initial info for first time interactions). A user could even get instant calorie information when handling food items in a grocery store. As the public begins to use AR technology more regularly, this type of information will seem more like reality, causing users to trust it more. But if an attacker can hack these AR devices, it presents a huge opportunity for them to poison this information and get end users into trouble. For example, what if an attacker wanted to hurt someone. Imagine a person driving in an unfamiliar town. Unbeknownst to them, the person is approaching a tight turn on the freeway. The turn has a physical speed limit warning sign telling the driver to drop speed to “25mph,” due to the corner. If an attacker could gain control of the driver’s AR system, they could overlay those 25mph signs with “60mph,” putting the driver in the dangerous situation of hitting a corner far too fast. And that’s just the start of what they could do if humanity becomes “programmed” to trust their AR HUD. Frankly, once AR becomes advanced enough to overlay anything we see with real-time CGI, and we use AR as a second-nature, trusted “sense,” there is no limit to how hackers might alter our realities. 3. Your perfect digital VR clone The ideal future of VR depends on a few things. First is full body digital tracking, where literally every movement of every appendage is tracked very finely, and recreated in digital space. Second, is the perfect digital avatar. In the future, the cameras or devices tracking us, will create quick maps of our physical 3D space, apply our texture to that model, and make a perfect replica in a virtual world. This might sound like sci-fi, but it’s closer than you think. However, imagine the social engineering possibilities if malicious actors got a hold of a person’s 3D model, and a history of all their movements in VR. Animators and computer scientists have already created many methods to make a person sound like they said something they didn’t, based on previous recordings of their voice. They can even alter video of a person, to give them different expressions and lip movements. In fact, you can see a scary example of this on a site created by NPR’s Radio Lab. While these fake videos haven’t been perfected yet, imagine how VR tracking data and accurate 3D models could change things. One of the unique identifiers of an individual is their unique movements and verbal or physical “ticks.” If compromised, these personal intricacies could allow hackers to socially engineer a person’s friends, or digitally impersonate a user. You really shouldn’t worry about the security of AR/VR today. Most people are only using these technologies for entertainment, and they don’t introduce much new data or attack surface. The worst VR risk today is making you unaware of your real physical surroundings. That said, as these technologies mature, expect criminals to target them more. I look forward to us realizing AR and VR’s full potential, but we should head into this future with our defenses ready.
<urn:uuid:93420dff-cc28-4ef0-b48b-4a93503bca1c>
CC-MAIN-2022-40
https://www.helpnetsecurity.com/2017/09/06/hacking-virtual-augmented-reality/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030338280.51/warc/CC-MAIN-20221007210452-20221008000452-00081.warc.gz
en
0.939643
1,808
2.765625
3
A method for scrambling data to protect it from the super powerful computers of the future has received a speed boost from a team of researchers from Duke and Ohio State universities and the Oak Ridge National Laboratory. The method uses quantum key distribution to guard data from prying eyes, wrote Nurul T. Islam, Charles Ci Wen Lim, Clinton Cahall, Jungsang Kim and Daniel J. Gauthier in an article published in ScienceAdvances. The problem in the past with the technology is it’s slow, they explained. Transfer speeds typically are measured in kilobits per second. However, the researchers found a way to increase key transmission rates between five and 10 times, bringing them into the megabit per second range. Making quantum key distribution practical is viewed as a way to counter the future threat to encrypted data. “One powerful aspect of quantum encryption is that it is secure against quantum computers,” said Kevin Curran, a senior member of the IEEE and a professor of cybersecurity at Ulster University in Northern Ireland. “Its strength does not depend on mathematical complexity, like current cryptography, but on physical principles,” he told TechNewsWorld. Hedge Against Future Computers Quantum computers pose a serious threat to cybersecurity because most current cryptosystems potentially could be broken with a powerful quantum computer in a reasonable amount of time, the researchers wrote. Quantum key distribution with symmetric encryption is one of the very few methods that can provide provable security against an attack aided with a quantum computer, they maintained. A major limitation of most current QKD systems, though, is the rate at which the secret key is generated, the researchers wrote. It is orders of magnitude lower than existing digital communication rates — a limitation that ultimately prevents QKD from being useful for a wide range of communication tasks. With their system, though, the researchers were able to generate provably secure cryptographic keys at megabit-per-second rates over metropolitan distances — 10 to 50 miles. What’s more, the system can be built with mostly off-the-shelf components, and the protocol the researchers developed is robust against coherent attacks on the data it’s moving. The researchers were able to obtain their speed boost by finding a way to pack more information on the photons used to carry a key’s data. With adjustments to release time and the phase of the photon, they managed to get two bits of data on the particle instead of one. However, that trick could have adverse consequences if the researchers’ scheme tries to reach beyond metro ranges, noted Jon Geater, CTO of Thales. “QKD has always held much promise, but it suffers from practical issues. Actually, the speed improvement here doesn’t really address the biggest issues, which are stability and range,” he told TechNewsWorld. “While the theory and physics are somehow perfect, practical implementations in glass fiber can only survive around 150 miles before breaking down and becoming unusable,” Geater said. “It is unlikely that doubling the carriage density of each photon will help with this.” While there is concern that a new generation of quantum computers could break current encryption schemes, fighting quantum with quantum isn’t the only solution to the problem. Researchers have been working to solve future problems by creating “quantum-proof” encryption algorithms or other solutions, noted Stephen Ezell, vice president for global innovation policy at the Information Technology & Innovation Foundation. For example, public-key systems can be designed to generate random keys with each message sent. “This means that if a quantum computer compromises one message in a system, all others are not also compromised,” Ezell told TechNewsWorld. Such systems have advantages over new schemes like quantum key distribution, said Jonathan Katz, director of the Maryland Cybersecurity Center. “These replacement systems being developed that are thought to be resistant to quantum computers would fit in more easily to the current way key distribution is done than quantum key distribution would,” he told TechNewsWorld. Still, betting on a “post quantum” mathematical solution could be a risky wager. “What the United States would like to do is replace the hard math problem used in the current cryptographic schemes with a different math problem that we hope is not breakable by a quantum computer,” explained researcher Gauthier, who is a professor of physics at The Ohio State University. “The rest of the world is definitely looking at those,” he told TechNewsWorld, “but they’re not putting all their faith in that one approach.”
<urn:uuid:abeffff5-f93f-404c-8c50-d0193e0ea6fc>
CC-MAIN-2022-40
https://www.linuxinsider.com/story/quantum-key-distribution-gets-a-speed-boost-84983.html
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030338280.51/warc/CC-MAIN-20221007210452-20221008000452-00081.warc.gz
en
0.937075
983
3.296875
3
Have you been hearing about the General Data Protection Regulation? Do you collect, use, or process personal data of subjects in the European Union? What is GDPR? Who must comply? How can you prepare? Should you complete a GDPR assessment? With the repercussions of data breaches hitting the headlines more often every day, it’s important to understand how this privacy legislation is going to affect your business and to ask yourself: are you ready for GDPR compliance? What is GDPR? The Data Protection Directive (DPD) has been in place for 20 years, setting a minimum standard for data protection laws in European Union Member States. Many of these states have taken legislation to the next level when it comes to protecting personally identifiable information, making it increasingly difficult for EU citizens to know how their rights and information are being protected across Europe, and for organizations to determine which laws apply when working across multiple Member States. Born out of cybercrime threats, technology advances, and concerns about data misuse, this legislation will require all data controllers and data processors that handle personal data of EU residents to “implement appropriate technical and organizational measures…to ensure the ongoing confidentiality, integrity, availability, and resilience of processing systems and services.” GDPR is enforceable and is equivalent to a US Federal Law, and failure to comply with GDPR can lead to fines of up to €20 million or 4% of annual global turnover – whichever is greatest. Who Must Comply with GDPR? GDPR applies to any entity collecting, using, or processing personal data of any data subject in the European Union. It’s important to note that this doesn’t just apply to organizations working within the EU, but also any organization in the world providing services to data subjects within the EU, and according to a general consensus, approximately 66% of US companies will be subject to this data protection law. GDPR requires organizations based outside the EU to nominate a representative organization within the EU where the target data subjects are based, and must be mandated by the data controller or processor to be addressed by the supervisory authority or data subjects on all issues relating to the processing of personal data. What is Personal Data? According to GDPR, personal data is defined as any information relating to an identified or identifiable person, or data subject, who can be identified by a name, an ID number, location data, or physical, physiological, genetic, mental, economic, cultural, or social identity. What is a Data Controller? The data controller is the organization that determines the purpose for processing personal data and what processing will be done. Processing, according to GDPR, includes the collection and storage of information. It is possible for an organization to be accountable as the data controller, but not otherwise involved with any actual processing of personal data. What is a Data Processor? A data processor is any organization or entity that processes personal data on behalf of a data controller. Processing is essentially anything done to the data, including storing, archiving, or reviewing. It is possible, and sometimes common, for an organization to be both a data controller and a processor. GDPR became effective May 25, 2018, so if you’ve not begun preparing, you should start now. In a highly data-driven world, it’s our responsibility to help protect organizations from data and privacy breaches. For more information on how you can complete a GDPR assessment, contact us today.
<urn:uuid:149528c8-f1ec-45f3-b594-724ffb16285c>
CC-MAIN-2022-40
https://kirkpatrickprice.com/blog/are-you-ready-for-gdpr-compliance/
null
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030331677.90/warc/CC-MAIN-20220924151538-20220924181538-00282.warc.gz
en
0.919681
708
2.65625
3