text
stringlengths
234
589k
id
stringlengths
47
47
dump
stringclasses
62 values
url
stringlengths
16
734
date
stringlengths
20
20
file_path
stringlengths
109
155
language
stringclasses
1 value
language_score
float64
0.65
1
token_count
int64
57
124k
score
float64
2.52
4.91
int_score
int64
3
5
What is Cloud Security? Cloud security, also known as cloud computing security, encompasses a broad range of controls-based technologies, policies, and procedures that enterprises deploy to protect data, information, applications, communication channels, and infrastructure in the cloud. As with on-premise systems and data, businesses should vigilantly protect cloud-based information assets. - Cloud security refers to controls, processes, and technologies used to protect information and systems in the cloud. - Cloud security safeguards cloud data, supports regulatory compliance, and protects user privacy - Cloud computing security is essential since it protects applications and information from frequent and sophisticated cyber threats in the cloud - Cloud security involves a joint responsibility between a business and a cloud service provider - Cloud security offers benefits like centralized management, affordability, automation, and reliability Organizations can configure and manage cloud computing security in one place, offering reduced administration overheads and empowering IT teams to focus on other vital tasks. The Joint Responsibility in Cloud Security Cloud security delivery depends on the individual cloud service provider and security controls in place. In most cases, implementing cloud security is a joint responsibility between the cloud provider and the customer or business. Cloud security shared responsibility model has three categories: - Provider’s security responsibilities: the cloud service provider safeguards the cloud infrastructure itself, as well as other security components like patching and physical host and network configurations - Customer’s security responsibilities: cloud consumers manage users, control access privileges, and protect cloud accounts from unauthorized access. Cloud customers also meet compliance with various cloud security regulations - Service model-based security responsibilities: these security responsibilities vary depending on the service model, such as infrastructure as a service (IaaS), platform as service (PaaS), or software as a service (SaaS) Importance of Cloud Security As businesses migrate their applications and data to the cloud to leverage capabilities like rapid elasticity, ubiquitous access, resource poling, and on-demand services, they require robust cloud computing security controls. Cloud-based workloads face constant risks from sophisticated and frequent cybersecurity threats. For this reason, enterprises should work with service providers offering in-depth security controls tailored to the company’s needs. Cloud security protects cloud storage and networks against malicious cyberattacks. The security measures also deter human error or negligence that causes data breaches. Cloud computing security provides procedures that enable data recovery during and after an incident. Overall, the controls reduce the impact of any cloud system or information compromise. Core Components of Cloud Security Cloud security focuses on these primary areas:· - Data security - Identity and access management (IAM) - Legal compliance - Business continuity (BC) planning - Governance (cloud security policies) - Data retention Cloud security focuses on designing and integrating the above categories to protect various resources, such as physical networks, data storage, data servers, computer virtualization frameworks, middleware, operating systems, data, applications, and endpoints. Benefits of Cloud Security - Centralized security – just as cloud computing centralizes workloads, cloud security consolidates protection. Cloud security offers centralized and advanced capabilities such as web filtering, traffic analysis, and network monitoring for numerous endpoints and applications. - Affordable – cloud computing security eliminates the need for investing in dedicated hardware, like in the case of an on-premise security strategy. Apart from removing capital expenditure, cloud security shared responsibility reduces administrative overheads. - Automation – reputable cloud service providers or cloud security platforms automate security configurations and security update administration. Cloud security services eliminate manual tasks that can massively drain resources and the workforce. - Reliability – cloud computing security allows users to access data and application safely from any location. Popular Cloud Security Challenges Some of the challenges facing cloud security include: - Extensive attack surface: cloud-based environments are large and attractive attack surface for hackers who exploit different vulnerabilities in cloud computing - Low visibility and tracking: cloud service providers have control over the cloud infrastructure. Cloud users lack visibility and control to identify cloud assets and protect them from cyberattacks effectively - Dynamic workloads: providers provision and decommission cloud computing resources dynamically. Traditional security controls cannot protect and enforce policies on the dynamic cloud environments - Emerging cloud concepts: embracing DevOps (continuous integration and continuous delivery) culture requires new security controls that embed in code and templates during the development cycle - Complex environments: managing security consistently in hybrid environments (on-premise and cloud) requires tools and procedures that work across different deployment environments - Increased compliance requirements: cloud users must comply with standards like PCI DSS, NIST, HIPAA, and GDPR. Poor visibility and cloud platforms’ dynamic nature complicates cloud compliance
<urn:uuid:8386c8a1-3bcb-46ef-9e9e-1e2166678716>
CC-MAIN-2024-38
https://cyberexperts.com/encyclopedia/cloud-security/
2024-09-07T20:06:25Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00520.warc.gz
en
0.906044
969
3
3
In an age of intelligence and interconnectedness where data reigns supreme, protecting it becomes paramount. Living in a world where information flows freely, and digital footprints mark our every move, cyber security has gone beyond being optional – it is now a necessity. With this article, we journey through the labyrinthine landscape of cybersecurity computer technology. From exploring the basics that form the bedrock of this complex field to gleaming insight into the emergent cyber threats that keep expert security professionals on their toes, from gaining an understanding of the revolutionary advances in cyber security tech to studying notorious breaches that jolt the industry. All of this while also trying to crystal-gaze into what may lay ahead in the future of cyber security – Here’s a holistic endeavor to unravel the intricate, often enigmatic world of cyber security. Understanding the Basics of Cyber Security Many view cyber security as the proverbial “guardian at the gates,” the digital fortress that protects our most valuable and sensitive information. Aligned with this perspective, for those tech enthusiasts keen on adopting the latest technological trends, it’s imperative to unravel the fundamental principles embedded within the concept of cyber security. Firstly, Confidentiality is paramount. Like a vault that safeguards a multimillion-dollar asset, cyber security mechanisms must ensure the secrecy of information. This keeps unauthorized individuals from accessing sensitive information, ensuring only those with the correct credentials can access the data they’ve been granted to. Then comes Integrity, ensuring that the data remains unaltered in transit and is exactly the same when it reaches its destination. Cyber security measures must pack quite a punch in being able to detect and ward off threats that aim to modify or destroy original data. Availability, on the other hand, ensures that authorized users have uninterrupted and timely access to resources and information. Everyone knows the frustration of being unable to retrieve necessary data due to system downtime. To eradicate this inconvenience, adequacy and robustness in computing resources are essential. Next up – Authenticity; here is where cyber security ensures that transactions, as well as communication between various entities, are genuine. It’s about making sure that those you’re interacting with online are indeed who they claim to be. Non-repudiation is another significant principle. In this context, it means that a sender cannot deny sending a message, and likewise, the receiver cannot deny receiving it. This stands to remove the possibility of false denial of a transaction or message transfer, which could otherwise leave space for digital dispute. Last but not least, the principle of Responsiveness comes into play. Cybersecurity mechanisms should have the propensity to respond immediately to any unauthorized intervention or possible threats. Quicker response times essentially mean threats are efficiently nipped in the bud before they can cause more damage. In essence, cyber security is not just about protection against threats—it’s a holistic approach amalgamating several principles to uphold the integrity, confidentiality, and availably of digital assets. It’s about building a digital fortress strong enough to withstand the ever-evolving saga of cyber threats while ensuring smooth sailing for authorized users navigating its currents. Evidently, in the breakthrough era of technology, where much of our lives and identities are tied to the digital realm, it becomes indispensable to incorporate these principles, utilizing technology not just to exist but to persist and thrive. Emerging Cyber Threats Safe, Secure, Savvy – Navigating Today’s Emerging Cyber Threats As technology’s rapid pace continues to forge ahead, moving us into the future, it has become increasingly critical for us to stay abreast of the shifting sands of cyber threats. These ever-evolving risks are not only shaping the landscape of cybersecurity but also dictating the way we interact with and treat our tech. Moving past the essential touchstones, such as confidentiality, integrity, authenticity, and responsiveness, it’s important to delve into the emerging threats now redefining the boundaries of online safety and data protection. AI-fueled attacks, for one, are on the rise. Harnessing the power of artificial intelligence, tech-savvy cyber criminals now craft deceptive deepfakes and engage in intelligent phishing strategies, personalizing malicious material with a level of sophistication previously unseen. If not nipped in the bud, these AI threats could compromise sensitive data, making a robust cybersecurity infrastructure non-negotiable. Supply Chain Attacks, another rising concern, target less-protected elements in the supply chain network, causing havoc. The infamous SolarWinds attack in 2020 is a resonating example of such attacks’ potency. It emphasized the dire need for fortified security measures across the entire supply chain trajectory. Cloud jacking also looms large – a cyber threat linked to organizations’ increasing reliance on cloud services. By infiltrating these digital spaces, hackers gain illicit access to valuable information. Therefore, implementing the latest cloud security protocols is central to maintaining non-repudiation and ensuring the availability of services. Lastly, let’s turn our gaze towards the threat of 5G-enabled attacks. As 5G edges us towards a new era of connectivity, it concurrently expands the risk scope for cyber threats. The increased number of connected devices on this high-speed network provides more entry points for hackers, mandating stringent safety measures for IoT devices. The updated threat landscape unquestionably poses unprecedented challenges to data security, demanding a shift from traditional security strategy to more comprehensive, highly advanced solutions. Recognizing these threats early and developing robust security measures is crucial, as is understanding which technology tools provide suitable defenses. Staying ahead of these threats requires ongoing vigilance, horizon scanning, and building security measures that are as dynamic and evolving as the threats they are designed to counteract. Moving forward, armed with the right knowledge and tools, we can respond to these threats proactively, reinforcing our defenses and ensuring we are well-equipped to navigate this continually evolving cyber landscape. Let it serve as a reminder to never be complacent in the face of technology, keeping the balance between harnessing its power and mitigating its potential risks. Advanced Cyber Security Technologies Delving into the world of cybersecurity, it is critical to underscore the role of advanced technologies in optimizing security measures. With innovations such as Machine Learning, Blockchain, and Quantum Computing transforming the landscape of cybersecurity, their efficacy hinges significantly on their seamless deployment and integration. Machine Learning (ML) and Artificial Intelligence (AI) offer game-changing transformations in cybersecurity. Rooted in the capability to catalyze threat detection, their proficiency lies in learning from historical cyber-attack patterns and subsequently predicting similar vulnerabilities. Machine Learning no longer waits for a human to identify and address cyber threats. It tracks and neutralizes threats real-time, greatly reducing the potential impact and severity of security breaches. Replicating the human brain via algorithms allows ML to sift through voluminous data during threat analysis, something that would ordinarily take humans innumerable hours or even be impossible due to sheer volume. It is also essential in identifying zero-day vulnerabilities – threats not yet identified in the system- by flagging unusual activity. Leveraging ML and AI, cybersecurity shifts from a reactive to a proactive stance, predicting and mitigating threats before they materialize. Pivoting to Blockchain, a buzzword often associated with cryptocurrency, is now turning heads in the cyber defense sector. Blockchain brings forward an immutable and decentralized ledger system, making data tempering an immensely strenuous process. Blockchain’s capacity to encrypt each transaction and link it irreversibly to transactions before and after it leads to an indelible audit trail, bolstering data integrity. Distributed Denial-of-Service (DDoS) attacks, prevalent cybersecurity threats, can also be mitigated by Blockchain through its decentralized nature. Savvy attackers often target centralized servers for a successful DDoS attack. Should these servers be replaced with a plethora of nodes (as in a blockchain network), attacker’s missions get complicated, and DDoS threats get substantially neutralized. Crowning the list of advanced technologies is Quantum Computing. In its basic sense, quantum computing revolves around quantum bits or qubits, offering enormous computational power. For cybersecurity, quantum computing promises ultra-secure communication links, commonly referred to as Quantum Key Distribution (QKD). In the event of any attempt at intrusion during communication, QKD facilitates the alteration of the encryption key, rendering the stolen data useless. While these technologies offer a wealth of advancements in cybersecurity, they are not entirely infallible. ML and AI can be tricked by skilled attackers into recognizing malicious code as benign. Blockchain, despite its touted security parameters, is still susceptible to a ‘51% attack,’ wherein an entity gaining control of the majority of the network could alter the data. The implementation of Quantum Computing, for all its potential, is still in the early stages, with adoption and optimization issues yet to be ironed out. To conclude, the technological warp speed at which we are operating demands an even more accelerated focus on cybersecurity. Leveraging advanced technologies serves as a powerful tool in contending with cyber threats. However, ensuring these technologies’ secure and effective implementation remains pivotal to realizing their full protective potential. “In cybersecurity,” as they say, “Innovation is the key, but vigilance is the doorkeeper.” Case Studies on Cyber Security Breaches In dissecting the high-profile cybersecurity breaches of recent years, a critical learning can be elucidated: the irreplaceable role of cybersecurity education and awareness. Cybersecurity is not a responsibility that solely resides with the IT department but extends to all personnel who access sensitive data and systems. The Yahoo data breach in 2013 serves as a painful reminder of this point, with 3 billion user records getting compromised due to an intricate web of deceptive phishing attempts and insider negligence. Constant training and information updates about the evolving nature of threats can help build a fortified human firewall. A well-informed workforce can serve as the first line of defense against phishing and other social engineering maneuvers. Emphasis should be applied across the organization – from the new intern to the top leadership. Everyone should understand the potential risks of lax cybersecurity practices. Further, the risk of legacy systems has been highlighted time and again in significant breaches. The WannaCry ransomware hits in 2017 dramatically demonstrated how outdated systems could cripple organizations globally. Intriguingly, the victims were not obscure companies with no IT budget but established entities such as NHS, FedEx, and Deutsche Bahn. Lesson? Transitioning to updated systems is non-negotiable for optimum cybersecurity. Strategically replacing or patching older systems can pay dividends in thwarting breach attempts. In line with this, prioritizing continuous security improvement is crucial. Cybersecurity is not a ‘set-it and forget-it’ arena. It demands continuous monitoring and improvement strategies. Case in point, the infamous Equifax breach in 2017 occurred due to a known vulnerability in the Apache Struts software. A diligent vulnerability management program entailing regular patching and updates could have averted the staggering loss of 147 million records. Beyond software, stronger emphasis should be placed on encryption and secure data-at-rest policies. The Uber data breach of 2016 is a testament to the importance of data encryption. The personal data of 57 million drivers and customers was left unprotected in cloud data storage, leading to disastrous consequences. Incorporating sturdier encryption techniques and regularly auditing to verify their proper implementation is a lesson worth applying. Lastly, the central theme recurring in all significant breaches is the inevitability of incidents – it’s a matter of ‘when’ more than ‘if’. Hence, continually building a robust incident response and recovery plan is essential. The Equifax breach resulted in an excruciating 76-day response time period due to a poorly managed incident response – a costly misstep. A well-structured incident response plan that includes clear communication protocols can help control damage and restore operations promptly. Collectively, these lessons reinforce how keeping pace with evolving threat landscapes cannot be achieved with technology alone. It requires a well-rounded strategy involving up-to-date systems, continuous security improvement, robust incident response plans, and, above all – perpetual cybersecurity education and awareness. Future of Cyber Security Progressing into the future, cyber security is heading towards a more automated, intelligent, and predictive manner of protecting systems and data. With the rise of more advanced, increasingly sophisticated threats, high-tech advancements like the development and implementation of behavioral analytics in data protection are becoming non-negotiable. Behavioral analytics plays a significant role in predicting future cyber attacks by learning from previous compromise attempts. This technology employs machine learning algorithms to recognize patterns and identify irregular activities. By analyzing vast amounts of log data, behavioral analytics can differentiate between regular and anomalous behavior while pinpointing potential threats. Another clear trend involves the coordination and integration of different security platforms to improve detection and response times. As cyber-attacks grow more complex, responding to a threat detected by one tool with another can be time-consuming and inefficient. Hence, the future lies in security platforms that communicate with each other, driving towards a more cohesive, streamlined cybersecurity response system. Furthermore, the Zero Trust model drives the future of cybersecurity by dismantling the traditional “trust but verify” method, instead following a “never trust, always verify” mentality. As organizations increasingly adopt remote working, the traditional corporate network is disappearing and, with it, network-based security. Zero Trust offers a solution supporting a distributed network while ensuring every request is thoroughly authenticated, authorized, and encrypted before access is granted. Meanwhile, homomorphic encryption is an advanced cryptographic method promising to revolutionize data privacy. Unlike traditional encryption techniques, which require decryption before conducting operations, homomorphic encryption allows computations to be performed directly on encrypted data, providing result that, when decrypted, matches the outcome of operations performed on the plaintext. This method dramatically enhances data privacy and security, particularly when processing sensitive data in public clouds. The future also spotlights secure access service edges (SASE). Integrating network and security services into a single cloud-based service, SASE converges functions like Secure Web Gateways, Firewalls as a Service, and Zero Trust Network access. Offering ease of management and upgraded security, SASE solutions are primed to be a significant piece of the cybersecurity puzzle. While we anticipate these promising advancements, it’s paramount to acknowledge continued risks. No technology is foolproof, hence the need for cybersecurity professionals to remain updated on the latest threat intelligence and adapt their security measures accordingly. Future-facing cybersecurity will be characterized by perpetual learning and the ability to adapt swiftly to new challenges. Admittedly, the cybersecurity landscape is evolving at a rapid pace, fuelled by emerging technological innovations. While these advancements present opportunities for improved security measures, they also heighten complexity and potential vulnerabilities, requiring cybersecurity professionals to remain agile, ever-vigilant, and ready for whatever comes next. After all, cybersecurity isn’t a final destination but a never-ending journey. As we look ahead, we can expect a future punctuated by increased intelligence, integrated responses, and security that’s woven into the very fabric of technology. As we persist in the Age of Information, fortification and resilience against cyber threats will increasingly determine survival and success. Equipping ourselves with a strong understanding of cyber security principles, awareness of common and emerging threats, practical knowledge from past breaches, and anticipation of future trends are all essential for traversing the ever-volatile cyber landscape. While technological tools enhance our capabilities, they are only as foolproof as the strategy and knowledge guiding their utilization. This exploration provides a strong and rounded perspective on cyber security, serving as a beacon for individuals and groups invested in preserving the sanctity and safety of the cyberspace thereby steering the helm of the digital future confidently.
<urn:uuid:f430a645-f1d9-482a-81bf-6ed20a99f4bd>
CC-MAIN-2024-38
https://cyberexperts.com/in-depth-look-at-cyber-security-computer-technology/
2024-09-07T20:50:08Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00520.warc.gz
en
0.923285
3,258
3.296875
3
Global Skills Report details growing rise in global AI and ML competency Finland has emerged as the world’s top data science nation, with the Nordic country “leading the way in AI education,” according to a report from online learning specialist Coursera. The Global Skills Report highlights efforts by the Finnish government and the University of Helsinki to teach 1 percent of the country basic AI skills as a catalyst for it becoming a leader in machine learning. The report draws on performance data from more than 77 million learners, 4,000 campuses, 2,000 businesses and 100 governments recorded since the start of the pandemic. Finland outranks Nordic neighbor Denmark in both global and European rankings, with the Danes following closely behind. The Fins, as well as Belarus, Bulgaria, and Italy, all outperform Russia, which led in cloud computing skills in 2020; the European nations are set to take advantage of the European Commission’s new €2 billion investment in common data spaces such as the Connecting Europe Facility 2 (CEF2). Coursera notes that eight of the top 10 global data science nations are from Northern and Western Europe – with the value of data in the EU expected to rise to around €550 billion by 2025, accounting for 4 percent of the bloc’s overall GDP. The UK is listed as “competitive” in both machine learning and data skills – but lags behind much of Europe, ranked as 34th globally – with the likes of Poland, Belarus, and Ukraine way ahead of the former EU member. The report praises the rapid development of ML capabilities in Serbia, which sits just one place below the UK in the global rankings, stating the Balkan nation holds "a number of complementary data science skills.” Northeast and Midwest ‘catch up’ to the West Coast Over in North America, the West Coast remains the leading region for data science – but the report reveals that the Northeast and Midwest are “catching up.” “Many top universities in these regions, such as Northwestern, MIT, and Carnegie Mellon University (CMU) have invested in recent years in new data science degree programs,” the report reads. Coursera found that learners in the South scored “particularly low” in mathematics. It suggests that the issue starts from an early age, with around 80 percent of Southern states being below the national average for eighth-grade math on standardized tests. Meanwhile, Canada is lagging in probability and statistics, with the report suggesting that this could be seen as part of a broader trend of declining performance in quantitative disciplines. Formerly a top-10 country in mathematics education, Canada has had its ranking and math scores “decrease consistently for the past 15 years.” Despite this, Canada is second to only the West Coast in terms of technology, with Coursera stating that no fewer than 500 global AI firms are currently headquartered in the country. Japan leads Asia in data science, while China ‘lags in tech skills’ In Asia Pacific, Japan is the leading nation in terms of data science. According to the report, the country’s aging population and increased healthcare costs provide it with “a unique opportunity” to utilize a “treasure trove of data.” Asian countries have invested heavily into national AI programs and those that did often perform well in technical education, the report states. Singapore is highlighted for the country’s National AI Strategy, which aims to invest $150 million into AI over five years. Coursera also notes China’s plan to become an AI world leader by 2030 – that envisions a domestic AI industry worth nearly $150 billion in just a few years. Despite topping rankings for databases and theoretical computer science, the report states that China “lags overall in technology skills.” The report references a study of undergraduate seniors in computer science, which found that Chinese students scored below US undergraduates, but were roughly in line with students in India and Russia. Compared to Asia Pacific, Central Asia was noted as lagging significantly behind its neighbors. However, some central Asian states are looking to address the gap, like Kazakhstan’s investment in a more secure high-speed infrastructure to transform and process data. About the Author You May Also Like
<urn:uuid:999a2523-0874-48f5-a603-a1823a5a7719>
CC-MAIN-2024-38
https://aibusiness.com/data/coursera-finland-is-the-world-s-top-data-science-nation
2024-09-09T01:36:11Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651053.52/warc/CC-MAIN-20240909004517-20240909034517-00420.warc.gz
en
0.962884
900
2.578125
3
How to Talk to Your Kids & Teens About Cyber Threats Starting a conversation about internet safety isn’t easy and comes with many challenges. How early should you start talking to them about predators and cyberbullying? How young is too young? Will they understand the significance of what you’re trying to teach them? Detective Cronister says kids and teens’ cyber awareness starts with you building relationships with them — the earlier (younger) the better. Communication is key to this process: - Be involved and active in your kids’ and teens’ lives — both on and offline. While parents often make it a point to get to know the friends that their kids and teens hang out with face-to-face, they don’t always take that same initiative in getting to know their online friends and acquaintances. - Familiarize yourself with common social media apps, especially those your children use. This is huge. How can you know if your kids are using apps safely if you don’t know how those apps work yourself? This is also helpful for teachers who want to teach kids about safe ways to operate online. - Be open with them and keep open lines of communication. Having open communication is key to any meaningful relationship — and the same can be said about the relationship between kids and teens and their parents. - Encourage them to ask questions and be prepared to answer them. Kids are innately curious and want to learn about things. Sometimes, this can lead them to search online for answers. Unfortunately, what they may end up finding can be dangerous or harmful. - Set clear boundaries and rules upfront. Establishing clear lines about what they can and cannot do online helps them know what’s acceptable and what’s not. Teach them about what you consider appropriate and inappropriate actions and behaviors. - Help them learn to trust their guts. If a situation or conversation doesn’t feel right, encourage them to talk to you or another trusted adult (such as a teacher). Nature gave us instincts for a reason, and helping kids learn to hone and trust their instincts is an important aspect of personal growth. - Control your reaction. The trick here is to not get angry or react if they come to you with an issue or tell you they’ve done something disappointing. This can be harder than you may think. Kids, like all of us, make mistakes. And reacting with a cool head goes a long way in establishing trust with your kids and teens. Talking to your kids about internet safety is a must in today’s connected world. According to Cronister: “We already do this as parents naturally when we talk to kids about real life strangers, right? And, so, I have always said that we already do this, we just also need to discuss it the same way on the internet — about internet safety.” So, how did he approach this conversation with his own kids? Cronister says he started talking with his kids about the differences between “good pictures” and “bad pictures.” “That way, they know, ‘hey, that’s something that Dad said was a bad picture and it’s making me uncomfortable, so I’m going to close this out or I’m going to tell my dad.’ And just exposing them to the thought process that there is such a thing as a good picture or a bad picture, a lot of times will give them enough of a conscious about those things that they’ll come tell you if they saw a good or bad picture.” Talking to your kids about internet safety is a must in today’s connected world. Be Part of Their Online World Dr. Kshetri also shares that it’s always a best practice for parents to be part of their kids’ and teens’ online social networks. It’s about gaining visibility into their “friend sphere” without being overbearing. “Maybe they can become their Facebook friends, for example. But if they become Facebook friends, they should let them know ‘I’m becoming your Facebook friend, but I’m not spying on you.’ So, this should be a clear line between us spying on children’s behaviors versus kind of observing in a way that is acceptable to them. So, I think they must be able to find some acceptable way so that there is less resistance from the children.” Privacy and Boundaries for Kids, Teens and Parents But what about concerns regarding giving your kids and teens privacy? Surely, parents shouldn’t invade their kids’ privacy and be overbearing, right? Both Cronister and Kshetri agree that there needs to be a balance between privacy and personal freedom in terms of allowing kids to grow. However, Cronister is quick to emphasize that there must be a limit in terms of what parents allow their kids do: “A lot of times, what I see with kids who are victimized is parents who have very few limitations on what their kids are allowed to do. So, it’s the 11-year-old girl who has a cell phone already and their parents think it’s just normal. They’re just supposed to have a cell phone, right? All kids have cell phones now. And the reality is, not really.” Cronister continues, saying that the threshold for age appropriateness is lowering in our culture right now. “I think the parents who are afraid of going against the grain in that respect are typically the ones who have problems because they don’t want to ‘invade their kids’ privacy’ because it’s good for them to have their own privacy. And I agree — to an extent. But I think there’s a middle ground: You can give your kids opportunities to have privacy and to be responsible, but at the same time not give them so much freedom that they make mistakes that are unfixable.”
<urn:uuid:b8a67579-914a-42e8-8ef1-300d82583403>
CC-MAIN-2024-38
https://dev.codesigningstore.com/talking-about-cyber-threats
2024-09-10T08:30:00Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651224.59/warc/CC-MAIN-20240910061537-20240910091537-00320.warc.gz
en
0.96828
1,266
3.71875
4
By Neal Bellamy, IT Director at Kenton Brothers Today, I want to talk about credential technology. While not an extremely exciting topic, it can be, and often is the weakest link in many organization’s access control system. Remember that an attacker doesn’t need to get through every defense in your system, most often they just need to get past the weakest one (or two). Let’s start with how cards and readers work. Any RFID reader, including the ones used for access control, puts out an electromagnetic field around the reader. This field is usually measured in inches, but in special readers like a Nedap long-range reader, fields can be measured in feet. When a credential (card, fob, wristband, sticker, etc.) passes through the field, it electrifies the antenna giving the chip on the credential enough electricity to transmit the data stored on the chip. Most often the data that is stored is the “Card number”. I put it in quotes because that “Card number” could be many things. Next, we need to talk about card numbers or more specifically card formats. Unfortunately, most card formats are simple and relatively easy to guess. The most common card formation is 26 bits in length. HID calls this H10301. The first 8 bits designate the facility code and the next 16 bits designate the card number itself. The facility code is a way to group the cards together and in theory, verify that the card belongs to the access control system. The low bit count means that there are only 256 possible facility codes and 65,535 card numbers. For those people paying attention to the details, the extra 2 bits are used for error checking. Most people start with card number 1 and work their way up. There are other card formats like 33-bit, 37-bit, 40-bit, and so on. Each increases the possible facility code and card number options. The important takeaway is that once an attacker has the card format, facility code, and card number of a person who has access, they can gain access to your facility. Like most things in commercial security, encryption is a way to combat the wrong people seeing the real card number. Encryption and card formats are independent of each other. You can have a 26-bit card that uses encryption and a 26-bit card that does not use encryption. That is based on the card technology. Card technology like Prox and Indala are not encrypted. This means that almost any card reader can read the actual card format, facility code, and card number, it just has to get close enough to a card that has access. Some technologies are encrypted but have already been cracked. Examples of these are Mifare Classic, HID iclass Classic, etc. Because the technology is already cracked, there are several ways of reading the encrypted data, and then applying the workaround to get to the actual card data again. Using a cracked technology is better than unencrypted, but it is still not advised. Some technologies are not yet cracked like Mifare EV3 and HID iclass SEOS. When an encrypted technology is in use, both the card and reader must be using the same set of keys. Public/Private key is a long topic, but effectively a matching pair of keys are used to encrypt and decrypt data. (More information here.) This means that readers and credentials are matched for the different manufacturers. If you are using HID readers, you almost always need to use HID credentials. Even with an encrypted, uncracked, card technology, the most commonly sold readers and credentials use the same key pair across all readers and credentials. This means that anyone can buy the latest HID reader to read almost every HID card ever sold. There are special programs where a business can “own” its own set of keys. Another option is to use a system that generates a unique key and then can use that key to encrypt the cards specifically for a given system like Gallagher. I know this was a lot of information, so let’s distill it a bit. First, make sure you are using encrypted card technology. Second, use the latest technology when you are using encryption. This will be based on the card readers you are using. Finally, if at all possible, own your public/private keys. Sign up for a unique key system like Corporate 1000, or use a system like Gallagher to generate a unique key for your system. If you have more questions or need help with your current/future commercial security solution, please give us a call.
<urn:uuid:d41a50d9-4555-48ab-af45-1ab74e1a1706>
CC-MAIN-2024-38
https://kentonbrothers.com/access-control/credential-technologies-you-may-not-be-as-protected-as-you-think/
2024-09-11T14:24:49Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651387.63/warc/CC-MAIN-20240911120037-20240911150037-00220.warc.gz
en
0.947188
957
2.578125
3
That computer your company is getting rid of might not be as clean as you think it is. In fact, it might still be holding critical financial, legal and personnel information — even as it goes out the door into someone else’s hands. A recent study by two MIT graduate students found massive amounts of sensitive data on old corporate systems donated to charity or stock piled at salvage companies. And 74 percent of the drives contained old data that could be recovered and read. Even 60 percent of users had reformatted the drives before getting rid of them, the systems still contained old data that could be recovered. Credit card information, patient data, financial records, love letters, pornography… you name it, it was found. Reformatting, it turns out, does not properly sanitize a disk because the Windows ”format” command doesn’t overwrite every block. ”The format command just reads every block to make sure that they still work,” says Simson Garfinkel, a graduate student at MIT’s Laboratory for Computer Science, which conducted the study. ”To properly sanitize the hard drive, you need to overwrite every block.” Here, we’ll look at the various tools that address this problem. They come in two broad categories — file wiping tools that wipe specific files or data, and drive scrubbers that completely remove all data from a hard drive. At the corporate end of the market, products focus on scrubbing drives before the company gets rid of the machines or hands it over to a business partner. Lawyers, for example, might have to give a laptop over to a rival attorney. The last thing they want is for old files to be dug up, particularly anything related to the current case. So in this situation, the drive is completely scrubbed of all data, then the relevant files can be loaded onto the computer. Many companies also have a policy mandating the scrubbing of machines before they go out the door. Some IT managers retire computers every three years, and recycling them has become a corporate mandate in some industries. Scrubbing tools, therefore, are now in greater demand than ever, in order to remove all vestiges of data before the computer is handed over to charitable organizations or otherwise disposed of. One of the leading companies in the area of drive scrubbing is New Technologies Inc. (NTI) (http://www.forensics-intl.com). NTI does not serve the consumer or individual market, and caters only to the largest corporations, law enforcement and government agencies. Its Data Elimination Suite is used by many Fortune 500 companies, large government agencies and law enforcement agencies. NTI also has a version of its scrubber that leaves the operating system on the drive but removes all data files. One enterprise alternative is Infraworks Sanitizer (http://www.infraworks.com/it_sanitizer.htm). This product has established a presence in the government market and has some big name corporate clients, such as Ford. Other scrubbers that appear to be geared to the corporate marketplace include PowerQuests DataGone (http://www.powerquest.com/datagone/) and Ontracks DataEraser (http://www.ontrack.com/dataeraser/). The consumer side of the market is dominated by a long list of products dealing mainly with wiping individual files, though a few also address complete drive scrubbing. Consumers, after all, are most typically interested in removing specific files from a computer. A thief, for example, was found guilt of robbery after getting away with the cash. His mistake? He typed up a note on a computer that he showed to the bank teller to order her to hand over the cash. Although he was smart enough not to save the file — he simply typed it up, printed it and then didn’t save it — a copy of the print file was recovered by investigators. He went to prison. There are way too many wipers on the market to list them all, so here we’ll only mention a few. East-Tec Eraser (http://www.east-tec.com/) appears to be more thorough than most, clearing swap files and other hidden files. It includes an ‘Erase beyond recovery’ feature. Mares and Cos RMD (http://www.dmares.com/maresware/html/rm.htm), Jeticos BCWipe (www.jetico.com) and Briggs Softwares Directory Snoop (http://www.briggsoft.com) also appear to have merit. ”Out damn spot,” said Lady McBeth. But no matter how hard she scrubbed, the blood on her hands would never go away. Unfortunately, it can be the same for file wipers and disk scrubbers. As forensics techniques continue to improve, much data can still be recovered even from files that were deleted or overwritten. A deletion, for example, goes to the recycle bin. But even if you ”erase” it there, the file can sit on the disk for some time before being overwritten by a new file. If you become more thorough and employ a file wiper, some traces of data may still remain even after it has been overwritten several times. ”It is effectively impossible to sanitize storage locations by simply overwriting them, no matter how many overwrite passes are made or what data patterns are written,” says security expert Peter Gutman of the Department of Computer Science at the University of Auckland. For some enterprises multi-pass file wipers and disk scrubbers may not be enough. The best advice is to find a vendor already delivering file wiping/disk scrubbing servicing to reputable firms and have your needs assessed. In some cases, a good enterprise-class file wiper may be enough, or you may need a heavier duty disk scrubber. But in other cases, it may be advisable to destroy all disks prior to disposal to eliminate even the tiniest possibility of data getting into the wrong hands.
<urn:uuid:c9c0a8fe-e257-4e8c-a51b-7206ffae91d1>
CC-MAIN-2024-38
https://www.datamation.com/security/are-you-giving-company-information-away/
2024-09-11T12:42:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651387.63/warc/CC-MAIN-20240911120037-20240911150037-00220.warc.gz
en
0.951849
1,270
2.515625
3
Jailbreak detection is a security feature that identifies whether the restrictions by Apple on iOS devices have been bypassed. The purpose of jailbreak detection is to prevent users from using the app on a jailbroken device, which could potentially compromise the security of the app or the data it contains. Jailbreak detection is a useful security measure for mobile apps that handle sensitive data or perform critical functions. Apple has put a number of security measures in place to maintain the security of iOS devices. Some of the security mechanisms used by Apple include: Jailbreaking is a process by which users remove the restrictions imposed by Apple to maintain the security of iOS devices. Rooting is the process for doing the same thing on Android devices. Once a device is jailbroken, users gain root access to it, which means they can access and modify sensitive system files and settings that are not normally accessible to them. Here are some things that can be done with a jailbroken device: However, bypassing Apple’s security measures creates risks. It can open the door to malware and other security threats, as well as cause performance issues or even render the iOS device unusable. Jailbreaking a device creates several security issues. Key security concerns associated with jailbroken devices are listed below. When implementing jailbreak detection in iOS apps, there are a few common methods on how to detect jailbreak on iPhone devices (or for that matter, on any iOS device). Jailbreak detection should be part of Mobile RASP (Runtime Application Self-Protection) and data protection solutions. Jailbreak detection is not foolproof, and there are ways for users to bypass it even though it can be a complex and technical process that requires a good understanding of iOS system internals and programming. For example, users can install a jailbreak detection bypass tweak, modify the app's code (this requires some reverse engineering), or use a jailbreak detection bypass tool. That said, implementing jailbreak detection is an important piece of a multi-layered security strategy for iOS apps that handle sensitive data or perform critical functions. Doing so adds another protective layer against exploits by hackers. Blue Cedar Enforce, a component of Blue Cedar Mobile App Security, provides jailbreak detection. Being able to detect a jailbroken device and preventing a mobile app from executing on a jailbroken iPhone is one of many mobile app security features provided by Blue Cedar Mobile App Security, which also includes Blue Cedar Connect. Blue Cedar also provides an easy way to incorporate mobile app security, including jailbreak detection, into a mobile app. That is via Blue Cedar Enhance, Blue Cedar’s no-code integration service that adds new functionality to mobile apps without requiring a single line of code to be written. Blue Cedar Enhance integrates mobile app security into iOS and Android mobile app binaries, regardless of the libraries and frameworks that underpin these app binaries. Blue Cedar Mobile App Security and Blue Cedar Enhance are delivered by the Blue Cedar Platform, a CI/CD friendly SaaS solution that also provides deployment services, such as app import and code signing, to streamline delivery of secured mobile apps.
<urn:uuid:e909da47-a190-4c40-8056-c0f1bd9de448>
CC-MAIN-2024-38
https://www.bluecedar.com/mobile-app-security-technical-glossary/jailbreak-detection
2024-09-15T02:53:26Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651614.9/warc/CC-MAIN-20240915020916-20240915050916-00820.warc.gz
en
0.925588
638
3.015625
3
How Cybersecurity Really Works: Protect Your Digital Life 2024 In today’s world, our lives are more digital than ever. Cybersecurity is key to keeping our personal and work lives safe. But have you thought about how it works and what you can do to protect yourself? Let’s explore the world of cybersecurity and find out how to stay safe online. Do you feel your digital life is secure? Learn about the secrets of cybersecurity and how to keep up with changing threats. Take charge of your online safety and secure your digital future. - Understand the fundamental principles of cybersecurity and how they protect your digital assets. - Discover the latest cybersecurity threats and vulnerabilities that pose risks to your online activities. - Explore the cutting-edge technologies and best practices that form the backbone of modern cybersecurity. - Learn to navigate the complex landscape of cybersecurity regulations and compliance requirements. - Develop a comprehensive strategy for managing cybersecurity risks and responding to incidents effectively. Understanding the Fundamentals of Cybersecurity In today’s world, cybersecurity is key. Our lives are more connected to technology, making it vital to protect our digital stuff and keep our online presence safe. Cybersecurity uses rules, methods, and plans to fight cyber threats. It keeps our information safe, whole, and easy to get to. What is Cybersecurity? Cybersecurity is about keeping systems, networks, and data safe from unauthorized access or harm. It’s a mix of tech, processes, and people working together. This helps fight threats like hacking, malware, and data breaches. The Importance of Cybersecurity in the Digital Age Now, cybersecurity is a must for everyone – people, companies, and governments. With more devices online, cloud use, and smarter cyber attacks, it’s crucial. Good cybersecurity keeps data safe, protects important systems, and keeps users trusting the digital world. Learning about cybersecurity helps us all get better at staying safe online. It prepares us for the new challenges we face in the digital world. Cybersecurity Principles | Cybersecurity Techniques | Cybersecurity Threats and Vulnerabilities In today’s digital world, it’s key to know the threats and weaknesses in our online spaces. Ark Solvers looks into the common cyber threats and attacks we face. They also share ways to spot and fix these weaknesses. Common Cyber Threats and Attacks Cybercriminals are always coming up with new ways to get into our networks and systems. Some top cyber threats are: - Malware: This is harmful software that can harm or take over computer systems without permission. - Phishing: This is when scammers try to get sensitive info by sending fake emails or making fake websites. - Ransomware: This kind of malware locks up your files and demands money to unlock them. - Distributed Denial of Service (DDoS) attacks: These are when hackers flood a website or network with too much traffic to slow it down. Identifying and Mitigating Vulnerabilities Our systems and networks have weak spots that cyber threats can use to their advantage. Finding and fixing these weak spots is key to avoiding cyber-attacks. Here are some ways to do this: - Regular checks and tests to find potential weak spots. - Using strong network security tools like firewalls and systems to stop intruders. - Keeping software and systems up to date to fix security issues. - Training employees on cybersecurity will make them more careful. - Creating and testing plans for when something goes wrong. By staying on top of cyber threats and fixing weak spots, companies can better protect against cyber attacks. This makes their cybersecurity stronger overall. Cybersecurity Best Practices and Strategies In today’s world, keeping your digital life safe is key. At Ark Solvers, we offer a set of cybersecurity best practices and strategies. These help protect you from new threats. Implementing Strong Cybersecurity Measures It’s vital to have strong security steps in place. Here are some important steps to make your cybersecurity stronger: - Use strict access controls and multi-factor authentication to check who you are and keep out unwanted visitors. - Keep all software, operating systems, and apps up to date to fix security holes. - Use top-notch threat detection and response to find and stop cyber-attacks quickly. - Have and test detailed plans for when things go wrong to keep your business running smoothly. - Teach your team about cybersecurity best practices, like spotting phishing scams and staying safe online. Cybersecurity Best Practices | Cybersecurity Defenses | Cybersecurity Frameworks | Threat Detection | Access controls, multi-factor authentication | Software updates, patch management | Incident response, disaster recovery | Advanced threat detection, employee training | By using these strong cybersecurity steps, you can protect your digital stuff. This helps lessen the effects of cyber threats. It also makes your organization more resilient against new security challenges. Cybersecurity Technologies and Solutions In today’s digital world, cybersecurity technologies and solutions are key to keeping our networks and data safe. Tools like firewalls and encryption help protect our digital assets. These technologies are essential for modern cybersecurity. Firewalls, Antivirus Software, and Other Security Tools Firewalls guard against cyber threats by controlling network traffic. They help prevent unauthorized access. Antivirus software fights malware and viruses to keep devices safe. Other tools, like intrusion detection systems and VPNs, boost network security. They work together to protect our data and networks. Encryption and Data Protection Techniques Encryption turns data into a code that only authorized people can read. Techniques like AES and RSA protect sensitive info. This way, even if data is intercepted, it can’t be read without the right keys. Other methods, like access controls and secure storage, also protect data. They keep information safe, ensuring it’s not shared without permission. The need for cybersecurity technologies and solutions is growing. These tools help us fight cyber threats and keep our online world safe. By using them, we can protect our digital assets and stay secure in a connected world. How Cybersecurity Really Works Cybersecurity is more than just using antivirus software or firewalls. It’s a detailed way to protect digital assets and data from threats. It uses complex mechanisms and protocols for a strong defense. Encryption is a key part of cybersecurity. It turns information into a code only authorized people can open. This keeps sensitive data like personal info and business secrets safe. Access control is also vital. It decides who can see or use certain resources. This includes strong passwords, biometric checks, and more to keep systems secure. Security protocols watch for suspicious actions in networks or systems. Tools like intrusion detection and SIEM help spot and act on threats quickly. Keeping software updated and using secure settings is also important. These steps help protect against new threats. Knowing about cybersecurity helps people and companies make smart choices to keep their digital stuff safe. It’s about using the right strategies to fight cyber threats. Cybersecurity Risk Management and Incident Response Dealing with cybersecurity means you need a strong plan that covers risk management and how to handle incidents. We’ll look at key steps to lower cyber risks and how to deal with security issues. Assessing and Mitigating Cyber Risks Risk management starts with checking your organization’s weak spots. This means doing vulnerability assessments often to find where hackers could get in. Knowing these spots lets you make plans to make your security better. Having good security policies and training your team is key. Make sure your company has rules for keeping data safe, controlling access, and handling incidents. Teach your team about the latest ways to stay safe online to keep everyone alert and ready. Responding to Cybersecurity Incidents Even with the best efforts, cyber attacks can still happen. Having a clear cybersecurity incident response plan is vital. This plan should say what to do, who does what, and how to talk to others to lessen the blow to your online systems. Quickly finding and acting on threats is crucial. Use strong monitoring and alerts to spot and stop threats quickly. Make incident response guides that help your team manage containment, removal, and recovery. This ensures a clear and organized way to lessen the harm. Getting good at cybersecurity risk management and incident response helps protect your digital stuff and keeps your business running smoothly. Always be ready to change your plans to stay ahead of cyber threats. Network Security and Data Protection Measures In today’s digital world, keeping your network and data safe is key. At Ark Solvers, we know how important it is to use strong security and protection for your online stuff. We’ll look at the main ways to make your digital world safer. Encryption Methods for Secure Data Transmission Encryption is a big part of keeping your network safe. It makes your data unreadable if someone else gets their hands on it. We’ll talk about the latest ways to encrypt your data, like AES, SSL, and TLS. Secure Coding Practices for Network Protection Writing code safely is also crucial for network security. By doing things right in software development, you can lower the chance of hackers finding weaknesses. We’ll cover the main ideas of secure coding to help you make your network stronger. Network Security Measure | Description | Firewalls | Firewalls act as gatekeepers, monitoring and controlling the flow of traffic in and out of your network, helping to prevent unauthorized access and detect potential threats. | Virtual Private Networks (VPNs) | VPNs create a secure, encrypted tunnel for your internet traffic, allowing you to safely access your network and data from remote locations. | Intrusion Detection and Prevention Systems (IDS/IPS) | These systems monitor your network for suspicious activity and can automatically detect and respond to potential threats, helping to mitigate the impact of attacks. | Using these security steps can help protect your online stuff and keep your network safe. Always be on the lookout and update your systems to fight off new cyber threats. Ethical Hacking and Vulnerability Testing In the fast-paced world of cybersecurity, ethical hackers are key to protecting digital systems. They use their skills to find weaknesses, test how strong digital systems are, and help companies improve their security. The Role of Ethical Hackers in Cybersecurity Ethical hackers, or “white hat” hackers, protect the digital world. They find security flaws through penetration testing and risk analysis. This helps them show companies how to beat cybersecurity risks before bad guys can. Ethical hacking is about finding and checking for weaknesses in systems or networks. It includes detailed vulnerability testing to see how well security measures work. Ethical hackers also simulate cyber threats to see how well a company can handle them. Thanks to their hard work, ethical hackers help companies stay ahead of cyber threats. By fixing weaknesses and using strong cybersecurity plans, companies can keep their digital stuff safe. They protect their customers’ private info and keep their stakeholders’ trust. Cybersecurity Compliance and Regulations In today’s fast-changing digital world, keeping up with cybersecurity laws is key for all businesses. Ark Solvers explores the complex rules and standards for protecting digital assets and data. By grasping these rules, companies can shield themselves from cyber threats and keep their operations safe. Understanding Cybersecurity Laws and Standards Rules like the General Data Protection Regulation (GDPR) and the Health Insurance Portability and Accountability Act (HIPAA) keep changing. Ark Solvers guides companies through these rules. They make sure strong cybersecurity policies and firewalls are set up to prevent cyber attacks. Also, using encryption and other data protection methods is vital for staying compliant and keeping sensitive info safe. Ark Solvers keeps its clients updated on the newest cybersecurity laws and standards. This helps them stay secure, protect their digital assets, and show they care about responsible data handling. Being proactive with compliance builds trust with customers, partners, and regulators. It also helps companies stand out in the market. What is cybersecurity, and why is it important? Cybersecurity is about protecting digital systems and data from unauthorized access or harm. It’s vital today because we rely more on technology and the internet. This makes us open to cyber threats and attacks. Good cybersecurity keeps our personal info, money, and important systems safe from cybercriminals and hackers. What are the common types of cyber threats and attacks? Common cyber threats include malware, phishing, ransomware, DDoS attacks, and data breaches. Hackers use these to get into systems, steal data, disrupt services, or demand ransom. Knowing about these threats helps us fight back with better cybersecurity. How can I identify and mitigate vulnerabilities in my digital systems? Finding and fixing vulnerabilities is key in cybersecurity. This means doing regular checks to spot weaknesses that hackers could use. Once you find them, you can fix them by updating software, using security patches, and improving access controls. What are some best practices for implementing strong cybersecurity measures? To boost your cybersecurity, use strong passwords, enable two-factor authentication, and keep software updated. Back up your data often and train your employees on cybersecurity. Using a framework like the NIST Cybersecurity Framework can also help you plan a strong security strategy. What are some key cybersecurity technologies and solutions? Good cybersecurity uses tools like firewalls, antivirus software, encryption, VPNs, and SIEM systems. These help detect and stop cyber threats, protecting your digital assets and keeping your systems and data safe. How do ethical hacking and vulnerability testing help with cybersecurity? Ethical hackers, or “white hat” hackers, are key in cybersecurity. They test digital systems to find and fix weaknesses. By mimicking real cyber attacks, they help companies see their security level, test defenses, and make needed changes to stay safe.
<urn:uuid:cb453148-e0c1-43a0-a84d-8e8bb5439e56>
CC-MAIN-2024-38
https://arksolvers.com/how-cybersecurity-really-works/
2024-09-16T08:53:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651682.69/warc/CC-MAIN-20240916080220-20240916110220-00720.warc.gz
en
0.923938
2,884
3.21875
3
Securing Our Electric Power Grid Is Critical Highly complex infrastructure systems require protection against cyberattacks. Electricity is so much a part of our everyday lives that we really only think about it when it is not there. That is why it is so important to build better security for our national electric power grid and other critical infrastructure. The power grid is a highly complex system, with multiple layers of defense, backup systems, safety mechanisms, and human operators. These layers successfully protect the system from most single-point failures. As Professor Richard Cook points out in his paper How Complex Systems Fail, catastrophe requires multiple small failures joining together in a cascading effect. The 2003 blackout in the northeastern part of North America clearly confirmed this scenario, moving so quickly that it only took seven minutes from the initial failure to the full blackout – too fast for human operators to counter. It then took between two and seven days to restore power to customers. Change introduces new forms of failure. The power industry is continually upgrading and evolving its systems, from generation to delivery. Smart meters enable time-of-day pricing, connected thermostats can be turned down during times of peak demand, and renewable energy sources need to be constantly monitored to adjust for fluctuations in their production. A lot of this involves equipment that is network-connected. And network connections mean the potential for cyberattacks. Whether it is a gang of criminals trying to disrupt the electricity for extortion, terrorists attempting to damage it for headlines, or nation states attacking it as part of their intelligence or combat strategy, the end result of a successful attack is blackouts, economic damage, and potentially weeks or months of repair. And the risk of a successful attack is not theoretical, as repeatedly demonstrated by simulated attacks, field trials, and cyberwar games, dating back to at least 2007. In our Internet of Things Security Solutions Group, we have been actively working on better protections for the electric power grid and other critical infrastructure. Our work with the Center for Strategic and International Studies (CSIS) has shown that this is a real and present danger. Of the 200 organizations from around the world that we surveyed, 85% have experienced network infiltration, 65% frequently find sabotage-capable malware on their systems, and 25% have been subject to cyber-based extortion. Building security into the power grid is challenging, due to the importance of service availability and the amount of legacy infrastructure. Since December 2013, we have been field-trialing a joint project with Wind River for critical infrastructure protection at Texas Tech University, where our solution withstood penetration testing and protected the system from the Heartbleed vulnerability and Havex attacks. This solution, developed in collaboration with the Discovery Across Texas smart grid project, separates security management from operations, providing device identity, malware protection, and data protection in a secure platform. By understanding the needs of the industry, the solution works with both new and legacy infrastructure, with little or no changes to business processes or application software. Electricity is critical to the daily operations of people, businesses, and governments around the world, and we need to improve its defenses against malicious attacks before some criminal group decides to demonstrate its capability to make us powerless. Read more about: 2015About the Author You May Also Like How to Evaluate Hybrid-Cloud Network Policies and Enhance Security September 18, 2024DORA and PCI DSS 4.0: Scale Your Mainframe Security Strategy Among Evolving Regulations September 26, 2024Harnessing the Power of Automation to Boost Enterprise Cybersecurity October 3, 202410 Emerging Vulnerabilities Every Enterprise Should Know October 30, 2024 State of AI in Cybersecurity: Beyond the Hype October 30, 2024[Virtual Event] The Essential Guide to Cloud Management October 17, 2024Black Hat Europe - December 9-12 - Learn More December 10, 2024SecTor - Canada's IT Security Conference Oct 22-24 - Learn More October 22, 2024
<urn:uuid:ee5d6fac-1e22-40cf-bc57-49eef89c7e29>
CC-MAIN-2024-38
https://www.darkreading.com/cyberattacks-data-breaches/securing-our-electric-power-grid-is-critical
2024-09-16T10:07:35Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651682.69/warc/CC-MAIN-20240916080220-20240916110220-00720.warc.gz
en
0.941596
806
2.5625
3
The idea of creating visual diagrams of network infrastructure in order to understand the relationships between component systems is decades old. However, most network visualisation tools have stuck to just two dimensions. Now, though, a UK-based company is introducing the third dimension, and it is using technology originally developed for computer games to make it possible. The technology in question is the brainchild of Dr Chris Doran, a Cambridge University astrophysics researcher. Doran founded Geomerics, whose graphics technology is used in video games to create realistic lighting effects. After winning investment from the UK’s Technology Strategy Board, Geomerics partnered with Cambridge IT consultancy Intergence to apply the same technology to network visualisation. The resulting product is called Hyperglance, and the company spun off to market it is named Real Status. Real Status CEO Royce Murphy argues that 3D visualisation gives clearer insight into networking infrastructure than the existing 2D systems. “The problem with 2D is that no matter how small you make the lines connecting the systems, those lines inevitably intersect and overlap,” he explains. “When you have a couple of hundred devices on the network, it gets very complicated and it is no longer decipherable. With a 3D modelling approach, you can navigate through the infrastructure model and see everything from the inside out.” The system works by collecting data from standard network discovery tools and third-party systems management platforms, and algorithmically calculating the best way to construct a 3D model of all the components. “One way to understand how the model is laid out is to imagine that all the network devices have a positive charge and want to repel one another, but the network links are tying them together,” explains Royce. The system detects both physical and virtual systems, and Royce says one of the most valuable uses of Hyperglance is mapping the interdependencies between the two. Many of Real Status’s advance customers (the product is still in beta testing) use it in conjunction with VMware’s virtualisation management tools to do just that, he says. “The system allows you to say, ‘I want any virtual connection to be a dotted line, I want a physical connection to be a solid line’. You can make virtual connections to storage orange, say, and virtual connections to CPU green,” Royce explains. “By using those visual cues, you can start to make sense of information that is just too complicated for current modelling techniques.” It can also be used to alert systems administrators to issues that need immediate attention, such as a security breach on a business-critical system. “Every company will have intrusion detection software, patch management software and some way of rating the relative importance of all their business systems, but the output from all of these will probably be log files or spreadsheets,” he explains. “Hyperglance allows you to create an expression that says if an intrusion is detected, and there are known patch management issues, and the system in question is business-critical, then make the node flash red on the model right away.” There is a danger for Real Status that Hyperglance might be dismissed as a toy – something nice to have but not essential – and its roots in video game technology may accentuate that impression. But Murphy insists that 3D visualisation makes a genuine difference to real-time systems management by capitalising on the pattern recognition capabilities of the human mind. “By visualising data from multiple systems on a single model, we allow people to spot correlations on the fly, rather than write correlation-detection algorithms for every scenario they can possibly imagine,” he says. “It’s that ability to correlate data from multiple systems and take decisions at the speed of sight that is the real benefit here.”
<urn:uuid:70e06c15-4031-4de1-83bc-b2c2de895b7a>
CC-MAIN-2024-38
https://www.information-age.com/real-status-applies-3d-video-game-technology-to-network-visualisation-26268/
2024-09-17T15:29:59Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651800.83/warc/CC-MAIN-20240917140525-20240917170525-00620.warc.gz
en
0.938681
799
2.78125
3
Welcome to the newest edition of Cyber Lingo, “Social engineering definition and examples”. As the title explains, today we’ll tackle one of the most common tactics used in the cyber security world: Social engineering. Although it is a common framework for cyber criminals, not all employees know what social engineering is. By knowing the definitions, identifiers, and examples of social engineering employees can be better prepared to spot social engineering attacks and stop them in their tracks. Let’s get started: Social engineering is a manipulation technique that cyber criminals use to trick victims into taking a certain action, typically sharing personal information, handing over money, or giving access to a network. It relies on psychological manipulation rather than technical hacking methods, making it a highly effective tactic for attackers. Cyber criminals build trust and then rely on emotions of fear, uncertainty, and urgency to force individuals into taking the wanted actions. These tactics can be used in a multitude of ways including over email, on a phone call, or in person. For just $325 USD, you can run a 6 week, automated program for gamified phishing awareness training and challenges. (Limited time offer. Normally valued at $999 USD) Use Promo Code: 6WEEKS Take proactive steps to invest in your business’s cyber resilience now to protect your organization from costly data breaches and disruptions. Start easily with our Quickstart Training Bundles. To learn more CLICK HERE. In almost all phishing attacks, there is some form of social engineering being attempted, to get you to trust the sender and take an action that benefits the attacker. In many cases, the attacker is impersonating somebody or a brand you trust. Social engineering terms often appear in discussions about cyber security training, policies, and news. You’ll likely have a training module and certain policies dedicated to only social engineering. This term is also prevalent in any news about data breaches and cyber attacks, so you’ll likely see this term in headlines online and during in-person discussions. If you work in a business environment or have an email or social media accounts, social engineering is used against you regularly. Some key environments to be aware of social engineering tactics would be: You are particularly vulnerable to social engineering attacks if it’s a busy time for your business, there are external distractions (ex. Tax season), or you are expecting to see or hear from people you don’t know. The key is to always be on alert for social engineering scams. Constantly being vulnerable to social engineering attacks may sound scary. But, there are ways you can protect yourself from falling victim to these advances. Here are some common tips for spotting and fighting against social engineering: One of the biggest social engineering breaches in history was the Sony attack of 2014. It is believed that this attack started because cyber criminals were able to use social engineering techniques to trick Sony employees into giving login credentials over email. Using those credentials, the cyber criminals gained access to Sony’s network and released vast amounts of sensitive data including employee information, unreleased movies, and internal communications. It was months of threats and releasing data until Sony was able to gain full control of their networks. Even after this, Sony still dealt with employee lawsuits and a poor reputation. Social engineering represents a significant threat in the realm of cyber security, exploiting human psychology to bypass technical defences. Understanding the definition and various applications of social engineering is crucial for developing effective countermeasures. By raising awareness, implementing robust security practices, and learning from real-life cases, individuals and organizations can better protect themselves against these sophisticated attacks. As the landscape of cyber threats continues to evolve, staying informed and vigilant remains our best defence against social engineering.
<urn:uuid:7994ef84-b8a3-4964-b5e8-b9c3cbf96fc4>
CC-MAIN-2024-38
https://clickarmor.ca/cyber-lingo-social-engineering-definition-examples/
2024-09-08T00:36:14Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650926.21/warc/CC-MAIN-20240907225010-20240908015010-00620.warc.gz
en
0.948513
765
3.21875
3
Federal agencies are increasingly moving their mission-critical systems and solutions to the the cloud: More than 65 percent report that some or all of their data is in the cloud, and the percentage is rising every year. While the cloud can provide cost-effectiveness, scalability and agility, agencies must take steps to ensure their data and applications will be readily available in the event of a disaster. In the past, planning for disaster recovery focused on rare occurrences: floods, power outages and the like. Today, the need for quick recovery is far more common, given the prevalence of ransomware attacks. Backup as a Service provides reassurance that data is protected in these cases. Here are some best practices for implementing and integrating Backup as a Service: The Most Crucial Element of Disaster Recovery Backup plans spell out exactly what should be backed up, how often, where backups should be stored and for how long. Often overlooked, however, is a detailed recovery plan, the most crucial element of disaster and recovery planning. Based on a risk assessment for critical functions that prioritizes the risk posed by cyberattacks, human error, software bugs, natural disasters and the like, the recovery plan should be developed long before disaster strikes. Such plans should spell out the policies and procedures that must be taken before, during and after a disaster, and must specify recovery point objectives (how much data loss is acceptable) and recovery time objectives (how quickly systems can be up and running). A critical step in recovery planning is checking the backups. Many organizations check backup quality and integrity early on, but then may go months without verifying that subsequent backup jobs — which run to completion — have backed up anything successfully. Click the banner below to learn how Backup as a Service boosts data protection. Adhere to the 3-2-1 Rule To ensure speedy recovery, multiple copies of data and applications are needed, both locally and remotely. The 3-2-1 rule states that you should have three copies of your data (your production data and two backup copies) on two different types of media (such as disk and tape), with one copy kept offsite for disaster recovery. Even if you use the cloud for backup, you should consider mirroring crucial data, just to be safe. Consider a situation where the cloud provider is inaccessible, when you are unable to access your network or when you lose internet access altogether. Another reason to choose an extra backup-and-restore method is that it takes a lot of bandwidth to restore entire systems, so full data recovery could take a long time. Having an alternative method could speed up recovery. Finally, cybercriminals often spend time inside the network elevating privileges and actively seeking out backups — including cloud backups — so they can delete them before launching a ransomware attack. They might even delete shadow copies stored locally, which would lengthen time to recovery or in some cases make it all but impossible. Apply Patches and Run Simulations to Avoid Devastating Downtime Disaster recovery often focuses on dealing with lost or inaccessible data but should also consider unplanned downtime. While planned downtime for such things as maintenance or updates can be disruptive, unplanned downtime can be devastating. The recovery plan plays a key role here. The person or role accountable for recovery must make sure the contact list is always up to date, and should have a plan for notifying affected parties as quickly as possible and keeping them updated with status reports. Preparing for unplanned downtime also means considering additional costs, not just for IT but also for the entire agency. Incremental technology purchases may be called for. Vendor support costs may increase, as well as staff overtime to resolve the incident. How to reduce the chances of unplanned downtime? Preventive measures such as updating and patching critical systems and applications are key. A vital component of planning for the unexpected involves running tabletop tests, simulations or active trials of the recovery plan. This helps uncover gaps in planning or overlooked issues such as recovery sequences to restore multitier applications or those dependent on Domain Name System or authentication services. Ask the Right Questions Before Investing in Backup Services Managed backup, or Backup as a Service, can help agencies with a variety of services ranging from assessing backup needs and overseeing backup jobs to helping with recovery when needed. When selecting cloud-based backup services, agencies should ask potential vendors these questions: Does the solution integrate well with your own backup solution? Does it include archiving, mirroring or other options? Where is the data stored? In general, the further away, the longer the recovery will take. Find out how fast backups will transfer in the event of unplanned downtime. What data security methods are used? Look for, at a minimum, AES 256-bit encryption. Find out who has access to the private decryption key. Is the vendor compliant with privacy and security regulations? This is especially important where regulatory requirements come into play. What recovery methods does the vendor use? Is it possible to recover directly from the cloud, operate remotely or fail over a system without the vendor’s assistance? MORE FROM FEDTECH: This process helps agencies protect email backups. The Bottom Line for Backups Cloud-based Backup as a Service solutions can give you peace of mind that critical data and applications are backed up and will be available when needed. This can be a cost-effective solution to ensure your agency does not lose data or suffer from unplanned downtime. Follow best practices to make sure that your plans are solid and tested frequently, you have sufficient redundancy in your backups, you know what to do when you experience downtime, and that your vendor is providing the depth and breadth of services you need.
<urn:uuid:68e1f75e-7344-4cb2-bb01-7cfa3468f39d>
CC-MAIN-2024-38
https://fedtechmagazine.com/article/2023/08/follow-these-4-best-practices-cloud-based-backup-and-recovery
2024-09-11T17:39:29Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651390.33/warc/CC-MAIN-20240911152031-20240911182031-00320.warc.gz
en
0.945984
1,171
2.59375
3
2012: No Apocalypse, but Space Weather Threat? The year 2012 will bring the next peak of an 11-year cycle of solar activity, which can produce sunspots and solar flares that can interfere with satellite and radio communications on earth. Is your data center ready? A look at some common-sense preparations. November 10, 2009 This NASA diagram depicts solar radiation storms interacting with the magnetic fields surrounding the earth, which can disrupt power grids and communications. What does 2012 hold in store? Will the supposed end of the Mayan calendar trigger a global apocalypse? Or perhaps something slightly tamer, like huge storms on the surface of the sun sending waves of "solar wind" hurtling towards earth to disrupt our communications and power grids? This week Hollywood hopes the first scenario will sell a lot of movie tickets to its disaster film "2012" starring John Cusack. The second scenario, however, isn't as far-fetched as you might think. In fact, there's enough science involved that data center operators are being encouraged to educate themselves about the issue. The year 2012 will bring the next peak of an 11-year cycle of solar activity, which can produce sunspots and solar flares that can interfere with satellite and radio communications on earth. The most significant potential impact is to power grids, which are susceptible to disruptions during these magnetic storms. A Real Threat to Power Grids A solar storm during the 1989 caused the entire Hydro-Quebec power grid to fail, leaving millions of residents of northeastern Canada without power for up to nine hours. The event also caused a transformer failure at a nuclear power plant in New Jersey. Is space weather a threat to your data center? Eric Gallant of Lee Technologies has been working to raise awareness of the 2012 sunspot cycle, so data center operators can assess the risk and determine whether additional steps are needed to protect their power and communications infrastructure. Similar to Hurricane Prep Contingencies for space weather are "really similar to things you do for hurricane or other disaster," Gallant says. That includes ensuring that you have well-maintained backup generators, an adequate supply of diesel fuel, and fuel supply agreements in place to cover an extended outage. Data center operators may also want to assess their surge suppression systems or consider switchgear that can manage transient power surges. Gallant also suggested monitoring web sites like the NOAA's Space Weather Prediction Center to be aware of problematic solar activity. "You need to keep track of space weather just as you keep track of the regular weather," he said. About the Author You May Also Like
<urn:uuid:e245c4cc-b39b-4cde-b7ed-d70c72f03ec7>
CC-MAIN-2024-38
https://www.datacenterknowledge.com/data-breaches/2012-no-apocalypse-but-space-weather-threat-
2024-09-11T16:29:33Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651390.33/warc/CC-MAIN-20240911152031-20240911182031-00320.warc.gz
en
0.95164
531
2.90625
3
October 24, 2023 Repeater vs. Access Point: What’s the Difference? In this article, we delve into the differences and similarities between a repeater and an access point, exploring their functions, applications and benefits. In the realm of networking, the terms "repeater" and "access point" hold significant importance, especially when it comes to extending network coverage and enhancing connectivity. Both devices play crucial roles in optimizing network performance and ensuring seamless communication. In this article, we delve into the differences and similarities between a repeater and an access point, exploring their functions, applications and benefits. Repeater: Enhancing Signal Strength and Range A repeater, also known as a range extender or signal booster, is a networking device designed to enhance the reach and strength of an existing wireless network. Its primary function is to receive a weak Wi-Fi signal from an access point or router and amplify it, rebroadcasting the signal to extend coverage to areas with poor connectivity. The working principle of a repeater involves capturing the signal from the existing network and transmitting it at a higher power to reach distant or obstructed areas. The repeater essentially acts as a bridge between the original access point and the devices that need to connect to the network. Advantages of Using a Repeater - Increased coverage: Repeater amplifies the signal, enabling the network to reach farther and cover larger areas. - Improved signal strength: By boosting the signal, repeaters enhance the quality of the connection and provide a more stable link to devices. - Cost-effective solution: Repeater installation is generally less expensive than setting up additional access points, making it a budget-friendly option for extending network coverage. Access Point: Extending Network Reach and Capacity An access point (AP) is a device that connects directly to a wired network and facilitates wireless communication for devices within its coverage area. Unlike a repeater, which amplifies an existing signal, an access point creates a new Wi-Fi network or extends an existing one, providing a dedicated wireless connection to devices in its vicinity. When an access point is added to a network, it connects to a router or switch through an Ethernet cable, allowing devices within its range to connect to the network wirelessly. Access points are often strategically placed in areas with high user density or where a strong, reliable Wi-Fi signal is required. Advantages of Using an Access Point - Dedicated network: Access points establish a separate wireless network, ensuring optimal performance and capacity for connected devices. - Scalability: As the network expands, additional access points can be deployed to enhance coverage and accommodate more users without compromising performance. - Greater control: Access points provide network administrators with advanced control over security settings, bandwidth allocation, and other configurations to tailor the network to specific requirements. In comparing repeaters and access points, it's essential to understand that they serve different purposes within a network environment. Repeater acts as a signal amplifier, extending the reach of an existing network, while an access point creates a new or extended network, offering dedicated connectivity to devices. If you need to enhance coverage in areas with weak or inconsistent Wi-Fi signals without the need for a new network, a repeater is a suitable choice. On the other hand, if you're aiming to create a separate, reliable wireless network or expand an existing one, an access point is the preferred solution. CDW sells wireless access points from top brands, including Aruba, Cisco and Extreme Networks. Shop today and get exactly the wireless connectivity solution you need.
<urn:uuid:554be14c-2b58-4219-bf3c-45240ca64ef1>
CC-MAIN-2024-38
https://www.cdw.com/content/cdw/en/articles/hardware/repeater-vs-access-point.html
2024-09-19T02:14:58Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651944.55/warc/CC-MAIN-20240918233405-20240919023405-00620.warc.gz
en
0.922149
733
3.515625
4
What is Attack Path Management? Attack paths are the highways and paths attackers take to reach something of material value within your organization. Attackers do this using combinations of conditions ranging from exploitable vulnerabilities, to exposed identities, to misconfigured systems. To put it more plainly, it's the map to an impact that can occur if left exposed. Attack path management is the science of reducing or eliminating the potential for material impact within your environment. By understanding how attackers can take advantage of your environment's exploitability, you can know where and how to take away their opportunities to cause harm. Attack path management is an important part of an overall exposure management approach. Anatomy of an Attack Path An attack path can be represented by many different devices, users, and relationships. The most basic attack paths include devices and users. This is what we think of when a device on the internet is exploited and an attacker gains access to a valuable user such as one with an administrator account. Epiphany is constantly evaluating, much like a very evil Google Maps, the best route an adversary could take to get to a destination (a "prize") at any time. When you break down these attack paths though, the pieces are fairly straight forward and are comprised of: Devices. Any computer system capable of supporting an operating system. User identities. A representation of a person or system that interacts with the authorization and authentication system. Objectives. The reason an attacker would want to use a path (to reach a prize). You can think of this as something that would cause a material impact to your organization. In the basic attack path, the objective is an admin user. Relationships. How the devices, identities, and objectives matter to each other. Example relationships are "what is installed on what" or "who is using what device." Epiphany keeps track of many more types of objectives and relationships from your on-premises network equipment to your cloud assets. But the key is to understand these basics first. Traversing the Attack Path Attacks don't just suddenly appear in the environment and go from there. They take advantage of an exploitable condition to gain their foothold into the environment. To better illustrate this, it can be broken down into a few different representations: a direct exploitation attempt, a user exploitation attempt, and stolen credentials. They are described next. Direct Exploitation Attempt In direct exploitation attempts, the attacker is trying to use a known vulnerability in an application or operating system to gain access (a "presence") to a device. This is the classic example that the basic attack path illustrated (above) and is what we most think of when we hear about something being exploited. These types of attacks are commonly used against firmware of internet-facing devices, such as the Pulse Connect Secure VPN exploitation from 2021. This type of exploitation can also be used by ransomware to spread within a network using weaknesses in operating systems to exploit and install its malicious payload. These types of attacks are easier to execute with devices on public Wi-Fi, shared office spaces, and homes of remote employees because these networks usually lack the protections provided within the corporate network. Consider how your users access your data and systems and how their devices are exposed. User Exploitation Attempt User exploitation attempts, called social engineering or in some cases arbitrary code execution, are centered around being able to convince a user to execute some dangerous code that is meant to exploit the operating system or application on the user's device. The first stage of a lot of ransomware, called the dropper, is usually benign but is used to download malicious code that will take advantage of known vulnerabilities. The most common attack vectors for these types of attacks are phishing attacks through email that result in the user executing a trojan document such as a PDF, redirecting the user to a malicious site, or redirecting the user to a site to steal their credentials. Using stolen credentials is the exploitation equivalent of using the front door. Often ransomware groups and advanced persist threats (APT) will use credentials stolen through other breaches or social engineering attempts to log directly in your corporate VPN and work from there. The only meaningful way to protect against that is by using multi-factor authentication (MFA) to validate against a remote access system. The relationship between the nodes within an environment is one of the most important parts of understanding the attack path. Epiphany uses these relationships to give you the ability to see why an attacker might want to use a relationship to carry out its attack. The implications of each relationship and the impact to the business is what Epiphany constantly thinks about. As you become familiar with the platform you will see the steps Epiphany takes to simplify this. The illustration above shows some of the types of relationships Epiphany takes into account when building the attack paths you see within the platform. In this very basic example, there are multiple mechanisms for exploitation that Epiphany considers. The normal relationship in Epiphany is between a device and a user. Inside Epiphany, it is labeled with "Used By." This is to simplify the ability to read the path from left to right, but can actually be technically represented by the presence of a token, an active session, or a credential that could be used. Each of these mean different things to the attacker, its tools, and its objectives. Epiphany also illustrates the relationship a group could have to the "Admin to" relationship of a device. In this case Epiphany uses the administrator and backup operator group from Windows Active Directory to show that an attacker may have different ways to access the objective device by using the user credential stolen earlier in the path. Epiphany keeps track of many different relationships among multiple systems within a platform, including network management, identity management, vulnerability scanners, application inventories, patch managers, endpoint protection, and many more. The objective is the whole reason the attacker is going after your organization to begin with. Each group or automated malicious application has an objective. The key to defending your organization is to understand the material impact caused by an attack path. To think of it another way, what in your organization causes a five-alarm fire if it goes down? What applications, users, and devices support that? Managing Attack Paths Managing attack paths is all about understanding your potential exposure, which is a function of a threat's ability to take advantage of your organization's weaknesses to create a material impact. You could think of it like this: This can be described simplistically as follows: for every exploitable condition present that the adversary can use, multiply them and then multiple that by the number of critical applications, users, and systems those conditions can impact. This index, an exploitability index, can turn out to be quite a large number and seem somewhat overwhelming. This is where Epiphany simplifies the problem. Its engine spends all of its time understanding all the components that contribute to your organization's potential exposure and then ranks the most exploitable paths, from the adversary's perspective, first. This allows you to focus on the outcome you want to achieve, instead of analyzing the problem. Using the skills described above, it is possible to understand the attack path and pick the most meaningful relationship to focus on. Managing the attack path is all about working backward from the problem. The Overall Path While the attack path in the next image looks intimidating, you can understand it if you apply the skills described above and understand that Epiphany is doing a lot of the heavy lifting for you. By going to the end of the attack path, you can see what Epiphany has automatically found as exposed and understand why it's important to the business. This is what Epiphany wants to stop the attacker from being able to reach. Notice the jewel icon in the upper right of the device icon. Epiphany uses the jewel to indicate that a device includes a "prize." Pivot points are quite common in attack paths and usually occur because the attacker's objective is on the other side of the device or firewall. This could be access to a high-value account or a restricted network that gets the attacker to its objective. The Attack Surface All attack paths can trace their access back to some point on the attack surface. This foothold could be caused by a vulnerability, misconfiguration, or direct access such as with a stolen credential. The Remediation Recommendation Once you've analyzed the path, pick the remediation strategy you think is best for your organization. Epiphany allows you to do multiple things with a remediation including creating tickets, assigning it for review, notifying on reoccurrence, or accepting the risk. Epiphany has a built in "easy button" to take you to the relationship you should focus on first. Just click the anywhere you see it and it will display Epiphany's top recommendation list. The process for managing attack paths built by Epiphany is very straight forward and can be repeated for different paths and variations in the same way. Find the prize and understand how, if it were compromised, it might impact your business. Applications, devices, and users all have different values to a business. Keep this in mind as you proceed. Look for a pivot or other direct relationship to the attack surface. This, for example, could be a misconfigured firewall or a user with special permissions. In the example path above, it is email@example.com that gives this access and the attacker gets it by using the firstname.lastname@example.org account to access Richard's device. Next, trace back to the beginning of the path. This is where the attack initially gains a foothold. If this is caused by a vulnerability, Epiphany will prioritize the vulnerabilities you should remediate using Epiphany's Vulnerability Management engine. Finally, determine what fix you'd like to apply to the path to remediate it. Epiphany thinks the best course, by default, is to break the path. But this is not always practical for business operations. Sometimes you should also consider strategies for increasing your ability to monitor a high-risk situation or increase resistance to an attacker by changing defensive control settings.
<urn:uuid:189498a5-0413-4ee2-8c2b-376d7adc3459>
CC-MAIN-2024-38
https://docs.epiphanysys.com/technical-documentation/epiphany-workflows/technical-analysis/attack-path-management
2024-09-08T03:27:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650958.30/warc/CC-MAIN-20240908020844-20240908050844-00720.warc.gz
en
0.95216
2,074
2.828125
3
Created on 02-28-2005 12:00 AM Policy Based routing exampleIn the following example, client PCs that are on the Internal and DMZ2 networks must have all of their outgoing HTTP and HTTPS traffic re-directed to a HTTP Proxy on the DMZ network. The HTTP Proxy will then establish a connection to the Internet for these web client requests. All other traffic, which originates from these Client PCs must egress directly from the External interface, onto the Internet. Port1 and Port2 interfaces are configured with Static IPs. 'Ping Server' is configured on each interface. Two default routes with the same Distance, are also configured. For further information, refer to the related articles "Conditions and Caveats of Policy Based Routing (MR7 and MR8)" and "Conditions and Caveats of Policy Based Routing (MR9)". The route policy order presented below is important, as the rules are matched in the order of first to last entry.
<urn:uuid:98e1f2db-fd14-42d5-b751-38100a59a404>
CC-MAIN-2024-38
https://community.fortinet.com:443/t5/FortiGate/Technical-Note-Policy-Based-Routing-example-to-redirect-HTTP/ta-p/197987?externalID=10781
2024-09-13T00:48:58Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00320.warc.gz
en
0.935608
201
2.8125
3
The Rise of Heat-Pump Water Heaters in the US Market The United States is currently undergoing a significant transformation in its approach to energy efficiency, with heat-pump water heaters (HPWHs) emerging as a key player. These changes are being driven by a combination of new federal efficiency standards, stringent local environmental policies, and attractive government incentives. This article examines the factors that are propelling HPWHs into the spotlight and reshaping the water heating market, positioning them as the preferred choice for consumers who prioritize sustainability and cost-effectiveness. Traditional Water Heating and Its Shortcomings The Inefficiency of Conventional Water Heaters Traditional water heaters, especially those that run on gas or use electric resistance, are notorious for their high energy consumption. Gas water heaters combust fossil fuels, which not only depletes natural resources but also releases harmful emissions into the atmosphere. Electric-resistance heaters, on the other hand, use a substantial amount of electricity to generate heat, often sourced from non-renewable energy, contributing to a larger carbon footprint. Consumers face long-term financial burdens due to the inefficiency of these models, which translates into higher utility bills and potential costs associated with environmental damage. Emerging Preference for HPWHs As the market becomes more educated on the benefits of energy efficiency, HPWHs are swiftly gaining favor. These sophisticated devices operate on a simple yet effective principle: absorbing heat from the surrounding air and transferring it to water in an enclosed tank. This process requires significantly less energy—typically one-third to one-quarter—that of their traditional counterparts. HPWHs offer substantial utility bill savings while simultaneously reducing households’ carbon emissions, effectively addressing the pressing issue of climate change. New Efficiency Standards Accelerating HPWH Adoption The Impact of Federal Efficiency Standards In a bold move to foster a more energy-efficient future, the Biden administration has introduced stringent new standards for electric water heaters. By setting the minimum uniform energy factor (UEF) at 2.3 for models with tanks larger than 35 gallons, outdated electric-resistance heaters are unlikely to keep pace. This regulatory push will catalyze the phasing out of less-efficient technologies, favoring the wide-scale adoption of HPWHs. Manufacturers are now incentivized to invest in and promote these eco-friendly alternatives to meet the market demands that the new standards will undoubtedly create. Anticipated Market and Environmental Benefits The implications of these new federal efficiency standards are profound. They are expected to reduce carbon emissions by 332 million metric tons and save a substantial amount of energy over the next 30 years. For consumers, the switch to HPWHs is projected to translate into average savings of $1,800 over the lifespan of their water heater. As such, the market share for HPWHs is forecasted to skyrocket—from a modest 3 percent currently to an anticipated 61 percent once the standards are fully enacted. This surge will not only benefit the environment but will also contribute significantly to household budget savings. State and Local Initiatives Complementing Federal Efforts Case Studies of Local Impact The state of California has set a clear precedent with its ambitious goal to ban new gas water heaters by 2030, underlining its commitment to the environment and public health. This decisive legislative action is a microcosm of the broader shifts taking place across the country. Similarly, other states and private sector actors, like Clayton, a leading manufacturer of prefabricated homes, are leading the charge in normalizing the shift to sustainable water heating technologies. These localized advances demonstrate the potential of policy and market forces working in tandem to drive innovation and consumer adoption. The Interplay of Policy and Market Forces State regulations and local air quality initiatives play a pivotal role in buttressing the rapidly growing sales of HPWHs. For instance, several states, not just California, have begun to offer incentives for residents transitioning to energy-efficient appliances. These local policies not only encourage consumers to choose greener options but also create a supportive ecosystem that complements federal efficiency mandates, paving the way for HPWHs to become the new norm in water heating technology. Incentivizing the Shift to HPWHs Tax Credits and Financial Incentives Recognizing that the higher initial cost of HPWHs could prove to be a deterrent, the Inflation Reduction Act of 2022 provides substantial tax credits that help alleviate the financial burden. These incentives have been crafted to make energy-efficient technologies like HPWHs more attractive to the average American consumer. Beyond federal initiatives, states and utilities are offering rebates that further offset the costs, advancing the competitiveness of these appliances in the market. Making HPWHs Accessible to Broader Populations Ensuring that energy-efficient technologies are accessible to a wide range of consumers, especially those in lower-income brackets, is critical for widespread adoption. The array of financial incentives is reducing the upfront cost of HPWHs, making them a viable option for more households across the socioeconomic spectrum. These measures are instrumental in fostering a market where HPWHs are not merely an eco-friendly alternative but increasingly the default choice for new and replacement water heaters.
<urn:uuid:61c517ed-6ad6-43f3-b4d6-4c67c08f8825>
CC-MAIN-2024-38
https://energycurated.com/infrastructure-and-technology/the-rise-of-heat-pump-water-heaters-in-the-us-market/
2024-09-13T01:52:07Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00320.warc.gz
en
0.926251
1,062
2.53125
3
A recent research states that e-reading systems with support for open EPUB format can be exposed to severe security risks. As many users may remember, this format relies on XHTML and CSS for building e-books, so web browsing engines often use it for content rendering. According to the report, it is the inherent characteristics of this format that represent a severe risk linked to the use of e-books in web browsers. During the analysis, the researchers used a semi-automated testing framework available on GitHub and found that 16 of the 97 systems analyzed allowed the leakage of file system information from an EPUB; in eight of these cases, the contents of the file were even allowed to be extracted. A threat actor could fully access the affected system by exploiting some specific features in the implementation of the read systems. While the severity of an attack depends on the platform used and the type of information stored, millions of users could be affected by these attacks. The researchers also conducted a manual evaluation of the most popular EPUB reading applications on Amazon Kindle, Apple Books, and the EPUBReader browser extension, which also led to the finding of multiple security weaknesses: “Most of the security measures on these readers can be evaded by abusing an input validation flaw.” , the experts add. Regular users of e-books may be the most surprised, but experts also found multiple flaws in Apple Books, which comes pre-installed on macOS, as well as security issues in the Windows version of Adobe Digital Editios. These issues were presented to the developers of the aforementioned tools and are expected to be corrected soon. To learn more about information security risks, malware variants, vulnerabilities and information technologies, feel free to access the International Institute of Cyber Security (IICS) websites. He is a cyber security and malware researcher. He studied Computer Science and started working as a cyber security analyst in 2006. He is actively working as an cyber security investigator. He also worked for different security companies. His everyday job includes researching about new cyber security incidents. Also he has deep level of knowledge in enterprise security implementation.
<urn:uuid:68b341af-8a70-4406-89e9-062dc35321f9>
CC-MAIN-2024-38
https://www.exploitone.com/cyber-security/experts-found-multiple-vulnerabilities-in-e-book-reading-systems/
2024-09-13T02:28:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00320.warc.gz
en
0.947439
433
2.703125
3
For the better part of four decades, Apple has bragged that not only are their devices more secure than PCs, hackers don’t bother building threats specifically for their operating systems because their security is so superior. For this reason, Apple has routinely refused advances from law enforcement to share workarounds so that police can get into phones. Apple’s rationale for this constant refusal is that it would undermine their ability to keep the most secure personal computing devices, secure. Federal law enforcement officials went ahead and developed their own workaround and the findings may surprise many Apple aficionados. Let’s take a look: After years of trying to go through Apple to gain access, they finally worked it out in 2020. In 2021, cryptographers published Data Security on Mobile Devices: Current State of the Art, Open Problems, and Proposed Solutions, which is a position paper that looked to answer three questions: - What security measures are currently in place to help deter unauthorized access to user data? - How do modern devices allow unauthorized access? - How can mobile security be improved to prevent unauthorized access? Researchers analyzed both the newest Android and iOS platforms and found that neither of them had security preparations that functioned any better than the other. Any person with the right equipment, and the inclination, can in fact, access the OS on either device. This may come as a shock to those people who have been lauding Apple’s devices to be impenetrable. Before you trash your iPhone, the researchers did “find a powerful and compelling set of security and privacy controls, backed and empowered by strong encryption” in iOS, but the tools presented were not used frequently enough to ensure security is maintained. Android’s issues were exacerbated, in comparison to Apple’s, due to the vast amount of manufacturers that make Android-run products. They found that many devices lacked communications between Google, resulting in slowly implemented updates and inconsistencies in some devices’ security and privacy controls. These are just the hardware and software vulnerabilities. In the rest of the report, the researchers detailed the specific vulnerabilities for each platform. One of iPhone’s best features is that it allows users to securely store data to iCloud. According to the researchers of this report, that isn’t all the data Apple takes possession of. When initiated, iCloud takes control of a lot of other data that is sent to Apple, where it is accessible by all different types of entities, hackers and law enforcement included. This problem is exacerbated as the defenses put forth by Apple are less effective than initially thought. Analysis of this relationship led researchers to suppose that a tool that has been around since 2018 allows attackers to bypass integrated protections to guess user passcodes. On the other hand, researchers found Android had some serious issues with its local data protection. An example of this can be found in Android’s lack of an equivalent to Apple’s Complete Protection encryption, which leaves Android more open to breach. This is why the FBI can effectively access data from either platform without help from developers. So What’s the End Result? Ultimately, both mobile OSs are much more open to data breaches than either manufacturer is willing to admit. It’s never a good practice to assume your data is safe; especially with the default data protection developers have in place. It just goes to show that there is no such thing as impenetrable security, and it is on the users (or the organization) to actively accept these results and do what they need to do to secure their data more effectively. To do this, you will need to manage these devices with a mobile device management platform and have your employees sign onto a Bring Your Own Device policy. This way your organization is covered in ways that individual devices and mobile platforms simply can’t. If you would like more information about Bring Your Own Device, mobile device management, or any other platform that helps keep your organization’s data secure, give the IT experts at COMPANYNAME a call at PHONENUMBER.
<urn:uuid:523060a8-7dfb-4a4e-ad1d-125a423b71b1>
CC-MAIN-2024-38
https://netcotech.com/are-apple-computers-really-more-secure/
2024-09-15T14:24:48Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651630.14/warc/CC-MAIN-20240915120545-20240915150545-00120.warc.gz
en
0.954794
836
2.546875
3
Combining Social Engineering & Malware Implementation Techniques Cybercriminals will often use a combination of social engineering methods and malware implementation techniques – in order to maximize the chances of infecting users’ computers: - Social engineering methods – including phishing attacks – help to attract the potential victim’s attention. - Malware implementation techniques – increase the likelihood of the infected object penetrating the victim’s computer. This was one of the first worms that was designed to steal personal data from users’ online accounts. The worm was distributed as an email attachment – and the email contained text that was designed to attract the victim’s attention. In order to launch a worm copy from the attached ZIP archive, the virus writers exploited a vulnerability within the Internet Explorer browser. When the file was opened, the worm created a copy of itself on the victim’s disk – and then launched itself, without any system warnings or the need for any additional action by the user. - Hello A spam email – with the word ‘Hello’ in the subject line – stated ‘Look what they say about you’ and included a link to an infected website. The website contained a script that downloaded LdPinch – a Trojan virus that was designed to steal passwords from the user’s computer, by exploiting a vulnerability in the Internet Explorer browser.
<urn:uuid:24757b88-4bc2-44c4-aa5a-0b595e510d16>
CC-MAIN-2024-38
https://www.kaspersky.com.au/resource-center/threats/malware-manipulation
2024-09-15T14:30:12Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651630.14/warc/CC-MAIN-20240915120545-20240915150545-00120.warc.gz
en
0.935914
279
3.046875
3
Overcool or Overheat… “Where? It’s a big data center.” Data centers have historically been overcooled to ensure uptime, but the trend is shifting towards more efficient cooling architectures. This article explores ASHRAE’s recommended temperature range for data centers and the benefits of optimizing temperature for energy efficiency. Importance of Optimizing Data Center Temperature Traditional Data Centers are over cooled, focussing on uptime. This is because they have limited or poorly designed containment, little understanding of the air flows, air temperatures and air pressures at key points in the heating and cooling cycle. The trend has been towards more efficient cooling architectures. These typically use hot or cold aisle containment. So what temperature should you be aiming for in your data center and where should you be measuring this temperature? Since 2005 ASHRAE (American Society of Heating, Refrigerating and Air-Conditioning Engineers) recommends operating anywhere between 18°C to 27°C (64.4°F – 80.6°F) with a humidity between 40-55%. This gives a starting point, but how do we decide where within this range to operate, and where to place sensors? Overview of ASHRAE Temperature Recommendations for Data Centers The safest option is to aim for 18°C. However, the safest option is not the most energy efficient. Modern servers operate at higher temperatures, such as 23°C – 24°C (73.4° – 75.2°F) server inlet temperature. There may be a slightly reduced lifespan of the components but this will be more than offset by the energy saving. Data centers can save 4-5% in energy costs for every 0.5°C (1°F) increase in server inlet temperatures. Going from 18°C to 24°C (64.4°F to 75.2°F) could save up to 43% of OpEx. This also reduces the carbon footprint of the data center, and the amount of water you need. Increasing the capacity of your existing facility without needing to outlay CapEX for additional cooling equipment is also possible. There is a trade off that comes with this. There is less margin for error. When you run your data center at an average of 18°C, a failure to properly measure server inlet temperatures is not critical. The closer to the upper temperature limits you run, the more you save in energy but the higher your risk of thermal overload. Factors Influencing Data Center Temperature The below factors can have an impact on your server rack inlet temperatures and energy consumption :- - The local climate - The number of racks, other equipment and how much heat they produce - The size and shape of the room - Power dissipated - Cooling system design If any one of these factors changes, your Data Center may be operating at too high or too low a temperature at some location in your Data Center. Benefits of Optimized Data Center Temperature The closer to the upper limit of ASHRAE recommended temperatures you operate, the more important a comprehensive monitoring system becomes. Monitoring the temperature of your data center offers several key benefits: - Minimizes downtime. Proper monitoring of temperatures helps avoid costly downtimes due to thermal overload. - Detailed analysis of hotspots and cold spots. Identify problem areas before they are critical. Locate where you have stranded capacity and overcooling. - Safely operate at the upper limits of ASHRAE recommended temperatures. The Role of Temperature Monitoring in the Data Center There are many options when it comes to monitoring solutions for the data center. From a single temperature sensor in the room, to detailed thermal mapping of the rack inlet and outlet temperatures, with ∆T calculations. Uptime Institute recommends a minimum of 3 temperature sensors on at least every other rack. The sensors are located at the top middle and bottom of the front of the rack. New technologies such as sensorCFD™ from AKCP integrate live sensor data and CFD (computational fluid dynamics) and real time analysis. This is useful in understanding airflows and temperature distribution. Here are some important features to look for when selecting a monitoring system for your data center. - 24/7 remote monitoring: This feature allows you to monitor your data center’s temperature in real time. - Automated alerts. Let you know when temperatures in the data center are outside of recommended ranges. Take action before issues arise. - Data logging capabilities:. Storing your data on a regular basis to spot long-term temperature trends and make adjustments. - Integration with DCIM platforms via SNMP or cloud platforms with MQTT. By using these features, data center professionals can rest assured that their facilities are running at peak efficiency and with maximum safety. Implementing DCIM Monitoring Software To ensure your data center runs at optimal conditions, you’ll need to invest in monitoring software. This will allow you to maintain more accurate readings and adjust the temperature as needed. With the right monitoring software, you can: - Monitor temperature conditions in real-time and receive alerts when specific thresholds are crossed - Accurately record and capture temperature fluctuations for future reference or analysis - Identify changes in temperature so you can take action before a critical situation arises - Optimize cooling systems performance by preventing system overloading - Save energy costs by avoiding unnecessary cooling system operation ASHRAE recommends implementing a comprehensive, integrated monitoring system. This can automate control of cooling systems and help maximize efficiency. It should generate detailed reports of past events and provide near real-time data on current conditions. With this information operators can make informed decisions. A top healthcare provider in the USA recently implemented thermal map sensors from AKCP on every rack in their data center. The sensors are monitored through Sunbird DCIM via SNMP. Sunbird provides data analytics, alarming and visualization tools to show hot and cold spots in the data center. This has allowed the data center to get an insight as to where they are wasting energy through overcooling, release stranded capacity by increasing server loads and safely manage hotspots when operating the data center at elevated temperatures. Example of thermal maps displayed in Sunbird DCIM Thermal map sensors consist of 3x temperature sensors at the front, and 3 at the rear of every rack, giving you measurements of rack inlet, outlet and ∆T. DCIM software from Sunbird uses this sensor data to create heatmaps of the data center. The healthcare provider discovered that they were overcooling servers throughout their data center, and could safely increase their data center operating temperature without creating hot spots. They successfully reduced their energy use by 18% while maintaining safe operating conditions for their servers. Based on the results achieved, they are preparing to roll out the same setup at their second data center facility.
<urn:uuid:c32efc4e-b411-4883-bc46-348e546ba51c>
CC-MAIN-2024-38
https://www.akcp.com/blog/overcool-or-overheat-data-centers-optimizing-temperature-for-efficiency/
2024-09-09T12:52:36Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651098.19/warc/CC-MAIN-20240909103148-20240909133148-00720.warc.gz
en
0.91174
1,429
2.6875
3
As our cities become increasingly connected and transform into Smart Cities, there is an opportunity to streamline emergency communications. Cities and municipalities can leverage a variety of advanced technologies and incorporate them into their own emergency communication plans. Emergency management decision makers tasked with improving city-wide emergency and disaster plans now have access to technology that can assure connectivity in the harshest weather or environmental conditions; increase visibility into dangerous environments; and, optimize response times. Wireless Short-haul for the Win Wireless short-haul solutions can create an industrial-strength Wi-Fi connection that was built to withstand earth’s most challenging conditions. These Sensor-2-Server (S2S) types of technologies are used for a variety of municipal and government use cases, but they are particularly suited for outdoor communication needs. While they are often used for day-to-day use, such as traffic management, they are a viable option for providing secure, reliable connectivity as part of any city or local government’s emergency communication plan. VVDS for Emergencies With an industrially hardened, high-speed wireless short-haul solution in place, cities can experience the benefits of Voice, Video, Data and Sensor (VVDS) information, even when cell towers are overloaded. In a world where we increasingly rely on connectivity, it is essential to keep government and municipalities online during the worst-case scenario. Industrial-grade Wi-Fi that is tested and proven in the most extreme weather conditions is designed with that in mind – keeping local government officials and first responders online. As a result, rescue efforts stay motion. With a VVDS-enabled technology in place, first responders achieve additional visibility into conditions. This real-time view allows for fast action that minimizes collateral damage. It also protects first responders, giving them an advantage in dangerous situations and offering a real-time view of environment they are heading into. Secure, Reliable Solutions Industrial wireless short-haul networks also offer the benefits of being highly secure. There are solutions with encryption capabilities that prevent data hijacking. As more cities become Smart Cities, decision makers will need to make Smarter emergency communication plans that align with the new technology landscape. There are S2S solutions on the market today that are designed for unrelenting performance in the outdoors. These solutions enable better response times, secure data transmission, increased visibility and higher-level risk assessment. When emergencies strike, every moment counts. Having a reliable connection can make the difference in saving lives. Is your city leveraging wireless short-haul solutions for emergency preparedness?
<urn:uuid:ab1ef62b-a350-4aeb-87ea-41ae8d79ceb1>
CC-MAIN-2024-38
https://www.freewave.com/whats-emergency-communications-plan/
2024-09-09T13:01:48Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651098.19/warc/CC-MAIN-20240909103148-20240909133148-00720.warc.gz
en
0.937204
527
2.90625
3
Another area of security and data privacy is law enforcement. Unsurprisingly, law enforcement and other national security agencies would handle private information, and such rules and regulations around protecting said information are of paramount concern. Here, we’ll discuss the FBI’s Criminal Justice Information Services division and its compliance requirements. What Is the Criminal Justice Information Services (CJIS)? Established in 1992, CJIS is the FBI’s largest division. It is tasked . CJIS supplies data security guidelines to law enforcement agencies, procures tests, and develops cutting-edge digital tools to help them in that mission. According to the “Criminal Justice Information Services (JIS) Security Policy,” the core document of CJIS compliance, the entire premise of CJIS is to “provide appropriate controls to protect the full lifecycle of CJI, whether at rest or in transit.” It’s essential to understand that CJIS serves several functions: - Centralized Criminal Justice Database: CJIS manages a comprehensive, centralized database with criminal justice information. Law enforcement agencies across the United States use this database to share and access critical information. - Overseeing National Crime Information Center (NCIC): CJIS manages the NCIC, one of the primary databases containing records of wanted persons, stolen property, missing persons, and other criminal justice information. - Administering National Instant Criminal Background Check System (NICS): CJIS administers NICS, which performs background checks on individuals purchasing firearms from licensed dealers. - Supporting the Uniform Crime Reporting (UCR): The UCR program collects and compiles crime statistics from law enforcement agencies nationwide, providing valuable data on crime trends and patterns. - Supporting Integrated Automated Fingerprint Identification System (IAFIS): This system allows for the electronic processing and storage of fingerprint records, enabling quick and accurate identification of individuals. - Defining Best Security Policies: CJIS establishes and enforces security policies and standards to protect the sensitive information in its databases. These policies ensure that only authorized personnel have access to the information and that it is used appropriately. - Providing Training and Support: CJIS provides training, technical assistance, and support to law enforcement agencies to help them effectively use the systems and resources available. Much like any other framework, that is a typical mission for security protocols in any industry or public service sector. However, as this document notes, local and state authorities increasingly rely on FBI information databases to locate or track criminals for the public good. That being said, it’s critical that controls and practices are in place to protect this information, no matter the person or the crime. Accordingly, CJIS is not a required standard that these local or state authorities adopt but rather a required minimum. They may adopt measures that extend CJIS standards or a standalone security system for their locality–so long as it satisfies CJIS requirements at a minimum. Changes to CJIS Requirements in 2024 Over the years, new requirements have been added to CJIS access policies. Some prominent changes were introduced in December 2023 as part of version 5.9.4. These changes include: - The addition of six new Policy Areas: Systems and Services Acquisition, System and Information Integrity, Maintenance, Planning, Contingency Planning, and Risk Assessment. - The new requirement is that any agency accessing CJI must implement Multi-Factor Authentication by October 1, 2024. What Are the CJIS Policy Areas? CJIS compliance is built around 19 policy areas that structure the practices expected of law enforcement. These policy areas aren’t built on specific technology pipelines. Rather, much like other systems like SOC 2 or HIPAA, its goal is to provide a technology-agnostic system that can set a minimum standard that individual agencies can meet as they can. The 19 policy areas in CJIS are: Policy Area 1: Information Exchange Agreements Information shared through communication must be protected. Before the exchange, agencies shall specify security measures through mutual agreements covering personnel, encryption, access, etc. All information will be protected from unauthorized disclosure with proper handling requirements. All state and federal agencies interacting with CJIS databases will have written and signed agreements with the FBI confirming their conformity with CJIS statutes. Policy Area 2: Security Awareness Training Agencies must enact security awareness training within six months of their initial compliance assignment and then update those policies once every two years. These security awareness training systems will do so based on established CJIS baselines: - Level 1: Covers topics such as training around expected behaviors handling CJI, knowledge or penalties around non-compliance, actions around incident response, and security around physical spaces. - Level 2: On top of Level 1 topics, Level 2 will cover media protection, protection and destruction of physical records, proper marking and handling of CJI, prevention of social engineering, and more. - Level 3: Includes Levels 1 and 2, plus knowledge of roles within a system, proper password usage and management, antivirus and malware protection, secure web usage, proper email usage, securing handheld devices, using encryption, using personal equipment, and more. - Level 4: On top of Levels 1, 2, and 3, includes protection against advanced threats, access control measures, network protection, data backup and storage, and others. Policy Area 3: Incident Response When disaster or security threats strike, this policy area requires agencies to have plans to respond. This includes reporting security events, managing incident handling, investigating and mitigating issues related to the incident, and training around incident response. Policy Area 4: Auditing and Accountability It’s critical that agencies can demonstrate compliance from the organization’s perspective and that of its employees. This area calls for IT auditing systems to track system and user events in IT infrastructure. This includes immutable records with time stamps and backup controls to store documents for at least one year. Policy Area 5: Access Control All IT systems must have controls to control authorized access to system resources. This area includes strict role-based access control, account management, access enforcement, and the enactment of least privilege access. Policy Area 6: Identification and Authentication Simply put, how the system securely manages user identities, authenticates against those user identities, and secures identity information against hacks or theft. This area can include minimum password standards, use of PINs, multifactor authentication (MFA), or one-time passwords (OTPs). Policy Area 7: Configuration Management An agency must have plans and procedures to manage system updates, upgrades, or component replacements. This area includes isolating components to minimum functionality, managing network hardware topologies, and proper security system update plans. Policy Area 8: Media Protection All storage media, no matter the type, must have specific physical and digital security measures to protect that data. This includes encryption, hardware security, and physical media (paperwork, images). This area also includes the sanitation and disposal of hard drives containing CJI, including demagnetization and overwriting. Policy Area 9: Physical Protection In addition to protecting physical media, agencies must protect locations where CJI is handled and stored. This includes perimeters around offices, locks and cameras around storage areas and data servers, logging of any entrance or exit of the premises, and other controls around private access points. Additionally, any individual with “unescorted access” either physically or digitally must have some minimum level of privacy training: - Basic Training: A ground-level overview of CJIA security requirements. - Awareness Training: Specific for people with physical access to information, like on-site clerks and secretaries. - Additional Awareness Training: Design for those who can alter information, like dispatchers and officers. - Advanced Awareness Training: This is for people handling critical infrastructure within the CJI system. Policy Area 10: System and Communication Protection and Information Integrity In short, data protection is stored and transmitted. Controls here include encryption (for data both at rest and in transit), firewalls, access controls around network access points, and other network security measures. These controls also apply to cloud computing, VoIP, and other forms of data transmission. Policy Area 11: Formal Audits All agencies must perform formal audits on their infrastructure and organization to ensure compliance. This includes any criminal justice agency (CJA) or noncriminal justice agency (NCJIS) with access to state or federal systems containing CJI. Policy Area 12: Personnel Security Agencies must identify any user accessing or working on their system, including personnel screening procedures, background checks, etc. Additionally, the agency must include security policies around transferring and terminating employees to control or restrict system access. Policy Area 13: Mobile Devices Agencies using mobile devices must use secured technologies, including 802.11 wireless protocols, secured Wi-Fi access points, and mobile device management for official purposes. Policy Area 14: System and Services Acquisition Organizations must have processes to protect the system’s integrity, including automatic software and firmware patch and update management. Policy Area 15: System and Information Integrity Agencies should continuously monitor systems to note vulnerabilities or attacks, software changes, or contained data changes. Policy Area 16: Maintenance Agencies accessing CJI or storing associated data must schedule, document, and record maintenance or equipment replacement. These maintenance events must have approval regardless of where they occur. Policy Area 17: Planning An agency should have plans to address emergency and non-emergency situations, including attacks, vulnerabilities, or updates. These plans must adhere to all CJIS privacy requirements. Policy Area 18: Contingency Planning Have a well-documented and tested contingency plan that spans the organization and addresses all defined IT missions, operational functions, or other requirements. Policy Area 18: Risk Assessment Any system containing CJI or related information should have a clearly defined risk management profile that identifies potential threats, vulnerabilities, and the system’s value or sensitivity. Manage Your CJIS Compliance with Lazarus Alliance CJIS compliance, like any other, requires regular vigilance and continuous management. You can find such management, expert support, and technical infrastructure with Lazarus Alliance. To learn more, contact us. - NIST 800-53 - FARS NIST 800-171 - SOC 1 & SOC 2 - HIPAA, HITECH, & Meaningful Use - PCI DSS RoC & SAQ - IRS 1075 & 4812 - ISO 27001, ISO 27002, ISO 27005, ISO 27017, ISO 27018, ISO 27701, ISO 22301, ISO 17020, ISO 17021, ISO 17025, ISO 17065, ISO 9001, & ISO 90003 - NIAP Common Criteria – Lazarus Alliance Laboratories - And dozens more!
<urn:uuid:5b24f28e-af72-44e4-b786-6d712a958f5f>
CC-MAIN-2024-38
https://lazarusalliance.com/what-is-cjis-compliance-in-2024/
2024-09-10T17:04:18Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651303.70/warc/CC-MAIN-20240910161250-20240910191250-00620.warc.gz
en
0.897772
2,232
2.515625
3
An industrial revolution is upon us. In fact, it’s the fourth time an industrial revolution has occurred. This time, it’s driving digital transformation trends and helping companies become Autonomous Digital Enterprises. In this article, we’ll: - Define the fourth industrial revolution - See how we got this far - Look at the 4 key design principles - Share trends so you can stay ahead of the competition What is the Fourth Industrial Revolution? The Fourth Industrial Revolution is the automation of manufacturing and industrial practices that would otherwise be manual tasks. Modern intelligent technology, machine communication in large machine-to-machine networks, and the advent of the Internet of Things (IoT) combine to create opportunities for businesses to capitalize on: - Smarter automation - More detailed communication and advanced self-monitoring - The creation of intelligent machines that manage, monitor and maintain systems - Solving problems without the need for developers A brief history of the Fourth Industrial Revolution In 2015 Klaus Schwab, executive chairman of the World Economic Forum, coined the term “fourth industrial revolution”. In 2016, the Forum began to study the fourth industrial revolution at a center in San Francisco. Schwab sums up this revolution as: This is the era of autonomous vehicles and other breakthroughs in automation that streamline efficiencies for business through robotics and intelligent automation mixed with quantum computing and biotechnology. As the name suggests, there have been three preceding industrial revolutions: - The first industrial revolution transitioned from hand-building techniques to steam power occurred roughly between 1760-1840. - In the late 19th and early 20th centuries, railroads and mail service emerged. - At the end of the 20th century, a third industrial revolution ushered in an era of digital technology. This third revolution, the digital era, happened as a decades-long response to World War II and the resulting industrialization slowdown. In time, computers and hand-held technology came to dominate the marketplace and business ecosystem, particularly around the need for more efficient ways of communication. What is Industry 4.0? You might also hear about Industry 4.0. The term Industry 4.0 refers to the German government’s complex, high-level strategy to promote digitization and computerization in traditionally human industries. For instance, in manufacturing, human operator error can cause accidents or damage thousands of dollars of equipment. Industry 4.0 has four design principles: - Information transparency - Decentralized decision making - Technical assistance Technologies of the 4th Industrial Revolution These popular concepts and technologies are what we mean when we talk about this 4th industrial revolution. You’ll see—with these technologies vast potential—why this revolution is playfully dubbed “the imagination age”. Virtual reality (VR) This emerging technology allows corporations to create immersive experiences using a digital VR headset. This can be anything, for instance: - Virtual safety training for employees in the transportation industry. - Virtual vacation services in the travel industry that help people take a staycation that feels more luxurious than just staying home. These immersive digital experiences are designed to simulate real-world experiences, sometimes with augmented perceptions that allow the user to react, respond, and learn from the simulation. Smart computing and automation Computing software today is designed for enterprise software needs. It can collect, organize, and assemble large chunks of data, provide visualizations, and even assist in practices like: - Virtual assistant functions 3D print makes physical tooling more accessible. This physical tooling—“printers”—can help quickly develop and deploy creative ideas into the marketplace: - Large-scale 3D printing has the potential to create medical devices and healthcare tools quicker and more affordably than traditional manufacturing. - At-home 3D printing has also come far, with hobbyists experimenting with print-at-home car parts, shoes, and more! With the development of 3D printing, fields from medicine and science to engineering and robotics can be decentralized. This means tools can be created as needed, more affordably, and closer to the source than even before. This technology is relatively new within this 4th revolution. Today’s scientists and engineers look for ways to capitalize on decreasing renewable energy costs in order to solve our global problem of limited resources. This emerging field uses biomolecular processes as a catalyst for developing new technology that can be used in industries like: - Areas where clean, efficient science is valuable The Internet of Things The Internet of Things, or IoT, is a term that describes the cloud as the ether in which a number of physical devices (“things”) can: - Provide feedback to one another - Create a network of tangible resources that can provide new opportunities for companies to connect internally and with customers. For example, when your phone can access a user portal with a company dashboard, you are connecting to IoT. Likewise, when a fleet of trucks use a centralized, proprietary dispatch and GPS system to move cargo, they are also part of the IoT network. Of course, the IoT has its own repercussions, something known as the Internet of Behavior (IoB). Technological advances in robotics continue to make it more usable and accessible. Eventually, perhaps, a robo-assistant will be a staple in any home or workplace. Today, however, that’s just a dream. Right now, robotics is emerging as a useful tool when machine learning, automation, and computation can be applied to a single machine that accomplishes a larger goal. Artificial intelligence (AI) Based on machine learning, artificial intelligence is the term for a computer using synthesized information to create something new or provide insight into a trend or challenge—without being prompted to do so. Blockchain is the method behind cryptocurrency, like Bitcoin, that allows tokens to be passed securely from one user to the next, to be accumulated and traded as part of a large and infinite ledger tracking each transaction as it occurs. Change is here—again Whether referred to as Industry 4.0 or the fourth industrial revolution, businesses are navigating a challenging time of organizational change. They are, once again, adapting to the technological disruption that will change their infrastructure and how they operate. - BMC Business of IT Blog - The Complete Digital Transformation Guide - Organizational Change Management (OCM): A Template for Reorganizing IT Around Your Vision - 7 Ways To Put Automation Everywhere Into Practice - Workload Automation (WLA) vs Robotic Process Automation (RPA): What’s The Difference? - 5G for Companies: Hype, Reality & Potential
<urn:uuid:a242c865-3dab-49be-82f7-d634d690c891>
CC-MAIN-2024-38
https://www.bmc.com/blogs/fourth-industrial-revolution/
2024-09-10T17:22:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651303.70/warc/CC-MAIN-20240910161250-20240910191250-00620.warc.gz
en
0.911616
1,392
3.453125
3
Experiments are the individual "projects" within a Use Case. They allow you to vary data, targets, and modeling settings to find the optimal models to solve your business problem. Within each experiment, you have access to its Leaderboard and model insights, as well as information that summarizes the data and experiment setup. There are two types of AI experiments available in Workbench: - Predictive modeling, described on these pages, makes row-by-row predictions based on your data. - Time-aware modeling, described here, models using time-relevant data to make row-by-row predictions, time series forecasts, or current value predictions "nowcasts". For each, you can build models using either supervised or unsupervised learning. - Supervised learning uses the other features of your dataset to make predictions. - Unsupervised learning uses unlabeled data to surface insights about patterns in your data, answering questions like "Are there anomalies in my data?" and "Are there natural clusters?" The following sections help to understand building predictive machine learning experiments in Workbench: Topic | Describes | Create experiments | | Supervised experiment setup | Specify a target to build models using the other features of your dataset to make predictions. | Unsupervised experiment setup | Use unsupervised learning to build models that surface insights about patterns in your data. | Advanced experiment setup | Use the Advanced settings tab to fine-tune experiment setup. | Manage models | | Manage the Leaderboard | Navigate and filter the Leaderboard; create feature lists. | Compare models | Compare up to three models of the same type from any number of experiments within a single Use Case. | Add/retrain models | Retrain existing models and add models from the blueprint repository. | Edit blueprints | Build custom blueprints using built-in tasks and custom Python/R code. | Explore model insights | | Evaluate models | View model insights to help evaluate models. | Reference | | SHAP reference | See details of SHapley Additive exPlanations, the coalitional game theory framework. | Troubleshooting the Worker Queue | Describes how DataRobot uses modeling workers and how to troubleshoot problems. | An experiment can only be a part of a single Use Case. The reason for this is because a Use Case is intended to represent a specific business problem and experiments within the Use Case are typically directed at solving that problem. If an experiment is relevant for more than one Use Case, consider consolidating the two Use Cases.
<urn:uuid:33e6dd61-1264-4cdf-ab94-ca2fefa8477f>
CC-MAIN-2024-38
https://docs.datarobot.com/en/docs/workbench/wb-experiment/wb-ml-experiment/index.html
2024-09-16T22:42:40Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651714.51/warc/CC-MAIN-20240916212424-20240917002424-00120.warc.gz
en
0.88753
539
2.859375
3
The FAIR data principles were first expounded in 2016 by a consortium of leading scientists and organizations in Scientific Data magazine. Their goal was to ensure that scientific data sets could be found and used by machines, with minimal human intervention. FAIR stands for Findable, Accessible, Interoperable and Reusable. Six years later, how can these principles help to create collaborative, data-sharing ecosystems up and down supply chains, in complex operational landscapes such as transport and allow servitization models in multi-party ownership scenarios such as utilities? Over the same period, we have seen the evolution of the digital twin. While digital twins mean different things to different people, the widely accepted interpretation is a digital data twin, where the twin is a virtual representation of something in the real world. Single pane of glass One of the first uses of digital twins was to make a ‘single pane of glass’ abstraction of an asset – which could be a thing, a person or a place. Data for assets is inevitably stored in incompatible ways in disparate systems for multiple, unrelated purposes. So, by focusing access to all this data through the digital twin, you create a single point where authorized people, applications - and even other twins - can go to find out about the asset, its current state and even subscribe to its updates. A big advantage of this model is that the owner of the asset stays in control of what goes in the pane of glass and who’s allowed to look at it. This approach works for purpose-built applications that have knowledge about a twin’s data built into their logic. For example, they may know that the dimensions of the asset are recorded in the twin as centimetres and the weight in kilograms. But this means that the data must be programmed by people; it is then fixed to that type of twin, and it’s difficult for the parties with whom you want to share the data to understand it. This only partially addresses the problem of multiple sources and has moved the interpretation of the data into application logic. Imagine the application is data-centric, in that it reacts to data and metadata – date about data. In our example, the weight data would ‘say’ it was in kilograms and the application would use this to interpret the data and respond accordingly. Now imagine how the ‘browser’ model that we use every day would work with digital twin data. You could search for twins and allow the application to react to metadata and display their data in the best way. You could show data from more than one twin and compare them. You could write code snippets to do this automatically. You could even take data from multiple twins, run synthesising algorithms on it and publish the results as more twins. But this is only possible if digital twins are made FAIR. Why twins must be FAIR ‘Search’ implies that the twins are made findable by their creator. ‘Choose’ implies that the twins are accessible - if I’m authorized. ‘React' and 'Compare’ imply that the data received is understandable and hence interoperable. Code snippets, synthesis and algorithms all imply that the data can be used for reasons other than its original purpose - hence reusable. In the internet world, HTML is used to ‘mark-up’ data to tell browsers how to render it. For example, tags like <table>, <tr>, <td>, etc tell the browser that this is tabular data. The FAIR principles don’t stipulate what method is to be used to specify metadata, but they do demand that: “(Meta)data use a formal, accessible, shared and broadly applicable language for knowledge representation”. There aren’t too many of these languages. RDF (Rich Data Format) and the Semantic Web technologies are the de-facto standards. But the digital twin browser application we imagined is not the end of the story. The originators of the FAIR principles had autonomous machine interoperability as one of their goals. If we apply this thinking to digital twins in an ecosystem, twins’ agents – the behavioural part of a twin – could search for twins near or related to them, interact with their data and then maybe drop the connection when they’ve moved on. For example, the twin of a train could search for nearby twins of pollen count data as it was moving, because pollen clogs the filters when the engine is running. Twins of engines on the train would know when they were running and update their metadata to reflect that they have been affected. Service engineers can look at the twins of the engines in the train to see when the filters need to be changed. FAIR data principles build on the foundations of each other. You can’t reuse if you can’t interoperate; you can’t interoperate if the data is inaccessible; you can’t try to access data if you can’t find it. Saying that the FAIR principles are for scientific datasets, is like saying Amazon is for books. Five years ago, the originators of the FAIR principles must have had a good idea that their principles applied to many things - including digital twin ecosystems. The best principles work like that. They give you yardsticks and guidance but don’t limit where you apply them.
<urn:uuid:7ce88b4a-304d-43ac-a254-42458d31f4e1>
CC-MAIN-2024-38
https://direct.datacenterdynamics.com/en/opinions/playing-fair/
2024-09-20T13:34:23Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652278.82/warc/CC-MAIN-20240920122604-20240920152604-00720.warc.gz
en
0.951691
1,107
3.328125
3
Urbanization, population growth, energy security concerns, and economic developmental objectives are the main reasons for depletion of conventional energy sources. Consequently, there is an increase in usage of non-conventional sources of energy that are renewable like solar, wind, geothermal, and energy from biomass. As a result of technological innovation in renewable energy, the cost of renewable energy has declined significantly in recent years. For example, the cost of solar photovoltaic (PV) modules is estimated to have dropped by 70% between 2008 and 2014, and this has resulted in a substantial increase in adoption of solar power. Moreover, there are growing interests from China as it is already the largest producer of solar equipment globally, and it is poised to be the leading country in solar power adoption in the future. A major component of the solar energy market is light-driven silicon based PV technology that is unable to generate power during night or on cloudy days and therefore, yields poor efficiency rates. Nonetheless, technology innovation is driving the efficiency improvement by using infrared-based PV that allows harnessing the infrared (IR) part of the solar spectrum. Such PV cells can work much efficiently even in the presence of clouds or at night. Introduction: IR based solar cells The IR- based solar cell is made of carbon nanotubes (CNT) that effectively captures sunlight in the near-infrared region. CNT-based PV cells are transparent to visible light and can be overlaid on conventional solar cells, opening up the possibility of harnessing the entire range of solar spectrum. Importantly, IR based PV has an efficiency rate of more than 50 percent as compared to conventional efficiency rate of only 11 to 15 percent. The increase in the efficiency makes IR based PV ideal for large scale deployment which will result in consistent energy generation and 24-hour operations. In addition to efficiency advantage, IR based PV provides similar benefits or advantage of conventional solar PV; generation of in-situ baseload and income generation via feed-in-tariff. Global Focus on Solar The aftermath of Fukushima mishap in 2011 saw Japan steering sharply toward renewable technologies; the country aims at generating 28 GW of solar power by 2020. Moreover, Japan, Taiwan, and China are expected to contribute more than 70 percent of the global solar PV production. These three countries are the high potential production centers with mature PV industry and CNT technology. This gives other global participants such as the United States and Germany a run for their money. Attractive demand centers for IR based solar cells are likely to come from giant economies such as India and China. China, which produces more than half the world’s supply of solar panels, is focusing on its internal needs for clean energy. In fast urbanizing country such as India where power disruption is common, it is more feasible to install local solar energy plants as greenfield projects due to its large availability of land. Moreover, price of electricity can make or break the overall solar industry. For example, the price of electricity from Japan’s power grid is higher than price of electricity from solar cells, making the country a ready market for use of solar cells. In countries with generally low electricity tariff, it is often usual that the adoption of solar energy is an uphill task with additional institutional and regulatory challenges. Opportunities and Challenges IR based solar panels are a step toward efficient utilization of available solar energy. Besides 24-hour function and efficiency, IR based PV cells are smaller in size when compared to conventional solar panel of similar capacity. Additionally, production cost of IR-based solar panels from carbon nanotubes is considerably lesser than visible-light based solar panels from silicon. The main challenge for market participants of IR based solar panels is the insufficient infrastructure for its implementation, as currently light based solar market is at the pinnacle of its growth. However, as the overall solar market continues to grow, the readiness for adoption of IR based solar panels is expected to be better in the next five years. Case Study: Solaxess- White Solar technology Solaxess, a Swiss-headquartered company, specializes in PV products, or building integrated PV (BIPV) to be specific. The technology used by it converts IR part of solar light into electricity while rendering the solar panel with different colors. For this it uses selective scattering filter to transmit the IR light and scatter the whole visible spectrum. This scattering of white light is necessary to give the panel a white or colored appearance, which is achieved by embedding a Nano-filter on a micro structured surface. Solaxess demonstrates best practices in innovating IR based PV in terms of efficiency, and to be uniquely different from other available solar panels in the market. Overall, the renewable solar energy market is expected to experience continuous shift in innovation and advancements. Japan and other emerging economies such as in India and China will proactively drive a sustained increase in the adoption of solar energy. With IR based solar panels in market, the production cost of solar power is expected to be drastically reduced. In fact, by combining it with conventional solar cells, end users may be able to harness most of the solar energy efficiently and effectively. As such several action items are relevant to the stakeholders in the market value chain: - Evolution of business model and solution portfolio - Ease of policies to encourage mass adoption of solar technologies - Awareness creation on cost and benefits
<urn:uuid:5d08d951-c3ff-4f4b-8a15-f2f421050631>
CC-MAIN-2024-38
https://dev.frost.com/growth-opportunity-news/off-the-grid-power-generation-focus-on-infrared-based-solar/
2024-09-09T14:36:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651103.13/warc/CC-MAIN-20240909134831-20240909164831-00820.warc.gz
en
0.944562
1,098
3.75
4
Information technology — Artificial intelligence — Management system About ISO 42001 ISO/IEC 42001 is an international standard that provides a comprehensive framework for establishing, implementing, maintaining, and continually improving an Artificial Intelligence Management System (AIMS) within organizations. This standard addresses the unique challenges posed by AI systems, including transparency, explainability, and ethical considerations, ensuring their responsible use and development. Benefits of ISO 42001 Ethical AI Development: ISO 42001 sets a global benchmark for ensuring AI systems are developed and deployed with a strong emphasis on ethics, security, and transparency. Enhanced AI Governance: The standard fosters a structured approach to AI management, promoting best practices that enhance reliability, safety, and stakeholder trust. Alignment with Global Standards: ISO 42001 aligns AI technologies with international standards and regulatory requirements, making it easier for organizations to operate across borders. Sustainable Development Goals: The standard supports global initiatives like the United Nations Sustainable Development Goals (SDGs), contributing to positive societal impacts. Importance of ISO 42001 ISO 42001 is critical for organizations that integrate AI into their processes. It ensures that AI systems are not only efficient but also ethically responsible, secure, and transparent. By adopting this standard, organizations can align with global ethical principles and regulatory requirements, fostering innovation while safeguarding against risks.
<urn:uuid:469b5bf5-0c72-4aea-a051-b47539ffd9df>
CC-MAIN-2024-38
https://www.allendevaux.com/iso42001
2024-09-09T14:12:16Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651103.13/warc/CC-MAIN-20240909134831-20240909164831-00820.warc.gz
en
0.88134
280
2.8125
3
In our world where the virtual landscape keeps on evolving at an unprecedented rate, ever-increasing cybersecurity risks are more than ever real. These risks pose a significant threat to your organization, its reputation, and, ultimately, its survival. Our focus for this month of July 2023 is on Cybersecurity Awareness Training, which acts as a crucial defence in protecting your organization, its assets, and your clients. Chapter 3 of The Global Risks Report 2022 revolves around digital dependencies and cyber vulnerabilities across the world. The main highlights correspond to the disturbing findings hereafter: an increase of 435% in ransomware and 95% of cybersecurity issues are tracked down to human error. 95 % de tous les problèmes relatifs à la cybersécurité peuvent être attribués à une erreur humaine1 What is Cybersecurity Awareness Training? Cybersecurity Awareness Training is a corporate-wide initiative to educate employees about the different kinds of cybersecurity threats that impact users’ accounts, devices, information technology systems and networks, and how to identify and avoid them.2 A Cybersecurity Awareness Training helps employees understand what cyber threats look like, how they operate and how to respond when they encounter such cyber threats. Why is Cybersecurity Awareness Training Essential? The digital world brings about a multitude of risks, ranging from security breaches and data leaks to full-scale cyberattacks. As a business organization, safeguarding the sensitive data you hold is not only a top priority but also a critical responsibility. The primary line of defence for your organization is not complex cybersecurity systems but rather a well informed, cautious, and vigilant professional team. Thanks to the implementation of Cybersecurity Awareness Training, you can generate a human firewall that shields your organization from a significant portion of cyber threats. What Topics Should Be Covered in a Cybersecurity Awareness Course? A wide range of Cybersecurity Awareness Courses are offered to help all kinds of companies strengthen their cybersecurity cultures. The Baseline cyber security controls for small and medium organizations require basic security training, including a focus on the following topics: - The use of effective password policies - Identification of malicious emails and links - Use of approved software - Appropriate usage of the Internet - Safe use of social media. Understanding the Cyber Threat Landscape within the Virtual World Before delving further into training specifics, it is sine qua non to grasp the cyber threat landscape and how it directly impacts your organization. The most common types of cyber threats include: - Phishing: This occurs when cybercriminals attempt to deceive recipients into providing sensitive data through seemingly legitimate emails and websites. - Ransomware: This malicious software denies access to a computer system or data until a ransom is paid. - Social Engineering: This involves manipulating individuals into divulging personal and confidential information. - Insider Threats: These intentional or accidental threats come from people working within the organization, such as employees, former employees, contractors, or business associates. They can be classified as either malicious or negligent in nature. Customized Cybersecurity Awareness Training When designing your Cybersecurity Awareness Training Program, it is noteworthy to consider your organization’s specific needs. Customize the content of that training to cover important areas such as password management, email messaging practices and Internet surfing, mobile device usage, social media risks, and recognizing and reporting potential cyber threats. Make use of real examples and case studies to illustrate the dangers and impacts of poor cybersecurity behaviour. In these training programs, it is recommended to perform a phishing simulation which will be covered in the next section of this Newsletter 58% of organizations report that their employees ignore cybersecurity policy and guidelines3 There are many organizations that offer Cybersecurity Awareness Training Program that will manage the deployments of courses and simulation exercises. There are a few of these organizations that are based in Canada. Among them is CIRA (Canadian Internet Registration Authority), a not-for-profit organization that offers such a service to Small and Medium Enterprises (SMEs). Some of you may know CIRA as the group that manages the .ca domain name on behalf of all Canadians and work to build a more trusted Internet for Canadians. Creating a proactive approach towards potential cyber threats is key. One of the most effective methods is by running periodic phishing simulation tests. This prevention process helps train your staff to recognize and promptly report phishing attempts. There are several free tools available: - Gophish: An open-source phishing toolkit designed for businesses and penetration testers. It provides the ability to quickly and easily set up and execute phishing simulations and cybersecurity awareness training. - Phishing Frenzy: An application that streamlines the phishing process while managing the results. Where to Find Online Cybersecurity Awareness Resources? Other online content provides an excellent platform to educate your staff about the basics of cybersecurity. Such courses often make available interactive and engaging content, rendering learning more effective. Among others, some free resources are namely: - ISC2 offers a Cybersecurity Awareness Training Course that is free to enroll. - Cybersecurity Awareness training Amazon internal awareness Quiz offer externally available in English and French. - Jigsaw, a Google company (Jigsaw | Phishing Quiz) offers a multilingual Quiz on phishing. - Rogers Cybersecure Catalyst part of Toronto Metropolitan University offers a free awareness training course for Small and Medium-Sized Businesses (SMBs), training aimed at corporate executives and IT personnel. - The Center for Development of Security Excellence Cybersecurity Awareness provides a series, of course, targeted to the US military and the US military industrial infrastructure. - Cybrary (https://www.cybrary.it/ ): offers a range of cybersecurity courses, including a Cybersecurity Awareness Course. These courses are designed to help businesses understand and combat potential risks associated with the cyber threat landscape. - SANS Cyber Aces (https://www.cyberaces.org/ ): offers free online courses about the core conceptual ideas on which cybersecurity is grounded. It is an excellent platform for increasing awareness and learning about the basics. - Get Cyber Safe is a national public awareness campaign created to inform Canadians about cyber security and the simple steps they can take to protect themselves online. Regular Training and Updates Cyber threats are not stagnant, they are ever-evolving. As such, your Cybersecurity Awareness Training must reflect such evolving changes. Ensure that your team is updated regularly about emerging cyber threats and cybersecurity practices. Such practices could involve periodic emails, workshops, or refresher courses Promoting a Cybersecurity Culture Beyond Cybersecurity Awareness Training Programs, promoting a culture of cybersecurity within your organization is paramount. This includes encouraging staff to take ownership of their cybersecurity both at work and home. This also involves nurturing transparency and reporting of potential cyber threats, and celebrating a culture of cybersecurity consciousness. The aim is to foster a cybersecurity-first mindset where every individual considers themselves as part of the organization’s cybersecurity solution. Assessment and Reinforcement Measuring the effectiveness of your Cybersecurity Awareness Training and its impact is crucial for the well-being of your company. Use assessments to gauge staff understanding and identify areas for improvement. Follow up with reinforcement activities to concretely apply the cybersecurity practices learnt during the training sessions, and create lasting changes which positively influence the behaviour of your personnel. Remember that the cost of Cybersecurity Awareness Training is just a tiny fraction compared to the potential damage that could be triggered by a cybersecurity breach. Actually, investing in the cybersecurity knowledge of your team, and fostering a culture of cybersecurity awareness is the best investment you can make for safeguarding your company against surging cyber threats. According to the far-sighted findings of The Global Risks Report 2022, 17th Edition, published by the World Economic Forum, 95% of all global cybersecurity issues can be traced to human error. Rapport Netwrix 2020 sur les cybermenaces
<urn:uuid:753debf0-f3d7-411f-a066-aa3bc476bdc1>
CC-MAIN-2024-38
https://insecm.ca/en/newsletter/why-cybersecurity-awareness-training/
2024-09-13T07:43:03Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651510.65/warc/CC-MAIN-20240913070112-20240913100112-00520.warc.gz
en
0.927002
1,622
2.515625
3
OK I admit that this is a bit farfetched. I don’t ever recall a test manager whose biggest concern would have been how to efficiently solve and test Sudokus. I was simply in a mood of having some fun and what would be better than working with a Sudoku problem. I bet that you have firsthand experience on solving those and just like me, you have spent a good deal of your air miles cracking those darn puzzles. Well at least I have, so bear with me. Instead of just opening up the last page of the local newspaper, I decided to create a Conformiq model that would solve my Sudoku problem. Probably not the most common use case of our technology but, like I said, I wanted to have some fun. What I did was that I created a Conformiq model that simply describes the rules of Sudoku. This is what we do with Conformiq technology; we describe the problem using rather abstract terms and let the technology solve the problem for us instead forcing us to do the hard mental work. In the case of classic Sudoku, the rules are pretty darn simple. [well]Classic Sudoku involves a grid of 81 squares divided into 9 blocks each containing 9 squares. Each of those 9 squares need to contain numbers from 1 to 9 and each number can only appear once in a row, column, and each 9 item square.[/well] The real problem is that, while the rules are simple and they are quick and easy to describe, the whole logic with all the combinatorial complexity makes solving Sudokus such a challenge. But if you think of it, this same thing applies to pretty much every non-trivial piece of logic out there: while the logic itself is quite simple, it lends itself to hugely complex scenarios. And this is where an efficient algorithm comes in. An algorithm is so much better in solving these combinatorial problems than our brains. Yes, our brains are arguably the most complex, elegant, and intriguing structure in the whole universe, but a human brain is just horrible in solving combinatorial problems. Compared to a computer, we really suck in solving combinatorial puzzles. We suck big time. So realizing that I’m no match for a computer, I simply gave my model with those simple rules to our Conformiq test generation engine and in no time I got my answer. It automatically generated the data needed to solve the rules to test the model. This brings me to the task of test design. How would you test the logic of a Sudoku application? You probably would have to solve multiple puzzles and verify the correctness of the results. But how would you do the design of such puzzle tests? Sure in the case of Sudoku you might say that the world is full of Sudoku puzzles and you would quickly find a set of examples to use. But what if you would not have that luxury? What if your Sudoku problem would be that Greenfield application that you are supposed to test? How would you go about then? Uh, that’s quite a pickle I would argue. However the reality is that testers and test managers face this situation each and every day. In the case of Sudoku I did what we humans are good at and described the problem, but then instead of trying to solve the problem myself, I outsourced the whole process of solving the problem, i.e. the process of test design, to a computer simply because it is so much better at it than I am. I was in a position to do so because I simply had proper tools available. I didn’t need to bother going through that hugely complicated, time consuming, and error prone process of solving the puzzle in my head and instead I merely described the rules of Sudoku and I handed over the real work to a computer. This Sudoku example, albeit something that as such we don’t need to deal with in our daily work lives, quite nicely demonstrates the fundamental difference between our Conformiq technology and manual test approaches and even with our competitor’s tools. Indeed model based testing solutions today that claim to automate the test design fall short here. Yes, on the surface level, the tools all look more or less the same. We all have boxes and arrows and by a click of a button you get tests. The reality however is that our competitors only solve the design of test flows (i.e., they provide the elementary machinery for iterating over the model paths) while test data design is left outside. Often the lack of this crucial functionality is hidden by all kinds of stories around great integrations with test data management tools and more while the real deal is that the vast majority of your test design, even after deploying these tools, still needs to be carried out manually. These tools wouldn’t help you a bit here. In the case of Sudoku this would mean that you would still need to manually solve the entire puzzle yourself even after having your test design “automated”. What…? Yes, you are left with solving the whole puzzle all by yourself. So much for automation… I do realize that most of you test managers out there do not lose sleep over a game of Sudoku, but I hope that this example highlights the need of not only automating the design of test flows but also the test data. If test data design is left outside or decoupled from the model logic you are bound to suffer from unproductive, error prone, and limited testing efforts. So maybe next time when you are in a position to affect your test automation tooling you should ask, “How would this solution help me in solving a Sudoku?” And me? Well, I got to have great fun with Sudoku modeling!
<urn:uuid:ed9dab35-14df-40aa-ae31-75476352a842>
CC-MAIN-2024-38
https://www.conformiq.com/resources/blog-solving-a-game-of-sudoku-using-a-test-generator-02-01-2018
2024-09-15T19:35:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651647.78/warc/CC-MAIN-20240915184230-20240915214230-00320.warc.gz
en
0.962727
1,183
2.546875
3
With many believing open-source software is more secure than proprietary software, we are now also seeing attempts to apply a similar theory to hardware development. At the 36th Chaos Communication Congress (36C3) hackers’ conference last month, however, experts Andrew “bunnie” Huang, Sean “xobs” Cross, and Tom Marble raised doubts about whether employing open-source development is enough to solve trust problems in hardware. Huang spoke at length on the topic. Differences between hardware and software in terms of trust Open-source software’s safety lies not only in its openness, but also in widely used tools that help ensure the program you run at the endpoint is true to the published source code. Programmers sign their software with a digital certificate, for example, and the system checks the certificate before running the software on a user’s computer. With hardware, it’s a different story. With no hardware analogs for hashing or digital signatures, users have no tools to check hardware’s authenticity against published information about it. The last time a device or chip is actually checked is at the factory. And the longer the gap between factory check and device use, the greater the chance of a successful MITM attack. What can go wrong? Generally speaking, anything at all can happen to chips or entire devices between leaving the factory and being used for the first time. To begin with, firmware can be replaced. (Sure, firmware is actually a software problem, so it can be verified, but you still have to rely on hardware during verification.) That’s why Huang focused on problems — component replacements, modifications, and implants — having to do strictly with hardware. These days, a totally unauthorized module can fit into a charging cable’s USB connector. Naturally, it’s even easier to tamper with more sophisticated multicomponent equipment that provides much more room for implants. The only good news here is that it’s relatively easy to detect added chip. The simplest substitution trick is to tamper with marking. One real-life example: A misbehaving microcontroller showed, on visual check, to have the right mark (from STMicroelectronics) on an altogether different chip. That time, the cheat was an expensive component replaced with a cheap one, but the replacement could have contained anything at all. People tend to think that chips cannot be modified once out of the factory, but that is not so. In many cases what we see as a single chip is actually several separate microcircuits in one package. An experienced adversary can use the same technology to put one more tiny piece of silicon into the very same package and connect this implant to existing contacts. In fact, equipment to do just that is relatively inexpensive and readily available (according to the speaker, a used wirebonding machine from China costs about $7,000), although the falsified results will be detectible in X-rays. Wafer-level chip-scale packages (WL-CSP) are much costlier to modify, but X-rays won’t reveal the deception. Integrated circuit (IC) modification Typically, companies design chips for their field-specific tasks but outsource them for production; only large market players can afford to produce their own chips. In this kind of arrangement, there is more than one way to modify the end product such that it still complies with the terms of reference. Moreover, after a chip or device is out of the designers’ hands, it’s rare anyone bothers to cross-check the resulting product against the original specifications. At what point can hardware be altered? The presenter offered several substitution scenarios ranging from fairly tricky (in-transit interception of cargo as an extreme example) to comparatively easy. Broadly speaking, anybody can buy a product, tamper with it, and return it to the seller, who can sell it again. And, formally, at various stages of procurement, the manufacturer’s packing team, customs agents, and many more parties have access to the equipment, and any of them can tamper with it if they choose. For all intents and purposes, using open-source hardware will not improve security much. Toward the end of his presentation, Huang speculated about what hardware production changes could enable end users to verify the safety of chips and devices. Those interested in the movement’s philosophy, as well as the technical details of chip modification, should view the presentation video. Not all of the many ways to make hardware dangerous are expensive or laborious, and most important, there is no direct correlation between an attack’s complexity and how difficult it is to detect. As for business users, stay mindful of the threat and do not rely solely on endpoint security products; corporate infrastructure protection systems fend off advanced threats and targeted attacks.
<urn:uuid:b9ad752b-41d4-4a18-b13a-3dfb0cee5d57>
CC-MAIN-2024-38
https://www.kaspersky.com/blog/36c3-open-source-hardware-dangers/32015/
2024-09-17T02:33:43Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651722.42/warc/CC-MAIN-20240917004428-20240917034428-00220.warc.gz
en
0.94531
1,004
2.859375
3
From TVs to watches, fridges, lightbulbs, or coffee machines, it seems everything needs to be connected now to be marketable. The Internet of Things (IoT) environment is growing in homes and workplaces, but it has established itself way ahead of regulation. IoT devices do not currently have to comply with any specific cybersecurity standards and malicious actors are already making use of these endpoints. Famously, a senior researcher at Avast even demonstrated in 2020 that he had successfully hacked his own coffee machine with ransomware. It’s amazing what some of us got up to in lockdown! With awareness of the risk growing, it is unsurprising that this month the European Union announced plans for a Cyber Resilience Act. The legislation aims to establish cybersecurity standards and stricter conformity assessment procedures for IoT. Connected devices are often seen as organisational cybersecurity’s weakest link, and regulation will doubtless improve cyber resilience—just as regulation improved data protection standards with the GDPR. Baseline cybersecurity standards for all connected devices, and stricter conformity assessment procedures for critical products, are certainly much needed. But this proposal raises many questions, such as, will legislation cover legacy products and those already in use? And how will small manufacturers manage the expense? It is not yet known if the intention is for a simple pass/fail system, or if there will be gradings assigned to help customers understand the value of more stringent security. What is known is that the threat landscape is not static, and so static certification for devices may not be appropriate, and may cause confusion as new threats emerge and attackers’ techniques evolve. If known vulnerabilities in IoTs are to be registered, consumers and companies will need to be educated so that they are able to make informed choices based on information about vulnerabilities and the risk exposure they open. Given the often low value and single use case for IoT devices, is it realistic to expect consumers to engage with potentially complex threats intelligence? On the supply side, organisations will face the challenge of not only having to architect, procure, implement, and roll out an adequate technology stack; they will also need to ensure they have processes and resources in place to operate as an ongoing concern. In the past, the EU has favoured the introduction of this type of legislation in one go, with a single date for compliance (in the way it introduced the GDPR). But my suggestion would be an incremental roll-out to enable policymakers to evolve the legislation. Make sure the end goal is clear, but start with an MVP or a pilot. This allows everyone to learn—including the regulator itself—by launching in one category or looking at minimal requirements before moving on to a potentially more stratified grading system. Overall, the regulation must be clear and precise on where responsibilities lie. The current proposal reads, “Obligations would be set up for economic operators, starting from manufacturers, up to distributors and importers, in relation to the placement on the market of products with digital elements, as adequate for their role and responsibilities on the supply chain.” This sentence hints at how broad the ecosystem affected will be, with complex R&D, manufacturing, and supply chain networks. In contrast to this complexity, the proposal currently calls for an “appropriate level of cybersecurity,” a very vague statement of intent which does not seem clear and concise enough not to trigger loopholes or to see stakeholders blaming each other in the future. It is time that IoT devices were better regulated for cybersecurity and resilience, but the devil—as they say—will be in the detail.
<urn:uuid:69faa823-9e7e-432b-a635-0acf24c43163>
CC-MAIN-2024-38
https://www.netskope.com/pt/blog/the-eu-cyber-resilience-act-thinking-out-implementation
2024-09-17T01:48:30Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651722.42/warc/CC-MAIN-20240917004428-20240917034428-00220.warc.gz
en
0.956249
728
2.625
3
Cloud Computing Disaster Recovery is a cost effective way to stay in the game when everything comes crashing down. Cloud Computing refers to storing your files onto servers across the world that will ensure your data is always available. When you use a cloud computing service for backups there are a few ground rules. When you first setup to the cloud, you are required to either upload your data or “sync” specific folders on your computer to the cloud network so that there is always an exact copy available online. Second, if you suffer a hard-drive crash and wish to restore your data, you should consider buying a new hard-drive and downloading your backup data onto the new drive. Third, and this is a leap of faith, many users are working directly on the cloud and accessing files remotely without saving them to their hard-drive. Working in the cloud exclusively is becoming standard practice among certain groups, but be aware and always “sync” your local drive to the cloud so that you can access your files if your internet ever goes out, or your access to the cloud service is unavailable. Although the nature of cloud computing means more uptime, Murphy’s Law will always triumph. The beauty of cloud computing disaster recovery is that your data is stored in multiple locations and if the cloud network ever does go down, another “node” in the cloud picks up the slack. During emergencies, cloud based systems are redundant and established enough to redirect all internal network routing to the nearest available node, sacrificing mere milliseconds in the process and bringing the world closer together when we need to be connected the most. Article by Scott Huotari, President CCSI, Google | LinkedIn
<urn:uuid:8d263937-37aa-4f98-944a-6b137ae449dc>
CC-MAIN-2024-38
https://www.ccsipro.com/blog/cloud-computing-disaster-recovery/
2024-09-20T18:08:11Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725701419169.94/warc/CC-MAIN-20240920154713-20240920184713-00820.warc.gz
en
0.925859
342
2.8125
3
The Ethics of ML and AI AI will enable breakthrough advances in areas like healthcare, agriculture, education and transportation; it's already happening in many ways. But new technology also inevitably raises complex questions and broad societal concerns. As we look to a future powered by a partnership between computers and humans, it's important that we address these challenges head on and address: - How do we ensure that AI is designed and used responsibly? - How do we establish ethical principles to protect people? - How should we govern its use? - And how will AI impact employment and jobs? To answer these questions, technologists will need to work closely with government, academia, business, civil society and other stakeholders. And focus on ethical principles - fairness, reliability and safety, privacy and security, inclusivity, transparency, and accountability - to guide the cross-disciplinary development and use of artificial intelligence for business and cyber. In this talk we'll share the principle ethics of AI & ML and have a discussion about how we can all work together to forward AI and ML use responsibly.
<urn:uuid:c3de585d-5dfa-4a99-abcf-74eb1e441802>
CC-MAIN-2024-38
https://ffiec.bankinfosecurity.com/webinars/ethics-ml-ai-w-2274
2024-09-09T20:50:43Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651157.15/warc/CC-MAIN-20240909201932-20240909231932-00084.warc.gz
en
0.957051
218
2.9375
3
China is expanding the scope of its biometric network in response to the COVID-19 pandemic. The Guangzhou Public Transportation Group is now using biometric tablets to check the temperature (and the identity) of every passenger who boards one of the city’s buses. The tablets themselves are being installed next to the driver’s seat on the buses in question. To use them, passengers must place their forehead close to the tablet to check their temperature. The tablet will also take a photo of each passenger, allowing them to be identified using China’s vast facial recognition network. Public health officials are hoping that the system will help curb the spread of the coronavirus. People with a high temperature may be displaying symptoms of the virus, while the use of facial recognition will allow state officials to trace anyone who has had contact with a passenger who later receives a COVID-19 diagnosis. The news suggests that China increasingly regards temperature tracking as an effective way to combat COVID-19. Al Jazeera recently reported that the country is already using infrared scanners in conjunction with surveillance cameras to monitor the progression of the disease, while the Chinese AI specialist Hanvon claims that it has developed a facial recognition solution that is 95 percent accurate when identifying people wearing medical masks. The Public Transportation Group claims that the data collected through the system will only be used for public health purposes. However, participation is not optional for Guangzhou commuters, and China has been criticized for its invasive biometric practices in the past. The Nikkei Asian Review notes that while the system may be a useful public health measure, it is unlikely to be implemented in Europe or North America, where there is far more resistance to government surveillance practices. Earlier this month, the ACLU sued the U.S. government for access to records about its use of facial recognition at the country’s airports. Source: Nikkei Asian Review March 30, 2020 – by Eric Weiss
<urn:uuid:ce2a861e-b586-4df1-9cc3-6c76f75b1e5e>
CC-MAIN-2024-38
https://findbiometrics.com/guangzhou-deploys-biometric-tablets-track-coronavirus-city-buses-033010/
2024-09-12T06:12:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651422.16/warc/CC-MAIN-20240912043139-20240912073139-00784.warc.gz
en
0.946386
396
2.625
3
As defined by most of the scholastic articles, the traditional security architecture is built on ‘castle-and-moat’ concept, but things are gradually changing – a new concept called Zero Trust is catching up. As per the traditional concept, once the person or device is authenticated, he or the device can access the resources within the perimeter of the network but as per the new concept (Zero Trust), even the person or the device is authenticated, they cannot move freely or access data at their will. They will be entrusted with a restricted access. Although the concept has been floating in the market for last 10 years, ever since Forester coined it but people were not convinced about it therefore not taking it so seriously. But as the situation of Covid 19 Pandemic is panning out and attacks like malware, impersonation of identity, email extortion campaigns, worms, phishing attacks, botnet, etc. have happened in these 45 days. Large very secured companies like Cognizant and LinkedIn have borne the brunt of the attackers, every CISO is now taking it seriously. Therefore, Zero Trust Security appearing to be the future of the information security. Zero Trust Security prescribes many procedures to isolate intruders. One of those is surface area reduction by microsegmentation. Microsegmentation is not new in network but in DC and compute it makes logical units with own unique security policies. By doing so, it limits the malicious actors to a restricted surface and easily traces if there is unwanted east-west or north-south traffic. Secondly, Zero Trust Security advocates for a threat modeling of no one is trusted by default from inside or outside the network, and verification is required from everyone trying to gain access to resources on the network. It means Multi-factor authentication (MFA) is also a tenet of this new concept which askes your password along wide 2FA to authenticate yourself. It also prescribes least-privilege access, which offers the users very limited access as per the company policy parameter – not beyond that. This obviously reduces the chance of exposing sensitive data. As far as devices are concerned, Zero Trust also monitors and controls different devices accessing network – again a step further to reduce the coverage area. Of course, Zero Trust Security also advocates for cognitive pattern recognition of the attacks by integrating machine learning with forensic network analytics. It automates identification of the modus operandi of the actors and therefore nullifies the attempts. So, in a nutshell, Zero Trust Security induces to review existing asset management policies, reduces and restricts overly permissive rules, strengthens vulnerable points and makes attack surface smaller. It also means taking control of IT security completely. Greg Bryan, Senior Manager, Enterprise Research at TeleGeography, says, “Zero Trust Security is making the move from buzzword to serious consideration.” But the question is since it is not a technology, can it be right practice or there is technology which creates a canopy of Zero Trust in the network? The other question is: what about customers, who want free access to their resources outsourced to the service providers? Would they be happy about sailing in a restricted zone in their own resources? “The challenge for CIOs is to understand the foundations necessary to make ZTS a viable security solution. For example, in order to have user or device-based security policies, you first need to identify every user and device on your network—no small task for many enterprises. Regardless, any time a new architecture is proposed there will be hesitation and a period of assessment before adoption ramps up,” he added. “Interest in Zero Trust Security remains high among WAN managers, while only 8% have implemented the IT security model. 31% are considering Zero Trust Security, 19% are in the adoption phase with a fifth of respondents unfamiliar with the concept.” Mind it, as per a report new organization falls victim to ransomware every 14 seconds. The period will further contract to 11 seconds by 2021 and damages would reach beyond $11.5 billion.
<urn:uuid:2a103af1-1328-4d21-b031-fa4281d50afa>
CC-MAIN-2024-38
https://www.enterpriseitworld.com/rise-of-zero-trust-security/
2024-09-13T12:41:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651513.89/warc/CC-MAIN-20240913101949-20240913131949-00684.warc.gz
en
0.954071
829
2.59375
3
The job of a DBA is special in any organization. Large companies might have anywhere from three to six (or even more) DBAs on staff. But for small to mid-sized organizations, oftentimes there is only a single DBA putting on a one-person show. And they might not even be a DBA by training; there is even a term that describes a developer or system administrator who takes on the role of DBA because nobody else is doing it (SolarWinds’ white paper, “The Accidental DBA”). But What Does a DBA Do? The DBA is responsible for managing the security, performance, and integrity of a database system. The DBA should get involved in database management from the planning and development phase going forward. Once the database is set up and in use, the DBA needs to make sure that the data is clearly defined and remains consistent; that data security, recovery, and backup processes are in place; and that users are given support with troubleshooting issues that they come across in their day-to-day work. Generalists Versus Specialists Database management involves a lot of different disciplines: logical database design, physical design, performance analysis and tuning, data warehousing, auditing and logging, server configuration, troubleshooting, and monitoring. Some DBAs prefer to get involved in all of these on a general level, doing a little bit of everything. Others specialize in one or two areas, learning and practicing those disciplines in-depth. Frequently, the DBA’s employer dictates whether the DBA is a generalist or a specialist, as well as what specialty to focus on. When an organization is small and cannot afford to have multiple DBAs, or the database accounts for only the supportive functions of the core business, it is more likely to hire a DBA with a generalized skill set. Since the capacity in each discipline will be relatively lower, a DBA with a good skill set and expertise should be able to handle it well. On the other hand, in larger organizations, where there can be many database systems and related tasks, it would be difficult for one or two DBAs to handle everything. This could also be the case when the database system is the core business of the organization, such as in organizations that sell data or provide analytics solutions. In such cases, the organization may hire several DBAs who are specialized in the most critical database disciplines used by the organization. There are several accepted database disciplines in which a DBA can specialize, as outlined below. A Database Architect is involved in the overall design and implementation of database systems. A database architect needs to gauge the business requirements and decide how best to design the database to cater to the business needs. They are only involved in the design and implementation of the database. A System DBA focuses on the administration-related activities of the database system. This includes server configuration, installations and upgrades, managing of support technologies, etc. An Application DBA focuses on database design, development, and writing SQL functions to perform CRUD operations for database applications. They are responsible for change management and performance tuning of the database for the particular applications they’re working on. A Performance Analyst focuses on database performance. Performance analysts constantly monitor the database for performance bottlenecks, including the SQL coding that results in performance issues, and handle performance-tuning tasks. They must have a good understanding and technical knowledge of advanced database concepts. Which types of DBA skills are needed for your organization? It is important to assess this, especially if you are hiring multiple specialized DBAs. To learn in detail about the different types of DBAs and their responsibilities, download “The Many Different Types of DBAs,” a white paper I wrote for Datavail (www.dbta.com/DBTA-Downloads/WhitePapers/The-Many-Different-Types-of-DBAs-6987.aspx). This will help you make sure that you are investing in the right set of people to take your organization to the next level. You can also find an in-depth discussion of DBA responsibilities, in my book Database Administration: The Complete Guide to DBA Practices and Procedures.
<urn:uuid:9b0fea03-e655-4b66-9ce2-90b7fb1fd190>
CC-MAIN-2024-38
https://www.dbta.com/Columns/DBA-Corner/What-Type-of-DBA-Are-You-121146.aspx
2024-09-15T23:45:45Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651668.26/warc/CC-MAIN-20240915220324-20240916010324-00484.warc.gz
en
0.946991
869
2.65625
3
This cybersecurity white paper explores the origins and motivations of cyberwarfare, and provides actions you can take in response. Elements of Cyberwarfare There are a wide range of cyber actions that can be used as an element of cyberwarfare. Think of these as primarily offensive actions, although some may have a defensive purpose as well. These weapons must be developed and maintained for a nation to be able to perform an act of cyber war. Most of the concepts typically related to hacking or penetration testing can be used as cyberwarfare tools, including reconnaissance, scanning, system identification, and vulnerability scanning. These are tools and techniques that assist the attacker in gaining more knowledge about the specifics of the target’s IT infrastructure. Once foundational knowledge is obtained about the target, the next element is to take advantage of a known or assumed vulnerability of the target. This is often known in the penetration testing community as gaining access, but it could also focus on damage and destruction instead of just remote control or intrusion.
<urn:uuid:ee7089f2-21ec-41dc-a81f-94ec76b4fbf2>
CC-MAIN-2024-38
https://www.globalknowledge.com/us-en/resources/resource-library/white-papers/cyberwarfare-origins-motivations-and-what-you-can-do-in-response/
2024-09-17T04:58:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651739.72/warc/CC-MAIN-20240917040428-20240917070428-00384.warc.gz
en
0.957499
205
3.03125
3
Yes, I realize that’s an ambitious title for a blog post about staying secure online, but there are a handful of basic security principles that — if followed religiously — can blunt the majority of malicious threats out there today. Krebs’s Number One Rule for Staying Safe Online: “If you didn’t go looking for it, don’t install it!” A great many online threats rely on tricking the user into taking some action — whether it be clicking an email link or attachment, or installing a custom browser plugin or application. Typically, these attacks take the form of scareware pop-ups that try to frighten people into installing a security scanner; other popular scams direct you to a video but then complain that you need to install a special “codec,” video player or app to view the content. Only install software or browser add-ons if you went looking for them in the first place. And before you install anything, it’s a good idea to grab the software directly from the source. Sites like Majorgeeks.com and Download.com claim to screen programs that they offer for download, but just as you wouldn’t buy a product online without doing some basic research about its quality and performance, take a few minutes to search for and read comments and reviews left by other users of that software to make sure you’re not signing up for more than you bargained. Also, avoid directly responding to email alerts that (appear to) come from Facebook, LinkedIn, Twitter, your bank or some other site that holds your personal information. Instead, visit these sites using a Web browser bookmark. Krebs’s Rule #2 for Staying Safe Online: “If you installed it, update it.” Yes, keeping the operating system current with the latest patches is important, but maintaining a secure computer also requires care and feeding for the applications that run on top of the operating system. Bad guys are constantly attacking flaws in widely-installed software products, such as Java, Adobe PDF Reader, and Flash. The vendors that make these products ship updates to fix security bugs several times a year, so it’s important to update to the latest versions of these products as soon as possible. Some of these products may alert users to new updates, but these notices often come days or weeks after patches are released. Krebs’s Rule #3 for Staying Safe Online: “If you no longer need it, remove it.” Clutter is the nemesis of a speedy computer. Unfortunately, many computer makers ship machines with gobs of bloatware that most customers never use even once. On top of the direct-from-manufacturer junk software, the average user tends to install dozens of programs and add-ons over the course of months and years. In the aggregate, these items can take their toll on the performance of your computer. Many programs add themselves to the list of items that start up whenever the computer is rebooted, which can make restarting the computer a bit like watching paint dry. And remember, the more programs you have installed, the more time you have to spend keeping them up-to-date with the latest security patches.
<urn:uuid:e68cdacb-0731-4470-bcd4-56c0ee47cfb9>
CC-MAIN-2024-38
https://krebsonsecurity.com/2011/05/krebss-3-basic-rules-for-online-safety/
2024-09-18T12:18:58Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651895.12/warc/CC-MAIN-20240918100941-20240918130941-00284.warc.gz
en
0.937192
670
2.515625
3
The European project Mignei It is a fact nowadays with the rapid technological development in all productive sectors and at different levels to have a replacement with the new technological methods. A project in the framework of a European research project called MIGNEI aimed to capture some ancient construction activities, with particular interest and to capture an overall picture, such as the materials they used, the manual movements and techniques of the producers, but also the general geographical, architectural and cultural context in which these arts were born. This approach of the research program was done in some areas that are specific to the production product. One of them is cheese-making. Computer scientists have decided to create a database that, in addition to cheese-making, includes loom weaving, silk weaving, traditional ceramics, dyeing clothes and blown glass. The way of approaching the connection of traditional arts and practices with the historical and geographical environment that has been developed is done with the interactive augmented reality and other modern tools. Τhe latest digital technologies help people to understand, protect and promote old and traditional crafts. The Mignei project was a way to help innovate ideas that engage today’s youth with older traditions. For example, in the cheese industry, technology was used to create a 3D scan of the tools used in traditional cheese production. So we have a form of digitizing all the processes of traditional cheese production with information that is transmitted using augmented reality including narration through a fascinating presentation of all the details of the production process. In many cases there are also educational applications to fully capture the traditional production technique using the internet and augmented reality. The case of glassmaking Glassmaking is a special art with many different uses for finished products. It is worth mentioning that the Conservatoire national des arts et métiers (CNAM) has been located in Paris since 1974. It is a museum of traditional and industrial arts and crafts as well as technological innovation. There are thousands of inventions, divided into 7 collections (Scientific Instruments, Materials, Energy, Engineering, Construction, Communication and Transportation) including glass. Glassmaking combines artistic skills with a thorough technical and sensory understanding of the material. With a set of simple tools and processes, glass blowers are able to produce an almost endless variety of items, from decorative and fragile to durable and functional. Glassmaking and glass blowing exhibit a range of expressions from craft to industrial, and have been practiced around the world for centuries. It is therefore an excellent opportunity to use important technological tools to capture the gestures of modern glassmakers (tools, light machines and glassware, etc.) and to explore “reverse engineering” techniques for “reinventing gestures”. In this way, the creators of rare classic objects can convey information, thus integrating it into motion-based narratives for its visitors, in an MRI environment. The case of weaving The structure of weavers (weaving) has not changed from the Neolithic era (from the first evidence of the existence of the weaver) to today’s weavers, made with the most advanced technologies. The technological development that took place due to the effort to improve the textile industry marked the Industrial Revolution that began in 18th century England and its aftermath. The evolution of weaving technology moved in two parallel directions: on the one hand, the need for easier and faster production led to the industrialization and eventual automation of repetitive movements of plain weaving (England 19th century). The introduction of patterns in the fabrics led to the design of the loom so as to introduce intricate fabrics during weaving (France 19th century). A development that inspired the later programming of computers. Both of these developments were combined in the form of the jacquard loom, which is perhaps the most important milestone in the development of textile technology. Advanced looms offered more opportunities for design and production. The use of jacquard marked a new era. The looms and the way of producing fabrics continue to evolve. The Haus der Seidenkultur (HdS) is an industrial monument that serves as a museum in Krefeld. The museum hosts numerous activities for all age groups and uses technology: workshops, tours, temporary exhibitions and events make the museum a vibrant experience where the world of silk and the history of Krefeld textiles invite visitors to participate in an exciting discovery journey.
<urn:uuid:9adcfdf2-48ca-4d17-bd18-45258cf7bd99>
CC-MAIN-2024-38
https://4imag.com/how-technology-helps-traditional-arts-and-professions/
2024-09-19T18:48:33Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652055.62/warc/CC-MAIN-20240919162032-20240919192032-00184.warc.gz
en
0.950969
890
2.96875
3
Data center failures are happening too often, according to a presentation at DCD SE Asia in Singapore, and the reason is a lack of information sharing. Recent data center failures have included the Singapore Stock Exchange (SGX, which had a major outage in 2014 due to a combination of factors), but this event was an exception, according to Ed Ansett of i3 Solutions, one of the experts called in to consult the organization after it happened. Where SGX shared its experience, other failures were shrouded in secrecy because of the competitive nature of the industry. Non-disclosure agreements prevent learning ”Data centers are bespoke complex homo-technical systems. The human side and the technical side cannot be separated. An understanding of both is required to reveal what went wrong - and all too often the failure is one that has been seen before,” Ansett told the conference. Other industries such as aviation have the same problem, but achieve much higher levels of reliability because of regulations requiring accident investigation, said Ansett, ”We are a younger industry, and we are unregulated. There is no authority that looks at data centers. The closest we come are safety regulations, or mandates laid down by financial services authorites such as The Monetary Authority of Singapore (MAS).” One reason for this is that data center failures do not cause loss of life, unlike plane crashes. A number of failures could be avoided if, for instance, data center operators refused to use residual current detectors (RCDs) in racked servers, said Ansett. These are designed to protect people from electrocution by devices such as lawnmowers, so they are inappropriate in a data center, and they can also trip unpredictably at a level lower than their 30mA recommended setting, causing a server failure that can cascade into something worse. Because failure data is not shared, knowledge like this is not available to all, and data centers achieve far lower continuous uptime than they could, said Ansett. He expects that this might change in future. Data centers are becoming more and more important to the increasingly instrumented world, and in his view, eventually a data center failure will happen that kills people. At this point, the industry will have to accept regulation, which will enforce sharing of the data.
<urn:uuid:b59d7ff2-5003-4f04-a56e-c55d298b2a55>
CC-MAIN-2024-38
https://www.datacenterdynamics.com/en/news/dcd-se-asia-many-data-center-failures-are-due-to-secrecy/
2024-09-08T20:03:36Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651017.43/warc/CC-MAIN-20240908181815-20240908211815-00284.warc.gz
en
0.96697
466
2.53125
3
Personal Identity Information or Personally Identifiable Information (PII) is information that permits the identity of an individual to be directly or indirectly inferred. An example of PII would be a person’s full name combined with a Social Security Number, passport number, or email address. Personally Identifiable Information is often found on public websites including your Full Name combined with your zip code, race, gender, address, and date of birth. Pro Tip: CyberHoot suggests answering challenge questions, used by websites to help you recover access to an online account, with made-up answers. Just be sure to save the bogus answer in your password manager. It's too easy for hackers to find our true PII on various public websites and break back into our online accounts if we tell the truth on PII Q&A. Multiple data protection laws such as the General Data Protection Regular (GDPR) and the California Consumer Protection Act (CCPA) have been adopted by various countries and US states to create guidelines for companies who gather, store, process, and sell the PII of their clients. The basic principles outlined by these laws provide restrictions and rights to consumers for managing and controlling access to their own sensitive information (PII). Most of these privacy laws allow users to ask the following questions or make these demands of companies: - What data do you have about me? - Don’t sell my data to any 3rd parties. - Delete or correct my data. What Does This Mean For My SMB? Legislation is in place throughout most of the world, so businesses should focus on examining their local country’s (or state’s) data privacy rights. Once you understand your legal obligations in your own home country or state, you’ll want to develop each of the following: - a breach notification plan. - An authentication and response plan to privacy requests from consumers (if you do collect PII on consumers). THE IMPORTANCE OF A BREACH NOTIFICATION PLAN It’s very important to have a plan to notify any users and applicable agencies involved in a data breach. The time to work out how to notify and what the applicable home country laws and requirements are is not during a breach but before one. In the US exists a patchwork quilt of data privacy laws by individual states. There is no federal law in effect. In preparing your breach notification process, make sure you seek outside counsel and expertise to help you comply with each state you may have data on (or country as the case may be). DATA PRIVACY REGULATIONS BY COUNTRY (A SAMPLE): - European Union’s General Data Protection Regulation (GDPR) - California Consumer Protection Act (2018) - Australia’s Privacy Amendment (Notifiable Data Breaches) Act of 2017 - Personal Information Protection and Electronic Documents Act (PIPEDA) - Germany’s Data Privacy Law Compared to GDPR (BDSG) CyberHoot recommends you know the laws affecting your company and you prepare a Breach Notification process document reflecting those laws. Don’t forget to test the process annually as well. Importance of Authenticating Data Privacy Requests Let’s say your company is a hotel. You receive a request email to email@example.com for all the dates you stayed at that hotel from John Doe during the past year. The email states this is a GDPR request and you have every right to the answer. It came from firstname.lastname@example.org. Should you respond? No. Jane Doe has no right to John Doe’s information under GDPR or CCPA or any other legislation. Your authentication process must engage via a unique identifier which is typically the email address owned by the person making the request. Under privacy legislation, you are required to verify the identity of the privacy requestor before providing any PII Data back. SMB PROTECTIONS BEYOND Data Privacy Regulations CyberHoot recommends the following best practices to protect individuals and businesses against, and limit damages from, online cyber attacks: - Adopt a password manager for better personal/work password hygiene - Require two-factor authentication on any SaaS solution or critical accounts - Require 14+ character Passwords in your Governance Policies - Train employees to spot and avoid email-based phishing attacks - Check that employees can spot and avoid phishing emails by testing them - Backup data using the 3-2-1 method - Incorporate the Principle of Least Privilege - Perform a risk assessment every two to three years If you would like to learn more about this topic, watch this short video: Source: NCSD Glossary, CNSSI 4009, GAO Report 08-356, as cited in NIST SP 800-63 Rev 1, Investopedia Secure your business with CyberHoot Today!!! CyberHoot does have some other resources available for your use. Below are links to all of our resources, feel free to check them out whenever you like: - CyberHoot’s Blog - Cybrary (Cyber Library) - Infographics by CyberHoot - CyberHoot’s Monthly Newsletters - CyberHoot Press Releases - CyberHoot Platform Instructional Videos (HowTo) – very helpful for our Super Users! Note: If you’d like to subscribe to our newsletter, visit any link above (besides infographics) and enter your email address on the right-hand side of the page, and click ‘Send Me Newsletters’. Sign up for the monthly newsletter to help CyberHoot with its mission of making the world ‘More Aware and More Secure!’
<urn:uuid:760023a5-9a29-4d43-9735-7f39c7702045>
CC-MAIN-2024-38
https://cyberhoot.com/cybrary/personal-identifying-information-pii/
2024-09-10T00:49:32Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651164.37/warc/CC-MAIN-20240909233606-20240910023606-00184.warc.gz
en
0.900959
1,203
2.765625
3
We highlighted the different application models for how an application can be created earlier. This section takes that discussion a step further by covering the different service models that can be leveraged for the application. These service models determine where the application is located and what elements of the application are owned and managed by the business. The following are the most common service models: On-premises: On-premises is the service model where a business owns and manages an application. A business will procure all of the infrastructure required to run the service and then fully manage, maintain, and operate it. In some situations, the management is outsourced but the infrastructure is procured and owned by the business. Software as a Service (SaaS): SaaS is where a vendor makes its software available to users, usually for a monthly or annual subscription service fee. Platform as a Service (PaaS): PaaS is where a vendor provides hardware and software tools, and people use these tools to develop applications. PaaS users tend to be application developers. Infrastructure as a Service (IaaS): IaaS is a pay-as-you-go service model for storage, networking, and virtualization—all of a business’s infrastructure needs. IaaS gives users cloud-based alternatives to on-premises infrastructure, so businesses can avoid investing in expensive onsite resources. Table 3-3 shows the comparison between these service models. Table 3-3 Service Model Comparison Service Model | When to Use | Business owned and managed. Available locally. Hosted within the business’s server environment. | Full control over all components of the application. | When a business requires full control of all components within the application. This is most often seen with security compliance and data classification requirements. | Available over the Internet. Hosted on a remote server by a third-party provider. Scalable, with service offerings based on need. | No need to install and run software on any computer. Everything is available to the end user over the Internet. Access to software can be from any device, at any time, with Internet connectivity. | When a business wants an application to run with ensured availability but without the headache of maintaining that application at any level. | Accessible by multiple users. Scalable. Built on virtualization technology. Easy to run without extensive IT knowledge. | Primarily used by developers to create software or applications. Developers don’t need to start from scratch when creating applications. | When a business wants to create a unique application without spending a ton of money or taking on all the responsibility. | Highly flexible. Highly scalable. Accessible by multiple users. Cost-effective. | On-premises IT infrastructure is expensive. The business maintains control over the infrastructure. | When a business requires complete control over its infrastructure and wants to operate on a pay-as-you-go basis. |
<urn:uuid:dc7d762e-1940-4512-a6cc-53358544b137>
CC-MAIN-2024-38
https://www.ciscopress.com/articles/article.asp?p=3178906&seqNum=3
2024-09-12T09:46:06Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651440.11/warc/CC-MAIN-20240912074814-20240912104814-00884.warc.gz
en
0.931184
611
2.546875
3
Microsegmentation enhances security by limiting an attacker’s movement and controlling access in specific areas of a data center or cloud setup. It aims to restrict lateral movement (east-west traffic) and improve security within the network. This technique helps prevent unauthorized access and reduces the potential damage caused by an attacker. By dividing the network into smaller isolated segments, microsegmentation provides an additional layer of protection against cyber threats and unauthorized access. This process has been around for a long time in network security. It has become easier to use with software-defined networking and more virtualization. Top microsegmentation vendors provide products that create secure areas in data centers and cloud networks. These products enforce strict rules on traffic between these areas to prevent security incidents. Microsegmentation limits the risk of breaches by reducing attack surfaces and restricts damage attackers can inflict by limiting their ability to expand access across a system. With virtualization and cloud, microsegmentation allows organizations much more granular control in security architecture.
<urn:uuid:e654a0ce-1488-496c-8774-1303c276ddc1>
CC-MAIN-2024-38
https://www.netskope.com/de/what-is-microsegmentation
2024-09-19T21:29:08Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652067.20/warc/CC-MAIN-20240919194038-20240919224038-00284.warc.gz
en
0.931894
205
2.609375
3
Without Maturity, there’s no Cybersecurity Date: 13 October 2020 There is no such thing as an "unhackable" system. In general, cybersecurity can be described as ensuring that there is no unauthorised access to an enterprise’s network, to applications, to databases and finally to all the data we want to keep safe and secure. Perhaps, the most “secure” system would be one which does not have any network connection whatsoever. Even then, this system is still vulnerable to physical attacks, theft, or, in some cases, radio frequencies being used to spy on the computer system remotely, and even influence its behaviour. In this blog, we cover: The limited knowledge computer users and system administrators have about computer network infrastructure and the working of its protocols does not help advance network security. In fact, it increases the dangers. In a mechanical world where users understand the systems, things work differently. For example, in a mechanical system like a car, if there is a fundamental mechanical weakness, the driver usually finds the weak point and repairs it. This, however, is not the case with computer networks. As we have seen, the network infrastructure has weaknesses and this situation is complicated when both system administrators and users have limited knowledge of how the system works, its weaknesses and when such weaknesses are in the network. This lack of knowledge leads to other problems that further complicate network security: - Network administrators do not use effective encryption schemes and do not use or enforce a sound security policy. - Administrators and users who are less knowledgeable quite often use blank or useless passwords, and they rarely care to change even the good ones. - Users carelessly give away information to criminals without being aware of the security implications. - Network administrators fail to use system security filters. According to security experts, network servers without firewalls “are the rule rather than the exception.” Core Concepts of Cybersecurity Cybersecurity is not about creating an “unhackable’ system. Cybersecurity is about reducing the risk of a system being breached (confidentiality), modified (integrity), or disrupted (availability) without authorisation. These three concepts - confidentiality, integrity, and availability - make up the core foundations of any cybersecurity programme. Here then are the key questions to ask to ensure that the three concepts have been paid heed to: - Confidentiality - Is the data protected from disclosure to unauthorised users? - Integrity - Is the data protected from modification by unauthorised users? - Availability - Is the data accessible to authorised users for review or modification? Too often, the third concept (availability) is overlooked, or ignored when planning a secure system. While confidentiality and integrity of data is important, that data is useless, if it is not available to authorised users. As I said above, security means no unauthorised access, which means no broken authentication by SQL injection (SQL queries), brute force (dictionary attack, common credential), buffer overflow (pass some random data), social engineering (getting the exact credentials), man-in-the-middle (stealing credentials) and session management (using cookies, session ID), that can lead to (sensitive) data exposure. Tenets for Success of Security Measures - Vulnerability: A flaw in a design, a possible means of compromising confidentiality, integrity, or availability. - Threat: An external or internal force, either man-made or natural - Risk: The probability of a threat successfully attacking a vulnerability For security measures to be successful, the security measures must cost less than the predicted loss, (should confidentiality, integrity, or availability be compromised, while making an attacker’s opportunity cost of defeating those security measures, higher than the value of a successful compromise?). Although the network infrastructure weaknesses seem simple, finding solutions will not be easy and it is an ongoing exercise of interest to lawmakers, law enforcement agencies and the network community. The ‘Holy Grail’ is to find a final solution to the complex computer network security problems. Even if this Holy Grail will succeed, the solution will not last long, for many of the following reasons: - The cyberspace infrastructure technology is constantly changing, adding new technologies along the way and as new technologies are added, new loopholes crop up, therefore, new opportunities are created for cyber-criminals. - Solutions to social and ethical problems require a corresponding change in the legal structures, enforcement mechanisms and human moral systems. None of these can change at the speed with which technology is changing. Soon, any solution will be useless, and we will be back to square one. - Yet, there is no fully functional national or international plan or policy that can withstand the rapid changes in technology and remain enforceable. - Most importantly, solutions that do not consider and are not part of a general public education plan, do not stand a chance of lasting for any extended period. For any solution to the computer network security problem to last, public education and awareness are critical. Possible Solutions to the Computer Network Security problem As a personal opinion, a workable and durable solution (if found), must include the following: • Public awareness and understanding of the computer network infrastructure threats, its potential consequences and its vulnerabilities. We cannot rely on education acquired from science-fiction novels. Otherwise, when such attacks really occur, the public may take them to be science- fiction events. • A well-developed plan based on a good policy for deterrence. • A clear plan, again based on good and sound policy, for rapid and timely response to cyber-attacks. However, we should not ignore the inconveniences, or the social and ethical disruptions that are perpetuated by technology. Our duty is to find ways to prevent future computer attacks. And our focus is to understand what they are, who generates them and, especially, why. Author: Oana Buzianu A passionate information security professional who has made cybersecurity a priority in her career. With 15+ years of experience as a cybersecurity specialist and a deep understanding of intelligence processes, Oana is focussed on shifting the focus away from rules and policies to values and ethics & doing the right thing even if no one is looking.
<urn:uuid:af318b06-3b48-455c-8b26-8436d8057d48>
CC-MAIN-2024-38
https://www.cm-alliance.com/cybersecurity-blog/without-maturity-theres-no-cybersecurity
2024-09-21T04:21:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725701427996.97/warc/CC-MAIN-20240921015054-20240921045054-00184.warc.gz
en
0.935939
1,286
3.359375
3
Did you know that 2,200 cyber attacks happen every single day? With that sort of relentless onslaught, it might make you wonder who has that much time on their hands. But, the reality we know all too well is that while human attackers may be orchestrating the attacks, much of the risk has to do with automated threats. Automated systems disrupt, tamper with, or gain unauthorized access to an application, network, or digital system. They differ from traditional security risks in that automated threats use — you guessed it — automated methods, such as bots, to orchestrate their attacks. In contrast, traditional security risks rely on simpler approaches to execute their attacks. Automated threats can cause a lot of damage to the attacked systems, and it’s not uncommon for them to lead to data breaches or system downtime. With AI moving at lightning speed, automated attacks will only escalate. Individuals and businesses, therefore, must take proactive measures to safeguard their digital assets against such malicious activities. App developers should be concerned about automated threats for several reasons. Firstly, these threats exploit vulnerabilities in applications to carry out malicious activities like data breaches, identity theft, and unauthorized access. This not only compromises user data and privacy but also undermines the integrity and reputation of the app. Automated threats also lead to service disruptions, such as DDoS attacks, which overwhelm the app’s servers, leading to downtime and loss of revenue. Moreover, these threats often evolve rapidly, making it challenging for developers to keep up with the latest security measures. These disruptions include skewed analytics and data that impact business decisions. Automated threats encompass a range of malicious activities executed through automated software. These include bots to carry out credential stuffing, where stolen account credentials are tested en masse against various websites. There are also scraping bots, which systematically extract large amounts of data from websites, potentially infringing on copyright or stealing sensitive information. Below is a breakdown. The emergence of Artificial Intelligence (AI) has paved the way for the automation of online tasks, with bots or robots playing a significant role in this process. While bots have helped streamline various tasks, they have also made it easier for cybercriminals to launch attacks on computer systems. Cybercriminals rely on bots to execute the most sophisticated attacks, such as: Twitter (X)bots are an example of automated threats in the real world. By creating many fake accounts, Twitter bots can behave differently from genuine accounts. They can like, retweet, or comment on posts that spread misinformation. Moreover, these fake accounts can respond to direct messages (DMs) and deceive unsuspecting individuals into parting with their money. Scrapers, otherwise known as web scrapers or web crawlers, are tools that extract data from websites. Just like bots, scrapers can bring businesses many benefits, such as search engine indexing, but when used maliciously, they become web scraping threats. Scrapers become automated threats due to the following reasons: LinkedIn is one example of a platform that experienced a massive data scrape for malicious purposes. Namely, the hacker used automatically collected data from LinkedIn user profiles to sell it on their forum for a significant sum. Credential stuffing is a cyberattack that involves accessing an account’s login information, such as usernames and passwords. Once the attacker has obtained this information, they can use it to take control of the account or gain access to sensitive data belonging to the account holder. The fact that Norton LifeLock, a global leader in consumer cyber safety, suffered from credential stuffing in 2023 only highlights how dangerous these automated attacks can be. To mitigate the risk of their applications falling victim to automated attacks, developers can implement a series of best practices. These measures aim to enhance security and ensure the integrity of the app’s operations. Key strategies include: Employing CAPTCHA Mechanisms — Integrating CAPTCHA challenges helps distinguish between human users and automated bots, effectively blocking many types of automated attacks. Rate Limiting — Implementing rate limiting on APIs and user actions prevents excessive requests from a single source, which is a common characteristic of automated attacks. Using Advanced Bot Detection Solutions — Leveraging sophisticated bot detection tools can help identify and block more nuanced automated threats that simple CAPTCHAs might miss. Regularly Updating and Patching — Keeping software up-to-date with the latest patches and security updates helps close vulnerabilities that automated scripts might exploit. Monitoring and Analytics — Constantly monitoring traffic and employing analytics can help in quickly identifying unusual patterns indicative of automated attacks. Authentication Mechanisms — Using strong authentication methods, like multi-factor authentication, adds an additional layer of security against automated exploits. Automated threats are merely one aspect of software security. The spectrum of vulnerabilities extends far beyond automated attacks. From code injections and data breaches to compliance issues and insider threats, the array of security challenges is vast and ever-evolving. Addressing these concerns requires a comprehensive and proactive approach to security. Kiuwan provides solutions for developers to ensure code security, vulnerability detection, enforcement of coding guidelines, and management of open-source components. Using Static Application Security Testing (SAST) and software composition analysis (SCA), Kiuwan can help you transform your software development life cycle. Sign up today to get a free trial and take advantage of our robust code security tools or click the link below for a free demo!
<urn:uuid:ed291013-6c19-4d7f-8aec-e7df15d53b39>
CC-MAIN-2024-38
https://www.kiuwan.com/blog/what-do-app-developers-need-to-know-about-automated-threats/
2024-09-07T18:16:43Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650898.24/warc/CC-MAIN-20240907162417-20240907192417-00484.warc.gz
en
0.936605
1,107
3.3125
3
Cloud computing is not a new concept. In fact, cloud computing dates back to the 1950’s, where the concept of shared, centralised compute resources arose, allowing multiple users within a company to access information on separate terminals. Fast forward to the early 2000’s, post the infamous dotcom crash, the reborn technology and data driven world begun using cloud servers to store and access data anywhere, anytime. In short, cloud computing is the delivery of computing services over the internet. The cloud is a service you are using right now and has become mainstream across business. Orchestrating the way in which we operate, the cloud is constantly introducing businesses to innovative, scalable and cost effective solutions for business growth. Underpinning a vast number of services, including your Microsoft Office 365 or Gmail account, through to the cloud services hosting all your data and running your applications, the cloud has become fundamental to all businesses. The cloud is ultimately made up of storage, compute and network. However, you’ll find there are unique ways to deploy cloud resources in public and private cloud, both providing similar benefits and a gateway to cloud computing. What is a Public Cloud? The most common and cost effective way to deploy cloud computing is a public cloud environment, which is accessible by anybody, anywhere, over the internet. Public cloud resources are owned and operated by third party cloud service providers including Microsoft Azure, Amazon Web Services (AWS) and Google Cloud. This means that all hardware, storage and network devices are shared with other organisations across the globe. The public cloud environment provides organisations the ability to rapidly scale and is often used for less sensitive applications that demand a varying amount of resources. It is more suitable for large scale redundancy and is constantly updated by the service provider to provide the best possible services for its users. Therefore, the public cloud is not suitable in facilitating legacy applications and is often used to provide web based email, online office applications and storage. What is a Private Cloud? Private cloud is cloud infrastructure exclusively utilised by one organisation. This is hosted in a service providers (such as 5G Networks) privately owned data centre or physically located on premise within your organisation. In the private cloud, business services and infrastructure remains secure by being maintained on a private network, with components dedicated entirely to your own business. Within the private cloud, you are able to seamlessly customise your cloud environment to meet specific business IT objectives, providing flexible solutions to facilitate any business requirement. More than 78% of Australian businesses are investing in private cloud technology to enable rapid service delivery. Unlike the public cloud, a private cloud environment has the ability to facilitate legacy systems. Acting as a gateway for organisations stepping into the cloud, the private cloud can be customised to support any existing legacy application. Further to this, organisations may have strict security compliances in place around existing legacy applications, which can only be fulfilled within a private cloud environment. Core differences between Public & Private cloud Security is one of the key differentiators between public and private cloud. Reflected in their respective terms, a private cloud will provide organisations the most secure cloud solution where an organisations data will not touch the public internet, significantly reducing the risk of intrusion. The private cloud is often used by government agencies, financial institutions and any other organisation with business critical operations seeking enhanced control over their environment. The public cloud provides a cost effective solution with no need to purchase hardware or software, operating on a pay as you go model. Within a private cloud environment, a business will generally enter a contractual agreement involving a fixed cost arrangement. Businesses often value these agreements as they are able to forecast operational costs over a period of time. A personal approach When utilising a private cloud environment through a service provider, you will benefit from a personable approach and service. This involves being exposed to a service providers expert team, generally comprising of a dedicated account manager, on site and help desk support and project managers to develop the best possible private cloud solution to meet your business requirements. A private cloud provides a customisable cloud solution with the ability to satisfy any business requirement. As you have no control over the cloud resources within the public cloud environment, organisations don’t have the luxury of making changes to suit their business needs within the public environment. Finally, organisations have mixed business requirements that require the use of both Public and Private cloud solutions. This is where Hybrid Cloud steps in. Hybrid Cloud is the combination of both public and private cloud services allowing you to benefit from the best of both worlds. Organisations gain the scalability and computing power of the public cloud for basic and non-sensitive computing tasks, while keeping business critical applications and data stored in a data centre or on-premise, safely behind a company firewall within the private cloud.
<urn:uuid:039cc524-efa7-46cb-9d31-f5634bf0dd5d>
CC-MAIN-2024-38
https://5gnetworks.au/insights/public-vs-private-cloud/
2024-09-11T10:40:31Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651383.5/warc/CC-MAIN-20240911084051-20240911114051-00184.warc.gz
en
0.942532
976
2.765625
3
Another new COVID-19 variant of concern has emerged. A variant is labelled “of concern” when it spreads more easily and is more virulent or resistant to vaccines. How does infrastructure help scientists detect new COVID-19 variants? As a technologist, do you wonder what architecture supports these very important applications? The Africa Centres for Disease Control and Prevention raised the alarm about a new variant. This variant is now named Omicron. Fortunately the African team was able to quickly identify a new variant. More importantly, they immediately shared their scientific observations with the world community of virologists, hoping the new data could be used to prevent another deadly wave of infection. Here are some of the African agency’s observations: - By 11/25, the new variant had been detected in 77 samples. These samples were collected between 11/12 to 11/20 from Gauteng province in South Africa, Botswana and Hong Kong. - The variant displays multiple mutations across the virus genome, including more than 30 in the region which encodes the spike protein responsible for virus entry into host cells. - Many of the identified mutations are not yet well characterized and have not been identified in other currently circulating variants. Don’t you wonder how they are able to figure this out so quickly? How does the COVID-19 virus work? Time for a biology lesson! Viruses hack the machinery of cells, per virologist Rich Condit. I love this presentation. Dr Condit describes biology as we would describe infrastructure. How does technology help scientists find new mutations? I found it helped to listen to Dr. Condit’s virology lesson. It helped me wrap my mind around how an application used to find viral mutations may need to be architected. For instance, Dr. Condit explains the genetic makeup of viruses. In fact, he provides detail about the genetic makeup of the SARs-CoV-2 virus (~ 14:00). Understanding that each strain of the virus has its own genetic markers makes it obvious which technical tools could be used to find new variants. For example, here’s one scenario of how SARs-CoV-2 variants are found. If you take a COVID-19 test and it’s positive for the virus, your sample may be screened with a genotyping panel. That screening panel may include a SNP (Single Nucleotide Polymorphisms) identification pipeline. SNP genotyping identifies genetic variations using sequenced SARs-CoV-2 samples. For example, the genotyping test can detect if the version of the virus that infected you is genetically different than known versions of the virus. Next, if a fast growing cluster of the new variant is found, health officials can be alerted. Finally, scientists will work to understand how the new genetic sequence of the virus works (how virulent it is, if there will be vaccine break-thru, etc.). The scientists get to work Currently, scientists are learning more about the Omicron variant. South African scientists detected the variant. When they noticed a cluster of infections, they informed WHO (the World Health Organization). WHO advised countries to enhance surveillance and genomic sequencing efforts. Additionally, scientists were asked to share any new information they find with GISAID. GISAID is a global database platform for virologists. It was built to enable rapid and open access to epidemic and pandemic virus data. Scientists use it to share data. First, they upload protein or nucleotide sequences in a FASTA file. Next, the server guides them to the correct sequences to compare. I can’t lie, the language is extremely scientific. However, if you’re a scientist who understands the genetic structure of a virus, you’ll probably understand the instructions. Incredibly this data is available in near real time. Of course, in a global pandemic reducing the time it takes to find solutions is a literal life or death matter. Now that we have an idea of how applications used to find new SARs-CoV-2 variants may work we can think about required infrastructure. One example is the CLIMB-BIG-DATA project (the Cloud Infrastructure for Big Data Microbial Bioinformatics). CLIMB is a collaboration of UK universities. The project helps scientists solve big data projects for bioinformatics. It most likely is the largest single system dedicated to Microbial Bioinformatics research in the world. This page lays out the system highlights: - It runs on OpenStack (on IBM supplied hardware) - Over 7500 CPU cores - 78TB RAM - Ceph object storage - 2304 TB cross site replicated storage - 27 Dell R730XD servers per site, each with 16x 4TB HDDs - 500 TB local storage (IBM GPFS) - Brocade networking - 1000 VMs It is easy to forget why we build infrastructure. While we nerd out on how to manage and move data around, our infrastructures support science that is equally difficult to understand. I’m sure there are many other examples of architectures that are supporting work being done to contain this pandemic. Spend some time nerding-out with your users. Maybe what you learn will lead to the next infrastructure innovation! Featured image via Shutterstock.
<urn:uuid:ac0c4e6e-4a72-464d-bc0c-4038debb0924>
CC-MAIN-2024-38
https://24x7itconnection.com/2021/11/30/how-does-infrastructure-support-the-discovery-of-emerging-covid-19-variants/
2024-09-13T18:13:54Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651535.66/warc/CC-MAIN-20240913165920-20240913195920-00884.warc.gz
en
0.921124
1,116
3.65625
4
You may have heard that a serious flaw in Apple’s implementation of SSL was recently discovered. Apple hasn’t offered much detail, but security researchers familiar with the issue have stated that the flaw can open up the possibility of a man-in-the-middle (MITM) attack. This means that unpatched systems are vulnerable to coughing up passwords, credit card numbers, or other sensitive information without the user being aware that their connection is insecure. At the time of this writing, a patch is available for iOS 7, but not for Mac OS X machines. Man-in-the-middle attacks on SSL are not at all new; in fact, I gave a presentation nine years ago at the 2005 RSA Conference, about a malware-enabled man-in-the middle attack one of our customers had discovered, our analysis of how it worked, and what could be done to prevent it. From the limited amount of information that’s been provided so far, it sounds like there’s not much that iOS and Mac users can do about this until a patch is released and installed. But can organizations maintain some assurance that the users communicating with them are not MITM victims? In many cases, the answer may be “yes,” with the help of client-side SSL certificates. To see how it works, let’s first look at how a man-in-the-middle attack works. Most of the time SSL is used, it involves only a certificate on the server, as shown below: To be successful, the man in the middle must be able to convince the client that it’s talking directly to their intended endpoint. This is supposed to be impossible, if SSL is implemented properly (which appears to have been Apple’s problem), and appropriate trust anchors are maintained (which is sometimes easier said than done). If either of these aren’t upheld, man-in-the-middle attacks become possible. But in mutual, or client-side SSL, the server requires that the client provide an authentication certificate as well. In this case, the server gains a level of assurance that it’s talking directly to the client: Now the attackers’ task becomes at least twice as difficult, as they must convince both ends of the connection that they are communicating directly with each other, rather than just one. Obviously, this specific problem with Apple iOS and Mac OS X will eventually be patched and go away. But it will almost certainly not be the last time we hear about a possible man-in-the-middle attack, and client-side certificates can help defend against MITM attacks the next time they come up as well. It’s worth noting, for example, that the server’s demand for a client certificate prevents the attack, even if the client’s machine is still unpatched and otherwise susceptible. Private Key Protection With any digital certificate implementation, it’s all about the private keys; if our man in the middle in the diagram above gets access to the client’s private key, he’s back in business. Diligent private key protection therefore becomes paramount; technologies such as On-Device Key Generation (ODKG) can help, as it limits the places where private keys can be captured or exposed. Any time keys are generated off-device, which frequently happens in mobile or BYOD scenarios, they are subject to compromise anywhere they are created, transmitted, stored, or archived. This is why CSS places so much emphasis on enabling ODKG in scenarios where it is otherwise difficult to achieve – including iOS and Mac devices.
<urn:uuid:368b94e4-53a4-4cdc-be3c-764a9834e0c0>
CC-MAIN-2024-38
https://www.keyfactor.com/de/blog/apples-ssl-flaw-another-man-middle-attack/
2024-09-14T23:44:18Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651601.85/warc/CC-MAIN-20240914225323-20240915015323-00784.warc.gz
en
0.952285
750
2.921875
3
Frost & Sullivan Shares Insights on the Occasion of World Environment Day (June 5th) The ubiquitous influence of plastics in mankind’s modern lifestyle has led to the production of about 300 million tonnes of fossil fuel-based plastics annually, according to the UN. About half of the total plastics produced to date are thrown away after a single use, and about 60% of the waste ends up in landfills or accumulates in the environment. This makes plastics a major contributor to land and marine pollution and loss of biodiversity. The devastating environmental impacts of improper plastic waste management and the depletion of fossil fuel resources are the major driving forces of the formation of stringent regulatory guidelines for the manufacturing, use, and disposal of plastics globally. Five Ways to Mitigate Plastic Pollution Maximizing the potential of plastic waste: The waste management hierarchy of “Reduce, Reuse, and Recycle” aims to extract maximum benefits from the generated waste plastics feedstock and enables end-users to manage the remaining minimum amount of waste. Technology to the Rescue: Emerging technological advancements have the potential to improve plastics waste management, paving way for a safer and cleaner environment. India’s Department of Science & Technology has been actively supporting the development of novel technologies such as recycling polystyrene waste with citrus peel extract and plastics waste-to-weave green technologies to develop fibers that can be used in the textile industry. Creating a Circular Economy: Efficiency of plastic recycling can be increased by integrating mechanical recycling technologies with co-gasification or pyrolysis. This creates a circular economy by closing the loop of the plastics value chain. This improves the ability to effectively utilize mixed plastic waste streams and transform the mixed waste into useful raw materials. France-based Carbios has developed a novel thermo-mechanical based infinite recycling process that converts waste PET plastics into monomers to be used as raw material for manufacturing plastics for the packaging industry. Formation of Strict Guidelines: Efforts are being made by policymakers across the globe, including banning single-use plastics and incentivizing recycling processes, to enhance the strategies employed in the waste management hierarchy. In 2019, the European Parliament approved the law banning single-use plastic items such as straws, cotton buds, and cutlery, by 2021. Adopting Sustainable Alternatives: While adoption of guidelines will help, it is up to individuals and organizations to make changes in their lifestyle and adopt sustainable alternatives to help mitigate plastic pollution. Chemical-giant BASF develops novel bio-polymers based on hemp, kenaf, and bast fibers as a sustainable alternative for fossil fuel-based plastics used in packaging, furniture, and automotive industries. Improving plastics recycling infrastructure has many societal advantages including reducing the burden on the environment and decreasing the dependence on finite fossil fuel resources. Plastic recycling and degradation technologies are also vital to realizing the potential of plastics waste as useful raw material while mitigating pollution. Plastic recycling reduces dependency on virgin raw materials made from fossil fuels and also substantially reduces the carbon footprint associated with plastic waste management. It provides the necessary tools to assign equal or greater value to the end-of-life plastics waste. Technological advancements in plastic recycling and degradation are necessary to overcome the current challenges associated with plastic waste management and to further reduce the negative impacts of plastics on the environment. Technologies Accelerating Plastic Recycling The two major classifications of plastics recycling are mechanical and chemical recycling technologies. Mechanical recycling technologies involve washing, shredding, and melting waste plastics, followed by pelletization and selective compression of all polymers to create monomers. Chemical recycling includes the use of catalytic degradation of High-density Polyethylene (HDPE), Low-density Polyethylene (LDPE), Polyethylene Terephthalate (PET), and polypropylene-based plastics wastes to create monomers. These monomers can be used as raw materials in many applications, including in petrochemical industries and oil refineries. Pyrolysis and co-gasification technologies enable thermo-oxidative de-polymerization that can be used for the effective conversion of mixed plastics waste feedstock (polyethylene, PET, polystyrene, polyamides, and others) into fuels and syngas. The derived fuels have properties similar to crude oil and can be further refined to be used in the transportation industry. The clean syngas sans dioxins and furans can also be used for the manufacture of hydrogen, mixed alcohols, and fertilizers. Technologies Accelerating Plastic Degradation Conventional plastics can take more than 400 years to degrade. It is, therefore, necessary to re-engineer the lifecycle of plastics using advanced biological processes that can ensure that the plastics degrade at a faster rate to avoid waste accumulation in the natural environment. Enzyme-engineered biological processes enable rapid degradation of waste plastics into smaller pieces of micro and nano-plastics that can be easily assimilated by the micro-organisms present in nature. The degradation rate can be enhanced by 20-30 times as the addition of engineered enzymes to plastics aids in accelerating the degradation with 90% conversion efficiency within the first four weeks under a conducive environment, according to Carbios. Work Done by Frost & Sullivan on COVID-19: To know more about Frost & Sullivan’s analysis on COVID-19 from across the globe, visit – https://www.frost.com/insights/covid19/. If you have any queries on the impact of COVID-19 across industries, need more information, or would like to schedule an interview/interaction with our spokespersons, please email Priya George at firstname.lastname@example.org.
<urn:uuid:4fc3d744-f781-4e10-943e-688cfb850852>
CC-MAIN-2024-38
https://www.frost.com/growth-opportunity-news/energy-environment/world-environment-day-2020-the-paramount-need-for-mitigating-plastic-pollution/
2024-09-16T06:29:53Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00684.warc.gz
en
0.906687
1,181
3.546875
4
Internet security experts say that misconfigured digital subscriber line and cable modems are worsening a well-known problem with the Internet’s DNS (domain name system), making it easier for hackers to launch distributed denial-of-service (DDoS) attacks against their victims. In fact, the percentage of DNS systems on the Internet that are configured this way has jumped from around 50 per cent in 2007, to nearly 80 per cent this year, according to Liu. Though he hasn’t seen the Infoblox data, Georgia Tech Researcher David Dagon agreed that open recursive systems are on the rise, in part because of “the increase in home network appliances that allow multiple computers on the Internet.” “Almost all ISPs distribute a home DSL/cable device,” he said in an e-mail interview. “Many of the devices have built-in DNS servers. These can sometimes ship in ‘open by default’ states.” Because modems configured as open recursive servers will answer DNS queries from anyone on the Internet, they can be used in what’s known as a DNS amplification attack. In this attack, hackers send spoofed DNS query messages to the recursive server, tricking it into replying to a victim’s computer. If the bad guys know what they’re doing, they can send a small 50 byte message to a system that will respond by sending the victim as much as 4 kilobytes of data. By barraging several DNS servers with these spoofed queries, attackers can overwhelm their victims and effectively knock them offline. DNS experts have known about the open recursive configuration problem for years, so it’s surprising that the numbers are jumping up. However, according to Dagon, a more important issue is the fact that many of these devices do not include patches for a widely publicizedDNS flaw discovered by researcher Dan Kaminsky last year. That flaw could be used to trick the owners of these devices into using Internet servers controlled by hackers without ever realizing that they’ve been duped. Infoblox estimates that 10 percent of the open recursive servers on the Internet have not been patched. The Infoblox survey was conducted by The Measurement Factory, which gets its data by scanning about 5 percent of the IP addresses on the Internet. The data will be posted here in the next few days. According to Measurement Factory President Duane Wessels, DNS amplification attacks do occur, but they’re not the most common form of DDoS attack. “Those of us that track these and are aware of it tend to be a little bit surprised that we don’t see more attacks that use open resolvers,” he said. “It’s kind of a puzzle.” Wessels believes that the move toward the next-generation IPv6 standard may be inadvertently contributing to the problem. Some of the modems are configured to use DNS server software called Trick or Tread Daemon (TOTd) — which converts addresses between IPv4 and IPv6 formats. Often this software is configured as an open resolver, Wessels said.
<urn:uuid:a316805f-8fb3-45e0-9c1c-5263b7525ed0>
CC-MAIN-2024-38
https://www.itworldcanada.com/article/dns-problem-linked-to-ddos-attacks-gets-worse/40179
2024-09-16T06:40:54Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651676.3/warc/CC-MAIN-20240916044225-20240916074225-00684.warc.gz
en
0.952362
659
2.625
3
In today’s digital age, cyber threats come in various forms, with vishing being a prevalent and dangerous one. What exactly is vishing, and how can individuals and businesses protect themselves from falling victim to this deceptive tactic? What is Vishing? Vishing, short for voice phishing, is a social engineering attack that uses voice communication to trick individuals into revealing sensitive information or performing certain actions. Cybercriminals often pretend to be legitimate entities, such as financial institutions or government agencies, to deceive victims into providing personal details like passwords, credit card numbers, or account information over the phone. Combatting Vishing Attacks Protecting yourself or your organisation from vishing attacks requires awareness, vigilance, and a proactive approach to cybersecurity. Here are some tips on how to combat vishing: - Be Sceptical of Unsolicited Calls: Never disclose sensitive information or make financial transactions based on unsolicited calls. Legitimate organisations will not ask for personal details over the phone without proper verification. - Verify Caller Identity: If you receive a call from someone claiming to be from a company or institution, hang up and independently verify their identity by contacting the official phone number listed on their website or documentation. - Educate Employees: Ensure that employees are trained to recognise vishing attempts and understand the importance of verifying caller identities before sharing information or taking action. - Implement Two-Factor Authentication: Utilise two-factor authentication for sensitive accounts and systems to add an extra layer of security, making it more difficult for attackers to access your data even if they obtain your credentials. - Report Suspicious Calls: If you suspect a vishing attempt, report the incident to the appropriate authorities or contact your IT security team for further investigation. How Network Fish Can Help At Network Fish, we specialise in cybersecurity solutions that help organisations defend against a wide range of cyber threats, including vishing attacks. Our team of experts can assess your current security measures, identify potential vulnerabilities, and implement advanced security controls to safeguard your business from social engineering tactics like vishing. By partnering with Network Fish, you can benefit from: - Tailored Security Solutions: We provide customised security solutions based on your organisation’s specific needs and industry requirements. - Incident Response Expertise: In the event of a vishing attack or security incident, our team offers prompt incident response services to mitigate the impact and restore normal operations. - Continuous Monitoring and Support: We offer ongoing monitoring and support to proactively detect and address potential security issues before they escalate into major threats. Reach Out to Network Fish for Assistance Don’t wait until a vishing attack compromises your organisation’s security and reputation. Contact Network Fish today to learn more about our cybersecurity services and how we can help protect your business from social engineering threats like vishing. Contact us now to schedule a consultation with our cybersecurity experts and strengthen your defenses against vishing attacks. Stay one step ahead of cybercriminals with Network Fish by your side.
<urn:uuid:dca4e6e6-8b3c-4acc-ad2a-59bf6bec2ac3>
CC-MAIN-2024-38
https://networkfish.com/it-security/unveiling-vishing-understanding-the-threat-and-how-to-combat-it/
2024-09-17T11:28:37Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651773.64/warc/CC-MAIN-20240917104423-20240917134423-00584.warc.gz
en
0.921712
616
2.90625
3
Recently, Google proclaimed that it had achieved quantum supremacy with its “Sycamore” quantum computer that can solve complex algorithms unsolvable by any other computer today. This milestone raises fundamental questions about how quantum computing can be used and how it will affect initiatives in the digital era. Every day, humans create more than 2.5 exabytes of data, and that number continues to grow, especially with the rise of the internet of things (IoT) and 5G capabilities. Machine learning (ML) and artificial intelligence (AI) are some of the ways to help manage and analyze data for competitive advantage, but continued innovation and the desire for meaningful insights may make data increasingly complex for organizations to collect and analyze. As classical binary computing reaches its performance limits, quantum computing is becoming one of the fastest-growing digital trends and is predicted to be the solution for the future’s big data challenges. Though quantum computing is still just on the horizon, the U.S. plans to invest more than $1.2 billion toward quantum information over the next 10 years in a race to build the world’s best quantum technology. What Is Quantum Computing? Quantum computing combines quantum physics, computer science and the theory of information – and most experts agree it has the potential to impact the future of digital business and security. How does it differ from classic computing? Traditional computing uses a base-two numerical system that follows set operations and processes and communicates data using bits. Every piece of digital information is stored as a bit in the form of either a one or a zero. A series of bits put together is called a binary code. For example, the letter “A” in classic computing is stored in binary as 01000001. The challenge of classic computing is that it can only run one calculation at a time. So, when there is a large data set to compute, this negatively impacts and decreases its computational power. Quantum computing is fundamentally different in that it uses what is called quantum bits (qubits), which expand the binary limits by following quantum logic represented as a one or zero of digital data and the logic of superposition in which a qubit is represented not just by one state but by both a one and a zero at the same time. It exists in an unrecognized combination, and once the data is called, it is compiled into one of the definite states as a one or zero. Superposition is what makes qubits more impressive because it decreases the number of operations needed to solve complex problems by being able to run calculations simultaneously at higher speeds and with less power consumption. Impact on AI and ML Analyzing massive amounts of data is very complex. In fact, it can be nearly impossible for humans to sift through a large volume of data and find correlated information that has actionable value. Organizations create so much data that large amounts of data go unused. Many are missing out on opportunities that could help gain competitive advantages. It is with these large data sets that ML can be quite helpful. ML requires a great deal of historical data and constantly needs to be fed with new information, so it can learn how the data changes and identify trends over time. However, as the volume of data increases, the complexity of computation also increases, along with the time needed to analyze, calculate, identify, interpret and provide any relevant output. Machines do not know how to tell a story with data, they simply provide the raw material. ML algorithms today are limited by the computational power of classical computers. Quantum computing is capable of administering large data sets at much faster speeds and can supply data to AI technologies to analyze data at a more granular level to identify patterns and anomalies. Quantum computing also can help integrate data by running comparisons between schemas to quickly analyze and understand the relationship between two counterparts. To give a bit of perspective, Google’s Sycamore is reported to have solved a problem in 200 seconds that would have taken today’s fastest supercomputer 10,000 years to solve. This opens new possibilities for the future of big data and analytics. As the world changes its approach to IT and information architecture, its approach to computing and big data must change as well. Organizations are facing different problems today, which means they need different solutions. ISG stays up to date with the latest digital trends in IT, so we can help enterprises make decisions about big data and analytics along the digital transformation journey. Contact us to discuss how we can help solve your current big data and analytics challenges and put your organization on a path toward a digital future.
<urn:uuid:52d80efa-1826-4128-a19f-a5d9a0fc542b>
CC-MAIN-2024-38
https://isg-one.com/articles/quantum-computing-and-the-future-of-big-data
2024-09-07T21:18:16Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00584.warc.gz
en
0.944623
933
3.8125
4
In celebration of National Engineers Week, STEM Lab now offers more free, online activities to inspire future engineers. Germantown, MD, February 19, 2019 – In support of National Engineers Week February 17 - 23, 2019, America’s No. 1 satellite Internet service HughesNet® and National 4-H Council announce the availability of two new hands-on projects as part of their popular 4-H STEM Lab. STEM Lab is a free, online resource designed to spark kids’ interest in science, technology, engineering and math (STEM). The U.S. Bureau of Labor Statistics projects that by 2026 the economy will require nearly 140,000 new engineering jobs to be filled in the U.S. To help meet the current and projected demand for graduates in STEM careers, HughesNet and 4-H are working together to inspire and empower the next generation of leaders through STEM Lab’s hands-on, fun projects for youth ages 4 to 16. Since its launch in May of 2018, STEM Lab has generated more than 60,000 page views and thousands of activity downloads. The new projects, “Building Bridges” and “Parachute Away,” teach principals of physics, civil and mechanical engineering and are designed to help youth become excited about STEM and engineering careers. “Building Bridges,” teaches children about bridge design, construction and the mechanics at work as they build their own bridges. “Parachute Away,” brings physics to life, showing kids how air resistance affects the speed of falling objects and how to construct a safe delivery system for fragile cargo. As with prior activities, they come with easy-to-follow instructions; explanations of foundational STEM concepts; discussion questions and supply lists (usually basic household items). “Through our sponsorship of 4-H, we’re making STEM education accessible to all kids, no matter where they live,” said Peter Gulla, senior vice president of marketing at Hughes. “Hughes employs some of the world’s most talented engineers, creating advanced satellite communications and networking technologies. We are proud to celebrate National Engineers Week with the launch of two new projects, helping to inspire kids to become the next generation of engineers and scientists who will shape our future.” “4-H has been inspiring future engineers for over 100 years. We’re excited to celebrate National Engineers Week by helping more youth learn about the wonders of engineering through the 4-H STEM Lab,” said Jennifer Sirangelo, president and CEO, National 4-H Council. “4-H youth, compared to their peers, are two times more likely to pursue STEM careers. Through STEM Lab we are ensuring that more kids have access to the hands-on STEM learning they need to explore future career pathways in engineering.” To learn more about STEM Lab and to dive into STEM activities, visit 4-H.org/ Founded by the National Society of Professional Engineers in 1951, National Engineers Week is dedicated to raising awareness and understanding about engineering careers among the nation’s future workforce.
<urn:uuid:e20a98c7-64be-440b-82c6-4b06a42f63eb>
CC-MAIN-2024-38
https://www.hughes.com/resources/press-releases/hughesnet-expands-4-h-stem-lab-projects-kids
2024-09-07T20:35:52Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650920.0/warc/CC-MAIN-20240907193650-20240907223650-00584.warc.gz
en
0.950907
637
2.703125
3
Organizations of nearly every size and industry are rapidly going through digital transformation. Although digitization is bringing along many operational efficiencies, it also puts organizations at greater risks of cyber disruptions. What is Cyber Disruption? In plain words, we can define cyber disruption as events over the cyber space that limit or eliminate the ability of an affected organization to maintain operations over its digital infrastructure. This can give rise to a host of other problems as well. Sources of Cyber Disruption When we contemplate the potential sources of cyber disruptions, the possible forces at play can be many. However, it is not possible for every organization to effectively evolve and implement a comprehensive solution for each potential source of disruption. Still, some of the most common sources of cyber disruptions include natural disasters, malware, ransomware and denial of service (DoS) attacks etc. Each entity will have to attach a relative weightage to these threats and any other ones as well. - 17 Best Practices to Attain Secure Remote Work Environments - Top 11 Guidelines for Cloud Security in 2020 - Jan 2020 Roundup of Data and Cyber Security Breaches What is Cyber Reliance? The term cyber resilience can be defined as the efficiency and effectiveness with which an organization can tackle cyber disruptions. Remember, as the digital space is constantly evolving, so does the need for comprehensive cyber resilience solutions. Cyber Resilience and Cloud Computing Cloud computing is a reliable and versatile platform for deploying and managing a cyber resilience solution. Here are some reasons why. Cloud Service Providers (CSP) like dinCloud have a very reliable and robust hardware infrastructure in the form of multiple data centers. This is an excellent support mechanism in scenarios where one data center becomes inaccessible due to some disruption. When you move your data to the cloud, it is converged in the CSP’s highly secure data centers. These data centers are secured by maintaining both physical and cyber security protocols such as firewalls, intrusion detection, encryption and more. When you build your cyber resilience over cloud infrastructure, your data is not scattered across multiple end point devices, which are nearly impossible to fully secure and patch against cyber threats. CSPs like dinCloud employ some of the most stringent and reliable protocols to guard organizational data from malware. Most small to medium organizations may lack both the resources and skill set to protect data from malware. Protection from Ransomware Ransomware is a huge menace in the cyber space and most entities that fall prey to this attack end up paying the ransom. A resilience model that is built around the cloud protects you from ransomware by maintaining multiple backups of your valuable data. Now, you will no longer have to end up paying ransom to the perpetrator of such an attack. Your cloud provider like dinCloud will immediately restore your data through one of its multiple backups housed safely in its physically dispersed data centers. When you build your cyber resilience around the cloud, you are gaining the additional edge of a reliable Business Continuity (BC) mechanism. In case of natural disruptions or otherwise, cloud providers have multiple data centers to support your critical operations. What this means for your organization is that you are also acquiring a reliable posture for disaster recovery (DR) and business continuity (BC), all wrapped up in the services of a reputed cloud solutions provider like dinCloud. Reduced Down Times In present times, the financial impact of even a small downtime can translate into thousands of dollars for a small to medium sized business. The reliable hardware of cloud providers, coupled with scheduled maintenances ensure zero to minimal IT downtime. In 2014, Gartner Estimated the average cost of a single minute of IT downtime to around $5,600. In 2016, a similar study by Ponemon Institute estimated this impact to be around $9,000, further raising the stakes of avoiding IT downtimes at all costs. The architecture of cloud providers (CSP) like dinCloud is designed in a way that a downtime in one of their data centers is compensated by some other data center that has idle capacity, making the impact of downtimes at the cloud user level to zero. Why Choose dinCloud for Cyber Resilience? dinCloud is a highly innovative and reliable Cloud Service Provider (CSP) that supports its thousands of cloud users through multiple data centers. Our data centers are certified for some of the cloud industry’s highest standards of security and regulatory compliance. We offer DR and BC built into our wide array of cloud hosted solutions that include virtual desktops, servers and database management solutions. We offer you a robust cyber resilience at an industry beating flat rate pricing model. Contact Us to help us understand your cyber resilience needs and priorities. We will design a solution that will deliver unmatched resilience for all contingencies.
<urn:uuid:b3c12896-ea72-4ea5-8d1f-5dcf49a98277>
CC-MAIN-2024-38
https://www.dincloud.com/blog/how-cloud-service-providers-enhance-cyber-resilience
2024-09-10T08:03:49Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651224.59/warc/CC-MAIN-20240910061537-20240910091537-00384.warc.gz
en
0.939694
977
2.640625
3
Artificial intelligence (AI) is a branch of computer science that focuses on developing systems able to perform tasks normally requiring human intelligence. Although AI research was first initiated in the 1950s, it has recently emerged as a powerful tool for software development. From gaming to medicine and transportation, AI is changing the way we live our lives and work. What is AI? AI is a broad term that refers to all types of technology that can mimic human intelligence. AI is not one thing, but rather many different approaches to solving problems. For example, you might use an algorithm-based approach like machine learning or deep learning; you could also build a system using rules and logic. AI can be used in software development for many purposes: - To make better decisions: If your company has an app or website where users can book appointments or order food delivery, then chances are good that it uses some sort of artificial intelligence behind the scenes. This could include things like customer segmentation based on past purchases or behavioral data points (such as whether someone spends more time browsing than ordering), which helps companies decide how much inventory to keep on hand at any given time. - To automate processes: Many businesses rely heavily on automation because it allows them to scale up their operations without hiring additional employees–a great example here would be Amazon’s warehouses, where machines do most of the heavy lifting! Why is AI so powerful? Artificial intelligence is a powerful tool for solving complex problems. AI can make decisions based on data in ways that humans cannot. For example, an AI system could be trained on your customer database and then use its knowledge of these customers’ needs to predict what products they might want next. For many businesses, this capability opens up new opportunities for innovation. For example, an insurance company could use machine learning algorithms to develop new policies tailored to individual clients’ needs (rather than having one-size-fits-all offerings). This allows them to make better predictions about risk, and thus charge higher premiums for those deemed high risk, while offering lower rates for low-risk individuals or groups. It also gives consumers more flexibility over their choices when buying coverage; if they want something different from what’s offered by default options, then they’re able to create customized plans without having any particular expertise in insurance policy design themselves! How can AI be used to build B2B solutions? AI can be used to build B2B solutions in many different ways. A few examples of this include: - Using various techniques and tools, such as artificial neural networks (ANN), machine learning and deep learning algorithms, natural language processing (NLP), computer vision and speech recognition technology. - Building intelligent agents that are capable of performing tasks on behalf of the user or understanding their needs through natural language interaction with them using question-and-answer dialogues with AI assistants like Siri or Alexa. What are the fundamental building blocks of AI-powered software solutions? What are the fundamental building blocks of AI-powered software solutions? This is a question that every software developer should ask themselves. If you’re not familiar with machine learning, neural networks and deep learning already, it’s time to catch up! These concepts form the basis of most modern AI applications that companies can use today. Neural networks are essentially sophisticated mathematical models designed to mimic how our brains work – they “learn” through experience by correcting their own mistakes over time until they reach an optimal state (i.e., they get better at solving problems). Neural networks have been around since the 1960s, but only recently have we been able to harness enough computing power from GPUs (graphics processing units) and other hardware components needed for these types of complex calculations in order for them to become commercially viable for businesses looking at using artificial intelligence technology as part of their product offerings or services offerings.” How do you implement your AI-powered solution? The process for building an AI-powered solution is similar to building any other software product. You’ll need to: - Design the solution. This includes defining your use case, gathering requirements, and building prototypes. - Implement your solution in code that uses the right technology stack (e.g., Python). - Test your code before releasing it into production by running unit tests or performing manual testing on staging environments with real data sets (and maybe even some fake ones). - Optimize as needed based on feedback from users or performance metrics Which programming language should you use to develop your intelligent solution? If you’re interested in developing an intelligent solution, the first question you will want to ask yourself is which programming language should I use? Python has been the most popular AI language for the last few years. It’s easy to learn and has a large community. There are many reasons why Python is so popular: Google, Amazon and Netflix all use Python; it’s great for data science and machine learning; it can be used as a web development language (for example on websites like Reddit). Importance of AI for SEO This is a great time to be an SEO professional. As more and more people turn to search engines, like Google and Bing, for information about products and services, it’s become increasingly important for businesses to hire a reliable As a B2B SaaS SEO services provider and start learning some ways AI can help you optimize your business. AI can help you understand your audience better. You can use AI-powered tools such as machine learning algorithms that analyze text data (such as search queries) in order to understand user intent behind different keywords or phrases used by visitors on your website or app. This analysis helps you identify topics related with each other so that you can create more relevant content around those topics using targeted keywords that are likely to attract visitors with similar interests, as well as improve the relevancy score of existing pages containing this information by adding additional relevant links within those documents themselves (e.g., internal linking). Businesses can apply artificial intelligence to many different problems and solve them more efficiently Artificial intelligence (AI) is a broad field that covers many different kinds of software development problems. AI can be used to solve any problem that requires a solution, including: - Data analysis and machine learning - Fraud detection and prevention - Predictive analytics - Automated decision-making As you can see, AI has applications in all industries and fields, from healthcare to finance, from manufacturing to retail. It’s not just for big companies, either; even small businesses can benefit from using intelligent software solutions to increase efficiency and productivity within their organizations. AI is a powerful tool that can help you build better B2B software. It will not replace your developers, but it will help them do their job faster and better. AI helps us understand our customers better, so we can deliver products that meet their needs more efficiently than ever before. The future of AI in software development looks bright, and we’re excited to see where it takes us next!
<urn:uuid:d3b3db4a-6d50-4816-9d7d-aa619b7705b8>
CC-MAIN-2024-38
https://www.kovair.com/blog/unlocking-the-power-of-artificial-intelligence-in-software-development/?related_post_from=8867
2024-09-11T12:14:44Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651387.63/warc/CC-MAIN-20240911120037-20240911150037-00284.warc.gz
en
0.952728
1,450
3.703125
4
The term "Internet of Things (IoT)" has already confused enough people, making it difficult to see the bigger picture of IoT architecture. Understanding the broader view of how a real time monitoring system that collects data from sensors works will help demystify various IoT terminology. Here's a look at how sensor-driven architecture can modernize and streamline a business. Defining IoT Architecture There are various components of IoT architecture that enable a business to collect valuable data in real time. Data insights captured from IoT sensors alone do not tell the full story of what IoT is about. In order for a business to justify an IoT investment, the company needs to understand the various layers and deployment stages of IoT architecture. Mixing them up can lead to expensive mistakes. The architecture layers for IoT are the sensing, network, processing and application layers. First, a physical layer within the environment must be established through connected devices and objects. The Internet of Things would not exist without smart technology such as smartphones or technology that converts analog information into digital data. Layers of IoT Architecture Part of the confusion over IoT is that it's not just one technology or process. It's a collection of technologies and processes categorized in layers. Each software architecture layer simply encompasses different types of devices, apps and methods that can be grouped together to serve a certain purpose. Here's a deeper look at what these layers mean: - Perception/Sensing Layer: Wireless sensors and actuators make up the perception layer. While a sensor collects and delivers data, an actuator measures changes recorded by the sensors. These sensors and actuators can be connected through Local Area Networks (LANs) and Personal Area Networks (PANs). - Network Layer: This layer encompasses the platform that facilitates data transmission through various connected devices. It allows devices to communicate with other servers and devices. The network layer allows you to monitor how data moves throughout a software application. It includes Data Acquiring Systems (DAS) and Internet/Network gateways. A DAS is responsible for the functions of data aggregation and data conversion from analog to digital. - Processing Layer: The analytics gathered from an IoT ecosystem are made possible due to data processing. IoT data is typically pre-processed and stored before it is delivered to a data center. Edge computing is becoming an important part of the processing layer since it involves IoT devices with data processing capabilities. - Application Layer: When users interact with IoT software, they are operating in the application layer. This layer includes smart apps that are used to access IoT data. Each of these software layers works together to make IoT technology possible. In order to call a system IoT-based, it needs devices that run on a network, where data processing is possible, then data can be accessed through smart apps. When you put all the IoT layers together, it creates an architectural system that reflects how successful communication works. Sensors act as the senders of information, while data is the message. The data moves through channels such as edge networks where it's processed. End users receive the data through their smart devices. The main noise in the system that can disrupt data transmission is the latency from sending too much data to the cloud. Planning for the IoT Architecture Model Before investing in an IoT architecture, it's best to get a broad view of the different stages involved with deploying it. Understanding how IoT architecture is broken down into technology categories will help clarify how the components interact with each other. Here are the sequential stages of deploying sensor-based architecture: Connected Objects and Devices All the devices that connect with an IoT network are considered part of the first layer, which must be set up first before any other deployment. Connected objects can be anything generating information that can be converted into digital form via electronic technology. The sensors and actuators in the perception layer are among the first to be deployed when setting up an IoT system. There are various types of IoT sensors now that can be used for different applications. Farmers use IoT throughout their properties to monitor environmental elements such as the quality of soil, air and water. The most advanced agricultural companies use drone equipment with sensors to take aerial photos of crop fields. Many smart homes contain electronic devices equipped with IoT sensors, such as smart thermometers. Some sensors can detect proximity, such as when a pedestrian walks near home. Proximity sensors are used by physical retail stores to detect when customers come near the door or a certain section of the store layout. Other sensors can measure pressure, chemicals and gases. Smoke sensors are found in modern smoke and fire alarm systems. While these sensors have existed for decades, they are much more reliable and effective in an IoT infrastructure. IoT-based smoke detectors have the capability of sending immediate alerts to homeowners on their smartphones when fire, gas, or smoke is detected on their property. Once IoT objects and devices are determined, the next stage in deploying this powerful interactive system is to set up an internet gateway. A gateway is a device that converts analog data to digital data to allow data processing over the internet. Today's internet gateways collect raw data for IoT sensors before it is sent to the cloud or a nearby server. After the internet gateway converts analog information to digital form, it aggregates and protects the data. The next features to deploy in an IoT system are analytics and security. Analytics software presents gathered data from various sources in a prioritized way to reveal system insights. Ultimately, analytics is the end product of an IoT system that allows business leaders to make quick improvements. Data security is no longer an option, as it must be viewed as a requirement. Many early IoT systems lack the sufficient protection of a zero trust environment in which all devices connected to a network must be protected. Not only do internet gateways help protect digital assets, but they can also improve a company's performance and efficiency. The third stage in the deployment of IoT architecture deals with the pre-processing phase and methods that enhance analytics. Edge computing is the use of network devices that have data processing capabilities, so that data does not need to be transmitted long distances. Sending big data to the cloud creates network congestion that can lead to latency and higher costs for bandwidth expansion. So far edge computing has proven to be more efficient than cloud computing for the processing of IoT data. Edge IT systems often incorporate machine learning technology in which data analysis of patterns can be accelerated to reach quick conclusions and solutions. By using self-processing IoT devices, an edge IT system reduces strain on your core IT infrastructure. Another advantage of edge computing is that it allows for data visualization techniques to deliver compelling system insights. In other words, it's a system that can that easily facilitates graphic displays, which are powerful tools that accelerate learning curves. Studies show people can grasp information faster from viewing an infographic or a video than from reading several paragraphs of text. Many cloud providers operate from large data centers where processed IoT data ends up. In the cloud, the data can be evaluated further for final analysis and reporting. Through data centers and cloud service providers company data can be easily managed and made available to customers. The cloud is often used as a place where customer data can be transferred through applications. Private clouds offer a much safer and more secure haven for confidential data than public clouds. But public clouds often allow for free storage, which is why many companies use them. Choosing a hybrid approach is sustainable because private clouds make critical data safer while public clouds help cut IT costs for storing less important big data. The combination of the cloud and IoT is a powerful solution for any business that generates a wealth of data. As long as the bulk of remote IoT data doesn't have to travel to the cloud, the cloud can be utilized as a hub for an entire business. Making sure you keep big data under control from IoT sensors is an essential piece of the sustainability puzzle. People and organizations can connect with each other more swiftly and seamlessly through IoT architecture. This system of data-collecting sensors helps businesses detect system problems in real time while offering immediate solutions through the use of machine learning software. A lot of IT companies are already emerging around the globe that offer services to design your business’ systems that are aligned with IoT architecture. So, this shouldn’t be a problem once you decide of doing so.
<urn:uuid:3a529717-44de-4b7e-b086-24588740bb4a>
CC-MAIN-2024-38
https://iotmktg.com/what-iot-architecture-and-its-impact-on-businesses/
2024-09-12T19:57:36Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651491.39/warc/CC-MAIN-20240912174615-20240912204615-00184.warc.gz
en
0.932974
1,690
3.1875
3
Have you ever wondered what forms the basis of May I know you page that facebook directed you when you were busy scrolling through or how your online signatures are verified ? Remember the crime documentaries where graphologist analyzes murder’s handwriting for finding the real culprit. Long gone are the days when all these nitty gritty tasks were in human hands, now artificial intelligence has taken over these assessments. In the modern era neural networks are assisting humans to survive the new age transitions in education, financial, aerospace and automotive sectors. But before knowing how they are giving different sectors a push, it is first important to understand the basic concept of neural networks and deep learning. Neural networks are a part of deep learning, which comes under the comprehensive term, artificial intelligence. Neural networks are a set of algorithms that are modelled after the human brain. These networks are also known as artificial neural networks (ANN). Sensory neurons, motor neurons and interneurons form the human brain. Artificial neurons, form the replica of the human brain (i.e. a neural network). Artificial Neural Network (ANN) Artificial Neural Network (ANN) is a collection of connected units (nodes). These connected units are known as artificial neurons. These units closely resemble the original neurons of a human brain. Every node is built with a set of inputs, weights, and a bias value. Weights of the neural network are held within the hidden layers. Weights and biases are learning parameters of machine learning models, they are modified for training the neural networks. Architecture of an artificial neural network Applications of Neural Networks Neural Networks are regulating some key sectors including finance, healthcare, and automotive. As these artificial neurons function in a way similar to the human brain. They can be used for image recognition, character recognition and stock market predictions. Let’s understand the diverse applications of neural networks 1. Facial Recognition Facial Recognition Systems are serving as robust systems of surveillance. Recognition Systems matches the human face and compares it with the digital images. They are used in offices for selective entries. The systems thus authenticate a human face and match it up with the list of IDs that are present in its database. (Must Check: Facial Recognition Work in Deep Learning?) Convolutional Neural Networks (CNN) are used for facial recognition and image processing. Large number of pictures are fed into the database for training a neural network. The collected images are further processed for training. Sampling layers in CNN are used for proper evaluations. Models are optimized for accurate recognition results. (Related Blog: How does Basic Convolution Work for Image Processing?) 2. Stock Market Prediction Investments are subject to market risks. It is nearly impossible to predict the upcoming changes in the highly volatile stock market. The forever changing bullish and bearish phases were unpredictable before the advent of neural networks. But well what changed it all? Neural Networks of course… To make a successful stock prediction in real time a Multilayer Perceptron MLP (class of feedforward artificial intelligence algorithm) is employed. MLP comprises multiple layers of nodes, each of these layers is fully connected to the succeeding nodes. Stock’s past performances, annual returns, and non profit ratios are considered for building the MLP model. Check out this video to know how the LTSM model is built for making predictions in the stock market. 3. Social Media No matter how cliche it may sound, social media has altered the normal boring course of life. Artificial Neural Networks are used to study the behaviours of social media users. Data shared everyday via virtual conversations is tacked up and analyzed for competitive analysis. Neural networks duplicate the behaviours of social media users. Post analysis of individuals' behaviours via social media networks the data can be linked to people’s spending habits. Multilayer Perceptron ANN is used to mine data from social media applications. MLP forecasts social media trends, it uses different training methods like Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and Mean Squared Error (MSE). MLP takes into consideration several factors like user’s favourite instagram pages, bookmarked choices etc. These factors are considered as inputs for training the MLP model. In the ever changing dynamics of social media applications, artificial neural networks can definitely work as the best fit model for user data analysis. (Related Blog: Detection of Fake and False News using CNNl) Aerospace Engineering is an expansive term that covers developments in spacecraft and aircraft. Fault diagnosis, high performance auto piloting, securing the aircraft control systems, and modeling key dynamic simulations are some of the key areas that neural networks have taken over. Time delay Neural networks can be employed for modelling non linear time dynamic systems. Time Delay Neural Networks are used for position independent feature recognition. The algorithm thus built based on time delay neural networks can recognize patterns. (Recognizing patterns are automatically built by neural networks by copying the original data from feature units). Other than this TNN are also used to provide stronger dynamics to the NN models. As passenger safety is of utmost importance inside an aircraft, algorithms built using the neural network systems ensures the accuracy in the autopilot system. As most of the autopilot functions are automated, it is important to ensure a way that maximizes the security. Applications of neural networks Defence is the backbone of every country. Every country’s state in the international domain is assessed by its military operations. Neural Networks also shape the defence operations of technologically advanced countries. The United States of America, Britain, and Japan are some countries that use artificial neural networks for developing an active defence strategy. Neural networks are used in logistics, armed attack analysis, and for object location. They are also used in air patrols, maritime patrol, and for controlling automated drones. The defence sector is getting the much needed kick of artificial intelligence to scale up its technologies. Convolutional Neural Networks(CNN), are employed for determining the presence of underwater mines. Underwater mines are the underpass that serve as an illegal commute route between two countries. Unmanned Airborne Vehicle (UAV), and Unmanned Undersea Vehicle (UUV) these autonomous sea vehicles use convolutional neural networks for the image processing. Convolutional layers form the basis of Convolutional Neural Networks. These layers use different filters for differentiating between images. Layers also have bigger filters that filter channels for image extraction. The age old saying goes like “Health is Wealth”. Modern day individuals are leveraging the advantages of technology in the healthcare sector. Convolutional Neural Networks are actively employed in the healthcare industry for X ray detection, CT Scan and ultrasound. As CNN is used in image processing, the medical imaging data retrieved from aforementioned tests is analyzed and assessed based on neural network models. Recurrent Neural Network (RNN) is also being employed for the development of voice recognition systems. (Must Check: Learning Recurrent Neural Network and applications) Voice recognition systems are used these days to keep track of the patient’s data. Researchers are also employing Generative Neural Networks for drug discovery. Matching different categories of drugs is a hefty task, but generative neural networks have broken down the hefty task of drug discovery. They can be used for combining different elements which forms the basis of drug discovery. 7. Signature Verification and Handwriting Analysis Signature Verification , as the self explanatory term goes, is used for verifying an individual’s signature. Banks, and other financial institutions use signature verification to cross check the identity of an individual. Usually a signature verification software is used to examine the signatures. As cases of forgery are pretty common in financial institutions, signature verification is an important factor that seeks to closely examine the authenticity of signed documents. Artificial Neural Networks are used for verifying the signatures. ANN are trained to recognize the difference between real and forged signatures. ANNs can be used for the verification of both offline and online signatures. For training an ANN model, varied datasets are fed in the database. The data thus fed help the ANN model to differentiate. ANN model employs image processing for extraction of features. (Related Blog: Hand Gesture Classification using Deep Learning with Keras) Handwriting analysis plays an integral role in forensics. The analysis is further used to evaluate the variations in two handwritten documents. The process of spilling words on a blank sheet is also used for behavioural analysis. Convolutional Neural Networks (CNN) are used for handwriting analysis and handwriting verification. 8. Weather Forecasting The forecasts done by the meteorological department were never accurate before artificial intelligence came into force. Weather Forecasting is primarily undertaken to anticipate the upcoming weather conditions beforehand. In the modern era, weather forecasts are even used to predict the possibilities of natural disasters. Multilayer Perceptron (MLP), Convolutional Neural Network (CNN) and Recurrent Neural Networks (RNN) are used for weather forecasting. Traditional ANN multilayer models can also be used to predict climatic conditions 15 days in advance. A combination of different types of neural network architecture can be used to predict air temperatures. (Must Check: Weather Forecasting Using Big Data Analytics) Various inputs like air temperature, relative humidity, wind speed and solar radiations were considered for training neural network based models. Combination models (MLP+CNN), (CNN+RNN) usually works better in the case of weather forecasting. Neural Networks have a myriad of applications, from facial recognition to weather forecasting the interconnected layers (human brain’s replica), can do a lot of things with some simple inputs. ANN algorithms have simplified the assessment and modified the traditional algorithms. With humanoid robots like Grace on its way the world can expect some sci- fi movies turning into reality pretty soon!
<urn:uuid:9bd0b5c3-756f-47b0-98c2-bbdfc09f8471>
CC-MAIN-2024-38
https://www.analyticssteps.com/blogs/8-applications-neural-networks
2024-09-12T19:37:57Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651491.39/warc/CC-MAIN-20240912174615-20240912204615-00184.warc.gz
en
0.92541
2,041
3.578125
4
CCSP Series – Chapter # 1 In today’s digital landscape, security has become of paramount importance. Organizations across the globe are faced with the challenges of protecting sensitive information, mitigating cyber attacks, and ensuring compliance with regulatory frameworks. To address these challenges, security professionals turn to industry-recognized frameworks and principles that provide a solid foundation for security practices. One such framework is the SANS Security Principles, which encompasses a comprehensive set of guidelines and best practices to strengthen security measures. In this blog, we will take a deep dive into the key principles of the SANS Institute’s security framework, exploring the importance of risk management, the role of asset inventories, the essentials of cloud security architecture, and more. 1. Understanding the Foundation of Cloud Security Cloud computing has revolutionized the way organizations store, access, and manage data. With the increasing adoption of cloud services, it is crucial to understand the foundation of cloud security. The security of cloud environments relies on several key factors, including risk management, asset management, and adherence to best practices. By establishing a strong security framework, organizations can protect sensitive information, safeguard critical systems, and mitigate security risks. 1.1 Importance of a Risk Management Approach One of the core principles of effective security management is adopting a risk management approach. By identifying, assessing, and managing security risks, organizations can make informed decisions that enhance their security posture. A risk management approach enables organizations to determine the potential impact of security threats, implement security controls, and align security practices with business objectives. It also helps in identifying and addressing vulnerabilities in information systems, thus ensuring the confidentiality, integrity, and availability of critical assets. Prioritizing risk management supports proactive security measures, keeping organizations ahead of emerging threats and enabling a robust incident response plan. 1.2 The Role of a Comprehensive Asset Inventory A comprehensive asset inventory is a cornerstone of effective security planning. By maintaining an up-to-date list of all assets, organizations gain visibility into their IT infrastructure, allowing them to track and secure resources effectively. An asset inventory enables organizations to identify critical systems, applications, and data, which in turn helps in implementing appropriate security controls and measures. It also provides insights for targeted security measures, ensuring that sensitive information is adequately protected. A comprehensive asset inventory strengthens the overall security posture of an organization, enabling it to respond effectively to security incidents and minimize the impact of potential attacks. 2. Essentials of Cloud Security Architecture As organizations increasingly migrate their systems and data to cloud environments, understanding the essentials of cloud security architecture is crucial. Cloud security architecture encompasses the design, implementation, and management of security controls in cloud environments. It involves following best practices and leveraging frameworks such as the Well-Architected Framework to ensure secure cloud deployments. By adopting cloud security best practices, organizations can safeguard their data, protect against cyber threats, and maintain compliance with regulatory requirements. 2.1 The Well-Architected Framework and its Best Practices The Well-Architected Framework, developed by Amazon Web Services, provides a set of best practices for designing and operating secure, scalable, and reliable cloud systems. The framework covers pillars of security, reliability, performance efficiency, cost optimization, and operational excellence. Following the Well-Architected Framework enables organizations to implement cloud security best practices, such as establishing strong network controls, managing access and authentication, and implementing data encryption. By adhering to these best practices, organizations can build and maintain cloud systems that are resilient to cyber threats and ensure the availability and integrity of data. 2.2 Common Security Pillars: Identity, Access Management, Encryption To establish a robust security framework, organizations must focus on common security pillars, namely identity and access management, and encryption. These pillars play a critical role in protecting sensitive data and mitigating security risks. - Strengthening identity and access management practices ensures that only authorized users have access to sensitive information. - Encryption, a process of transforming data into an unreadable form, provides an additional layer of security, rendering data useless even if it is intercepted by unauthorized individuals. - Implementing security measures, such as network segmentation and vulnerability management, further fortifies the overall security posture of an organization. 2.3 Selection Criteria for Cloud Service Providers (CSPs) When choosing cloud service providers (CSPs), organizations must consider several factors, especially those related to security. These criteria help ensure that data is protected, compliance requirements are met, and the overall cloud environment is secure. - Data protection: Organizations should evaluate the data protection measures offered by CSPs, such as data backup, encryption, and access controls. - Security compliance: Ensuring that CSPs adhere to security standards and regulatory frameworks, such as ISO 27001 or the Payment Card Industry Data Security Standard (PCI DSS). - Incident response plan: Evaluating whether CSPs have effective incident response plans in place to handle security incidents and minimize their impact. - Security awareness and training: Assessing whether CSPs prioritize security awareness and training for their employees, ensuring they have the necessary skills to protect customer data and systems. 3. The Intersection of Business, Security, and Frameworks To achieve holistic security, organizations must understand the intersection of business objectives, security requirements, and frameworks. By aligning IT resources with business requirements, organizations can ensure that security measures are implemented to support operational efficiency and strategic alignment. Frameworks like the Cloud Security Alliance’s Enterprise Framework provide guidelines for secure cloud operations, helping organizations implement security measures in a structured and standardized manner. 3.1 Aligning IT Resources with Business Requirements Alignment of IT resources with business requirements is essential for organizations to achieve optimal security and operational efficiency. This alignment ensures that security measures are implemented in a manner that supports business goals while maintaining compliance with regulatory frameworks. - Operational efficiency: Matching IT resources, such as infrastructure, applications, and security systems, with business needs allows organizations to optimize operational efficiency, reducing costs and improving performance. - Strategic alignment: Ensuring that IT resources are aligned with business requirements supports strategic alignment, enabling organizations to leverage technology to drive their long-term business objectives. - Risk management: Understanding and addressing business requirements through IT resource alignment facilitates effective risk management, allowing organizations to identify and mitigate potential security risks early on. 3.2 The Role of the Cloud Security Alliance’s Enterprise Framework The Cloud Security Alliance (CSA) provides a robust framework, known as the Enterprise Framework, that offers guidance for security architecture in cloud environments. This framework helps organizations implement security measures based on industry best practices, ensuring data protection, risk management, and compliance. - Secure cloud operations: The Cloud Security Alliance’s Enterprise Framework offers guidelines for secure cloud operations, helping organizations implement security controls across various layers of their cloud infrastructure. - Risk management framework: The framework provides a risk management framework that enables organizations to identify and assess security risks, implement controls, and develop incident response plans. - Security governance: By adopting the Cloud Security Alliance’s Enterprise Framework, organizations can establish security governance practices, ensuring accountability, compliance, and continuous improvement of their cloud security posture. 4. The Unity of DevOps and Security Traditionally, security has been perceived as separate from the development and operations processes, leading to potential vulnerabilities. However, the integration of security into DevOps workflows, commonly known as DevSecOps, helps organizations reduce security risks, ensure compliance, and deliver higher-quality software products. Another aspect of this unity is SecDevOps, which emphasizes the security considerations throughout the development lifecycle. By bridging the gap between development, operations, and security, organizations can achieve a more robust and secure software development process. 4.1 Understanding DevSecOps and SecDevOps DevSecOps and SecDevOps represent a paradigm shift in cybersecurity, emphasizing the integration of security practices into the development and operations processes. These approaches promote security awareness and collaboration across teams, ensuring that security measures are considered from the beginning of the development process. - Early security consideration: By implementing security as an integral part of the development lifecycle, organizations can identify and address security issues early on, reducing the risk of vulnerabilities in software or systems. - Continuous security testing: Ensuring continuous security testing throughout the development cycle helps in identifying security weaknesses, vulnerabilities, and threats, enabling organizations to address them promptly. - Automation in security practices: Emphasizing the importance of automation in security practices allows organizations to streamline security measures, enhancing efficiency and consistency. 4.2 The ‘Shift Left’ in Security Testing The concept of the ‘shift left’ in security testing refers to the practice of moving security testing activities earlier in the development process. By identifying and fixing security issues at the onset, organizations can reduce the overall risk of vulnerabilities and cyber attacks. - Early identification of security issues: Shifting security testing to earlier stages of the development process enables organizations to identify security issues, such as vulnerabilities or misconfigurations, before they become more challenging to address. - Improved security posture: Addressing security concerns early on enhances the overall security posture of applications, systems, and infrastructure, providing organizations with a stronger defense against cyber threats. - Automated security testing: Utilizing tools for automated security testing in the early phases of development helps organizations streamline vulnerability management, reducing the manual effort required to identify and mitigate security risks. 5. Evaluating Cloud Service Providers (CSPs) Selecting the right cloud service provider (CSP) is critical for organizations looking to leverage cloud computing benefits while ensuring data protection, compliance, and security. When evaluating CSPs, several factors need to be considered to build trust and make informed decisions. 5.1 Standards Utilization and Regulatory Compliance in CSP Evaluation Regulatory compliance and standards utilization play a vital role in evaluating cloud service providers. Organizations need to ensure that CSPs comply with regulatory requirements and adhere to security standards. - Regulatory compliance: Evaluating CSPs based on compliance with relevant regulations, such as data protection regulations, ensures that data is handled and stored securely, mitigating regulatory risks. - Standards utilization: Assessing how CSPs utilize security standards, such as ISO 27001, NIST Cybersecurity Framework, or industry-specific frameworks, helps organizations determine the level of security controls and practices implemented by the service provider. - Service providers: Understanding the security measures and controls implemented by the CSP, including incident response plans, access controls, data protection measures, and physical security, is essential in evaluating their overall security capabilities. 6. Adherence to International and Industry Standards To ensure security and compliance, organizations must adhere to international and industry security standards. Several frameworks and standards, such as ISO 27001/27002, ISO 27017/27018, PCI DSS, FedRAMP, and UK G-Cloud, provide guidelines for establishing robust security practices. 6.1 Overview of ISO 27001/27002, ISO 27017/27018, PCI DSS, FedRAMP and UK G-Cloud When considering information security, it’s vital to grasp the scope and requirements of ISO 27001 and 27002. Additionally, understanding the critical security controls within ISO 27017/27018 is paramount for robust cyber security. PCI DSS plays a crucial role in securing sensitive information like payment card data. Moreover, familiarity with FedRAMP is essential for evaluating cloud services’ security risk levels. Lastly, an overview of UK G-Cloud standards sheds light on government cloud services’ compliance purposes. In conclusion, understanding and implementing SANS security principles are crucial for maintaining a robust and secure cloud environment. By focusing on risk management, asset inventory, cloud security architecture, and aligning IT resources with business requirements, organizations can enhance their security posture. The integration of DevOps and security, adherence to international standards, and thorough evaluation of cloud service providers further strengthen the security framework. Embracing these practices ensures a comprehensive approach to cloud security that safeguards data, systems, and operations effectively. Stay proactive, informed, and diligent in your security efforts to mitigate risks and protect your valuable assets in the digital realm. The opinions expressed in this article belongs to the individual contributors and do not necessarily reflect the views of Information Security Buzz.
<urn:uuid:8ded3b95-e305-4719-9f49-6431f60498de>
CC-MAIN-2024-38
https://informationsecuritybuzz.com/sans-security-principles-a-deep-dive/
2024-09-14T00:31:30Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651540.77/warc/CC-MAIN-20240913233654-20240914023654-00084.warc.gz
en
0.910389
2,530
2.765625
3
Gartner defines edge computing as solutions that facilitate data processing at or near the source of data generation. For example, in the context of the Internet of Things (IoT), the sources of data generation are usually things with sensors or embedded devices. Edge computing serves as the decentralized extension of the campus networks, cellular networks, data center networks or the cloud. (Source: https://www.gartner.com/smarterwithgartner/what-edge-computing-means-for-infrastructure-and-operations-leaders) MEC is often advertised as an independent entity, but in practice, the processing of data for both the User Plane (UP) and Control Plane (CP) is still centralized. In other words, part of the processing is executed at the Edge site while the central site does the aggregation and analysis of the whole data. Many probe systems operate this way by collecting the data on the site, anonymizing the data and sending the data for central processing. Unlike with a direct cable connection, this data is sent over the company’s data network using TCP/IP. To comply with data security and privacy, this solution needs a more secure method of sharing the data. Cubro’s solution uses a VPN connection from the sites to the central data center. Cubro’s comprehensive approach integrates tapping, network packet brokering, packet routing, and data encryption through VPN services. Although VPN connections are considered slow, Cubro’s VPN has been tested to provide 300 Mbps. Cubro uses Omnia120 at both ends, one Omnia120 is for establishing a secure, encrypted VPN connection while the receiving end removes the VPN tunnel and provides the data to the recipient without encryption. This connection is established by using a physical cable. Omnia120 has a unique architecture with Silicone on Chip and two ARM processors. One ARM CPU is used for the VPN connection while the chipset and second ARM CPU can be used for other tasks such as packet deduplication and packet filtering including Regex or the user can use, for example, Wireshark. Enterprises may also use Omnic, Cubro SmartNIC, with the same software. Communication to the Omnia120 and Omnic is also encrypted using certificates that authenticate the user and encrypt the traffic. As the prominence of MEC and distributed processing grows within enterprises and CSPs, data transfer is streamlined and bolstered by enhanced security measures. Cubro’s expertise extends to facilitating GDPR compliance by addressing Personally Identifiable Information (PII) concerns across the data chain.
<urn:uuid:e60aa01d-393e-4393-b1cf-aedbf41d1905>
CC-MAIN-2024-38
https://www.cubro.com/en/blog/enhancing-security-measures-for-mobile-edge-computing-mec/
2024-09-14T01:58:58Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651540.77/warc/CC-MAIN-20240913233654-20240914023654-00084.warc.gz
en
0.916299
538
2.890625
3
VLANs are local to each switch's database, and VLAN information is not passed between switches. Trunk links provide VLAN identification for frames traveling between switches. Cisco switches have two Ethernet trunking mechanisms: ISL and IEEE 802.1Q. Certain types of switches can negotiate trunk links. Trunks carry traffic from all VLANs to and from the switch by default but can be configured to carry only specified VLAN traffic. Trunk links must be configured to allow trunking on each end of the link. Trunk links are required to pass VLAN information between switches. A port on a Cisco switch is either an access port or a trunk port. Access ports belong to a single VLAN and do not provide any identifying marks on the frames that are passed between switches. Access ports also carry traffic that comes from only the VLAN assigned to the port. A trunk port is by default a member of all the VLANs that exist on the switch and carry traffic for all those VLANs between the switches. To distinguish between the traffic flows, a trunk port must mark the frames with special tags as they pass between the switches. Trunking is a function that must be enabled on both sides of a link. If two switches are connected together, for example, both switch ports must be configured for trunking, and they must both be configured with the same tagging mechanism (ISL or 802.1Q). To enable trunking between the switches, use the following steps: Enable trunking on a port. Enable the trunk: set trunk mod/port [auto | desirable | on | nonegotiate | off] (global) interface type mod/port (interface) switchport mode dynamic [auto | desirable] (interface) switchport mode trunk (interface) switchport nonegotiate The most basic way to configure a trunk link is using the option on. This option enables the trunk and requires that you also specify a tagging mechanism for the trunk. For IOS devices, the command switchport mode trunk is equivalent to the set trunk mod/port on command. When specifying the option on, you must also choose a tagging mechanism (see Step 1b). Some IOS switches do not support Dynamic Trunking Protocol. For these switches, the only command that you can use to configure trunking is switchport mode trunk, which essentially turns trunking on. Many Cisco switches employ an automatic trunking mechanism known as the Dynamic Trunking Protocol (DTP), which allows a trunk to be dynamically established between two switches. All COS switches and integrated IOS switches can use the DTP protocol to form a trunk link. The COS options auto, desirable, and on and the IOS options of dynamic auto, dynamic desirable, and trunk configure a trunk link using DTP. If one side of the link is configured to trunk and will send DTP signals, the other side of the link will dynamically begin to trunk if the options match correctly. If you want to enable trunking and not send any DTP signaling, use the option nonegotiate for switches that support that function. If you want to disable trunking completely, use the off option for a COS switch or the no switchport mode trunk command on an IOS switch. Table 6-2 shows the DTP signaling and the characteristics of each mode. It is important to remember that not all switches support DTP and might not establish a trunk without intervention. Also remember that DTP offers no benefit when you are trunking with a non-Cisco switch. To eliminate any overhead associated with DTP, it is useful to use the nonegotiate option when DTP is not supported. When enabling trunking, it is not possible to specify a range of ports. Table 6-2 Trunking Mode Characteristics COS = on IOS = mode trunk Trunking is on for these links. They will also send DTP signals that attempt to initiate a trunk with the other side. This will form a trunk with other ports in the states on, auto, or desirable that are running DTP. A port that is in on mode always tags frames sent out the port. COS = desirable IOS = mode dynamic desirable These links would like to become trunk links and will send DTP signals that attempt to initiate a trunk. They will only become trunk links if the other side responds to the DTP signal. This will form a trunk with other ports in the states on, auto, or desirable that are running DTP. This is the default mode for the 6000 running Supervisor IOS. COS = auto IOS = mode dynamic auto These links will only become trunk links if they receive a DTP signal from a link that is already trunking or desires to trunk. This will only form a trunk with other ports in the states on or desirable. This is the default mode for COS switches. COS = nonegotiate IOS = mode nonegotiate Sets trunking on and disables DTP. These will only become trunks with ports in on or nonegotiate mode. COS = off IOS = no switchport mode trunk This option sets trunking and DTP capabilities off. This is the recommended setting for any access port because it will prevent any dynamic establishments of trunk links. Cisco 2950 and 3500XL switches do not support DTP and are always in a mode similar to nonegotiate. If you turn trunking on for one of these devices, it will not negotiate with the other end of the link and requires that the other link be configured to on or nonegotiate. Specify the encapsulation method: set trunk mod/port [negotiate | isl | dot1Q] (global) interface type mod/port (interface) switchport trunk encapsulation [negotiate | isl | dot1Q] The other option when choosing a trunk link is the encapsulation method. For Layer 2 IOS switches, such as the 2900XL or the 3500XL, the default encapsulation method is isl. You can change from the default with the switchport trunk encapsulation command. For COS switches or integrated IOS switches, the default encapsulation is negotiate. This method signals between the trunked ports to choose an encapsulation method. (ISL is preferred over 802.1Q.) The negotiate option is valid for auto or desirable trunking modes only. If you choose on as the mode or if you want to force a particular method or if the other side of the trunk cannot negotiate the trunking type, you must choose the option isl or dot1Q to specify the encapsulation method. Not all switches allow you to negotiate a trunk encapsulation setting. The 2900XL and 3500XL trunks default to isl and you must use the switchport trunk encapsulation command to change the encapsulation type. The 2950 and some 4000 switches support only 802.1Q trunking and provide no options for changing the trunk type. (Optional) Specify the native VLAN: set vlan number mod/port (global) interface type mod/port (interface) switchport trunk native vlan number For switches running 802.1Q as the trunking mechanism, the native VLAN of each port on the trunk must match. By default all COS ports are in VLAN 1; and the native VLAN on the IOS devices is also configured for VLAN 1, so the native VLAN does match. If you choose to change the native VLAN, use the set vlan command for COS switches or the switchport trunk native vlan command for IOS switches to specify the native VLAN. Remember that the native VLAN must match on both sides of the trunk link for 802.1Q; otherwise the link will not work. If there is a native VLAN mismatch, Spanning Tree Protocol (STP) places the port in a port VLAN ID (PVID) inconsistent state and will not forward on the link. Cisco Discovery Protocol (CDP) version 2 passes native VLAN information between Cisco switches. If you have a native VLAN mismatch, you will see CDP error messages on the console output. Specifying VLANs to Trunk By default a trunk link carries all the VLANs that exist on the switch. This is because all VLANs are active on a trunk link; and as long as the VLAN is in the switch's local database, traffic for that VLAN is carried across the trunks. You can elect to selectively remove and add VLANs from a trunk link. To specify which VLANs are to be added or removed from a trunk link, use the following commands. (Optional) Manually remove VLANs from a trunk link: clear trunk mod/port vlanlist (global) interface type mod/port (interface) switchport trunk allowed vlan remove vlanlist By specifying VLANs in the vlanlist field of this command, the VLANs will not be allowed to travel across the trunk link until they are added back to the trunk using the command set trunk mod/port vlanlist or switchport trunk allowed vlan add vlanlist. After configuring a port for trunking, use one of the following commands to verify the VLAN port assignments: show trunk [mod] [mod/port] (privileged) show interface type mod/port switchport show interfaces trunk show interface [mod] [interface_id] trunk The commands show interfaces trunk and show interface [mod] [interface_id] trunk are not available on all switches that run IOS. In this example the switches Access_1 and Distribution_1 and Core_1 are connected as shown in Figure 6-2. 802.1Q trunking is configured in the on mode between Access_1 and Distribution_1 switches. ISL is configured in desirable mode on the Distribution_1 switch to the link connecting to the core. The core is configured for autotrunking mode and encapsulation negotiate. The trunk connected between the access switch is configured to only trunk for VLANs 5, 8, and 10. The trunk between the Distribution_1 and Core_1 is configured to carry only VLAN 1 and VLAN 10. Figure 6-2 Network Diagram for Trunk Configuration on Access_1, Distribution_1, and Core_1 An example of the Catalyst OS configuration for Distribution_1 follows: Distribution_1 (enable)>clear trunk 1/1 2-1001 Distribution_1 (enable)>set trunk 1/1 desirable isl 10 Distribution_1 (enable)>clear trunk 2/1 2-1001 Distribution_1 (enable)>set trunk 2/1 on dot1q 5,8,10 An example of the Catalyst OS configuration for Core_1 follows: Core_1 (enable)>clear trunk 1/1 2-1001 Core_1 (enable)>set trunk 1/1 10 An example of the Supervisor IOS configuration for Core_1 follows: Core_1(config)#interface gigabitethernet 1/1 Core_1(config-if)#switchport encapsulation negotiate Core_1(config-if)#switchport mode dynamic auto Core_1(config-if)#switchport trunk allowed vlan remove 2-1001 Core_1(config-if)#switchport trunk allowed vlan add 10 Core_1 (config-if)#end Core_1#copy running-config startup-config An example of the Layer 2 IOS configuration for Access_1 follows: Access_1 (config)#interface gigabitethernet 0/1 Access_1 (config-if)#switchport mode trunk Access_1 (config-if)#switchport trunk encapsulation dot1q Access_1 (config-if)#switchport trunk allowed vlan remove 2-1001 Access_1 (config-if)#switchport trunk allowed vlan add 5,8,10 Access_1 (config-if)#end Access_1#copy running-config startup-config
<urn:uuid:47d23e31-9fd4-4b62-a063-1b15c16bfb3a>
CC-MAIN-2024-38
https://www.ciscopress.com/articles/article.asp?p=29803&seqNum=3
2024-09-15T02:51:06Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651614.9/warc/CC-MAIN-20240915020916-20240915050916-00884.warc.gz
en
0.819301
2,564
2.859375
3
Is Cyber Security Needed in the Future? | Stay Safe As technology grows, so does our need for cybersecurity. We’re more connected than ever, putting our digital lives at risk. This article will look into the future of cyber threats, data protection, and what we can do to stay safe online. - The digital landscape is constantly evolving, requiring vigilance and adaptability in cybersecurity measures. - Protecting sensitive data and personal information is crucial as our reliance on technology and online systems increases. - Proactive strategies, such as network security and cyber defense, will be essential in mitigating the impact of cyber attacks in the years to come. - Emerging technologies, like artificial intelligence, could play a significant role in the future of cybersecurity, both in terms of threats and defenses. - Cybersecurity will remain a critical priority as the digital landscape continues to shape our personal and professional lives. Cybersecurity: Safeguarding the Digital Age Our world is more connected than ever, making strong cybersecurity measures crucial. In this digital era, we face many digital threats that can harm our information, disrupt systems, and threaten our privacy. Cyber Threats: Evolving Risks in a Connected World Cybercriminals are always coming up with new ways to break into systems, steal data, and cause trouble. They use everything from ransomware to complex threats. Cybersecurity challenges are always changing for both companies and people. To stay ahead, we need to understand the latest cybersecurity trends and be proactive about cyber resilience. Data Protection: Shielding Sensitive Information With growing worries about digital privacy, protecting sensitive data is key. Companies and individuals must use strong cybersecurity solutions to guard against cybersecurity risks like data breaches and identity theft. Good risk management and ethical hacking can spot and stop these threats. The need for cybersecurity jobs is increasing, and we need people with the right skills. Keeping up with the latest cybersecurity innovations and trends is vital. This helps individuals and companies stay ahead and keep their systems safe from digital threats. Network Security: Fortifying Digital Infrastructures In today’s digital world, strong network security is key. Our world is more connected than ever, making it crucial to protect our digital spaces. We must focus on cyber defense strategies, data privacy, and protecting our digital infrastructure to manage cyber risks well. The cybersecurity team is crucial for keeping our digital systems safe. They use new tech like artificial intelligence and machine learning to fight off complex threats. By using strong cyber defense plans, companies can keep their networks safe and protect private data better. - Implementing robust access controls and authentication protocols to restrict unauthorized access to digital assets. - Deploying advanced firewalls and intrusion detection systems to monitor network traffic and identify potential threats. - Regularly update software and apply security patches to address known vulnerabilities. - Conducting regular risk assessments to identify and address emerging cyber threats. - Educating employees on best practices for cyber hygiene and reporting suspicious activities. By focusing on network security, companies can strengthen their digital setups. This helps protect sensitive info and keeps stakeholders trusting them. Cybersecurity is very important as we deal with the digital world’s challenges. Cyber Attacks: Mitigating Destructive Incidents Cyber attacks are a big worry in today’s digital world. They threaten our info and can hurt businesses and groups a lot. We need strong cybercrime strategies to fight these threats. Online Privacy: Preserving Digital Identities Our lives are more digital now, so keeping our online privacy safe is key. Cyber crooks keep finding new ways to steal our personal and money info. We must use strong online safety steps to keep our digital lives safe. Groups need to fight cyber threats with a strong plan. By being alert, using the latest security tools, and teaching users how to stay safe, we can lessen cyber attack damage. This helps protect our digital future. We must stay ahead and informed to beat these cyber threats. By knowing how cyber crooks work and using strong digital safety plans, we can protect our online privacy. This keeps our digital identities safe. Read More: Top 10 Most Common Types of Cyber Attacks Cyber Defense: Proactive Strategies for Protection The digital world is changing fast, making strong cyber defense strategies more important than ever. As threats grow, we must act early to protect our online stuff and stay safe. Cybercrime Prevention: Deterring Malicious Activities Stopping cybercrime is key to good cyber defense. We can stop bad actors by using smart strategies. This means using the latest in cybersecurity ethical hacking and plans to find and fix weak spots before they’re used. Keeping our online life and personal info safe is also crucial. This helps prevent identity theft, data breaches, and other bad stuff that can really hurt us. - Adopt robust cyber defense strategies to address emerging threats proactively - Implement comprehensive cybercrime prevention measures to deter malicious actors - Leverage cybersecurity innovations and ethical hacking techniques to identify and address vulnerabilities - Prioritize digital privacy protection to safeguard sensitive information - Cultivate a culture of cyber resilience within your organization or personal digital ecosystem By using proactive cyber defense and stopping cybercrime, we can make the internet safer and stronger. This helps everyone, from individuals to big companies, deal with the changing cyber threats. Digital Safeguards: Implementing Robust Security Measures In today’s digital world, strong security is a must. Cybercrime and digital risk management are top concerns. That’s why it’s key to use solid digital safeguards to keep information safe and follow online safety rules. Using strong network security is a big part of keeping data safe. This means setting up firewalls, intrusion detection systems, and VPNs. These tools help protect digital spaces from cyber threats. But it’s not just about tech. Teaching employees about cybersecurity is also vital. Training them on how to handle data and respond to threats makes them a strong defense against cyber attacks. Adding things like data encryption, secure cloud storage, and keeping software up-to-date also helps. These steps make it harder for hackers to get to sensitive info. They also lessen the damage if a data breach happens. As technology changes, keeping up with new cyber threats is key. It helps keep online spaces safe and protects against more complex cyber-attacks. Cybersecurity Strategies: Holistic Approaches for Resilience The digital world is always changing, bringing both new chances and big challenges for cybersecurity in 2030. Experts say we need to use a whole approach to stay strong against cyber threats. New trends and ideas will change how we keep our digital stuff safe and protect our privacy. Future of Cybersecurity 2030: Emerging Trends and Innovations Experts see a future where keeping data safe and securing networks will be top priorities. New techs like artificial intelligence and machine learning will change how we fight cyber attacks. We’ll also need strong plans for managing digital risks and following safety rules online to tackle new challenges. Here are some trends we’ll see in cybersecurity by 2030: - More use of AI to find and fight threats - New encryption that’s safe against quantum computers - Security for connected devices - Planning for risks and how to handle them - Working together between governments, companies, and security pros Looking ahead, cybersecurity experts need to be ready for the changing world of cyber threats. By using a complete approach to cybersecurity, we can make our digital world strong and safe. This way, we can keep our online activities safe and private in the future. Is cyber security needed in the future? As technology advances, the need for cyber security is more important than ever. Our digital lives depend on it. Cyber security is key to protecting our personal and professional lives. Cyber threats are growing and getting more complex. Hackers are finding new ways to attack our digital world. We need strong cyber security to keep our data safe and protect us from harm. Our world is getting more connected, thanks to the Internet of Things (IoT) and cloud computing. This means we face more risks. We need strong cyber security to keep up with these changes and protect our digital world. Without good cyber security, our digital future is at risk. We could face data breaches and system failures. This would put individuals, businesses, and governments in danger. Cyber security is vital for a reliable and trustworthy digital world. As we use more automation and connect more devices, we’ll need better cyber security. By investing in new technologies and strategies, we can keep our digital world safe. To answer the question, “Is cyber security needed in the future?” the answer is a clear yes. Cybersecurity will keep growing in importance as we rely more on digital systems. By tackling cyber security challenges, we can make a safer digital future for everyone. Read More: Why AI is the future of cybersecurity? Cyber Security’s Future: Challenges and Opportunities Looking ahead, the future of cyber security is full of both hurdles and chances. The field is changing fast, bringing new threats and areas to watch. We must stay alert and keep finding new ways to protect ourselves. What is the Future of Cyber Security? The attack surface is getting bigger as we connect more devices and systems. This means more ways for cyber threats to get in. Cybercriminals are getting smarter, using AI and machine learning for more precise attacks. But there’s hope on the horizon. New techs like biometrics, quantum cryptography, and cloud security are making things better. AI and machine learning are helping us spot and stop threats faster. And with 5G and IoT, we can talk securely over longer distances. Teaching people about cyber security is also key. It helps us all deal with the challenges better. By learning about online safety and managing risks, we can face the future with confidence. The future of cyber security will need a mix of efforts. We’ll need to work together to keep our digital world safe. By being innovative, supporting talent, and focusing on resilience, we can tackle the challenges ahead. This way, we can keep our digital future safe and bright for everyone. Read More: How is AI used in cyber security? Future of Cybersecurity with AI: Leveraging Advanced Technologies The world is getting more digital, and cybersecurity is changing fast with AI. AI is bringing new ways to fight cyber threats. It’s making threat detection, response, and overall security better. AI can look at huge amounts of data quickly. It spots patterns and finds odd things fast. This means it can find threats that humans might miss. This makes fighting cyber attacks faster and safer for companies. AI can also do some cybersecurity tasks on its own. This includes checking for weaknesses, updating security, and handling incidents. This lets people focus on harder security problems. It makes security work more efficient and effective. But, using AI in cybersecurity has its problems and ethical issues. We need to worry about privacy, bias in algorithms, and how AI could be used for bad things. It’s important to make AI tools responsibly to protect our data and privacy. As cybersecurity changes, using AI will be key to staying safe. AI helps companies stay ahead of threats. It makes them more secure, improves how they handle security and protects their online stuff better. Read More: Artificial Intelligence in Cyber security Cyber Security in the Next 10 Years: Predictions and Preparedness Looking ahead, the future of cyber security is set to be exciting and always changing. Technology is moving fast, and cyber threats are getting more complex. This makes cyber security more important than ever for protecting our digital lives. Experts believe cyber security will keep being key in fighting cyber attacks. New tech like artificial intelligence and quantum computing will change and test the cyber security world. Does Cyber Security Have a Future? Cyber security is not just important; it’s essential. As we rely more on digital systems and the internet, we’ll need strong, flexible, and ahead-of-the-game cyber security. Here are some predictions and things to think about for the future: - Increased focus on cloud security and the protection of data stored in the cloud - Advancements in biometric authentication and identity management to enhance digital security - Widespread adoption of machine learning and artificial intelligence to detect and mitigate cyber threats in real-time - Greater emphasis on supply chain security and the vulnerability of interconnected systems - The emergence of quantum-resistant encryption to safeguard data against the potential threats posed by quantum computing To keep cyber security strong in the future, we must stay alert, flexible, and proactive. By using new tech, working together, and keeping up with learning, we can make a safer and stronger digital future. Will AI replace Cyber Security? The debate on whether AI will replace traditional cybersecurity is growing. AI can detect and respond to cyber threats well. But, experts say it will only partially replace traditional methods. Instead, AI and cybersecurity experts will work together to protect us from future threats. AI is great at quickly analyzing lots of data and spotting patterns that might mean a cyber threat. It can automate some security tasks, making things more efficient. This lets cybersecurity teams focus on harder challenges. But, AI needs humans to make sure it makes ethical and clear decisions that fit with security goals. The threats we face are getting more complex, with hackers finding new ways to get past security. Just using AI alone would leave us open to attacks. The key to a strong defense is the partnership between AI and human cybersecurity experts. This team will be essential in keeping us safe from new cyber threats. Read More: Can AI replace cybersecurity? Is cyber security needed in the future? Yes, cybersecurity is vital for the future as our digital world grows. With new tech, threats keep changing. We need strong cybersecurity to protect our digital world and keep information safe. What is the future of cyber security? Cybersecurity’s future looks dynamic and challenging. New threats will keep coming. But we can fight them with new security tech like AI and machine learning. As tech changes, we must stay ahead to protect our digital world. Will AI replace cyber security? AI will play a big role in cybersecurity but won’t replace it. AI can help detect threats and respond faster. But we still need human skills and strategies to tackle complex threats. The future will blend AI and human efforts for better security. What are the future challenges and opportunities in cyber security? Cybersecurity’s future has both challenges and chances. We face fast-evolving threats and need to protect more devices. There’s also a need for more cybersecurity experts and complex rules to follow. But, new tech like AI and cloud computing can improve security. Plus, more people are paying attention to cybersecurity, which helps us work together and find new solutions. Does cyber security have a future? Absolutely, cybersecurity has a bright future. As we connect more devices, we’ll need better cybersecurity. It’s key for protecting important systems, keeping data safe, and making our digital world secure. With new tech, cybersecurity will keep evolving to meet new threats, making it vital for our digital future.
<urn:uuid:c4a753c8-3607-4f15-8e44-5e9567f94949>
CC-MAIN-2024-38
https://arksolvers.com/is-cyber-security-needed-in-the-future/
2024-09-16T08:47:31Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651682.69/warc/CC-MAIN-20240916080220-20240916110220-00784.warc.gz
en
0.926455
3,171
2.8125
3
The CREATE INDEX statement creates an index on an existing base table. The index contains the columns specified and is keyed on those columns, in the order they are specified. Indexes can improve query processing. If data is retrieved from a table based on an indexed column, the DBMS uses indexes, if available, to accelerate query processing. To obtain the greatest benefit, create indexes that contain all of the columns that are generally queried and keyed on some subset of those columns. Any number of indexes can be created for a table, but, for portability, each index can contain no more than 16 columns. To prevent the index from accepting duplicate values in key fields, specify the UNIQUE option. If the base table on which the index is being created has duplicate values for the index’s key fields, then the create index statement will fail. Similarly, if you attempt an insert or update that violates the uniqueness constraint of an index created on the table, then the insert or update will fail. This is true for an UPDATE statement that updates multiple rows: the UPDATE statement will fail as soon as it attempts to write a row that update violates the uniqueness constraint. Particular Enterprise Access products may support extensions to the CREATE INDEX statement (using the WITH clause). To ensure application portability, follow each CREATE INDEX statement with a COMMIT statement. An index cannot be updated directly. When a table is changed, the DBMS updates indexes as required. To destroy an index, use the DROP statement. All indexes on a table are destroyed when the table is dropped.
<urn:uuid:620c15a9-8d6c-4351-a7b1-46918c5f0b53>
CC-MAIN-2024-38
https://docs.actian.com/ingres/10S/OpenSQL/Description_6.htm
2024-09-18T20:50:54Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651941.8/warc/CC-MAIN-20240918201359-20240918231359-00584.warc.gz
en
0.788133
330
2.734375
3
4 Password Management Tips There is no better way to protect your data than simple password management. Some of the old wisdom still applies when it comes to password protection, but technology is always changing. Read more to learn these four password management tips. These changes bring new advice and new ways to help protect your IT system from getting infiltrated by those who wish to do you harm. Creating a Strong Password Building strong IT security is all about having a good foundation. A strong password is the beginning of your cyber safety. The goal is to create something that computers will not be able to guess through random input. Follow these pieces of advice to increase your password security. 1. Make a Password at least 12 characters long A strong password is all about trying to stop computers from successful brute force attacks. These attacks use automated systems to try countless times until they get into your system. 5 character passwords are easily broken. A strong program can break most 5 character passwords in less than a minute. For years it was common practice to suggest 8 characters to be considered strong, but the common thought is that they are no longer enough to stop sophisticated software. A study from Georgia Tech Research Institute found that the best passwords have to be at least 12 characters in length to satisfy the minimum security. However, more and more security companies are suggesting passwords of at least 15 and preferably 20 characters in length to have a sufficiently secure password. 2. Make your password complex and unique Using symbols, numbers, spaces, misspellings, and any other way you can think of to make your password unique is advised. However, adding symbols is not as helpful as adding length. An 8 character password with symbols and numbers is still easier to guess than a 20 character password of just lowercase letters. A good way to ensure a unique password is to use a password generator. Password generators can add safety, but you will want to ensure it’s from a safe source. A human password may look like “MyDogsNamE123$$” while a generated one could be “JE56QqrN7WwvJ5y” The computer-generated password will be much harder to solve while being the same amount of characters. 3. Use a Password Manager Each password that you use should be unique from the others. It can get complicated to keep track of everything, especially if you are using computer-generated passwords. This is why we recommend using a password manager like LastPass. A password manager will keep track of all the complicated logins that you now use. This will allow you to have more security for each individual login while keeping everything manageable. Password managers are a great way to add security to your IT infrastructure. 4. Use Two Factor Authentication Two Factor Authentication (2FA) is one of the newest, and best ways to secure your passwords. If you use 2FA then you will have a secondary application that confirms that it is, indeed, you trying to log in. If a program was able to get into your data then you would get an alert to confirm that it was you. If it was not, then you can decline the login, and change your passwords. This will let you know that an attack has happened, and will help you stop a breach before they are able to get in. It is easy to feel secure with a password that you have chosen years ago. However, keeping things fresh, using new approaches, and adding tools like 2FA to your security can drastically improve your chances if someone wants to break in. Good password management practices and using tools like 2FA can be the best way to help secure data and protect your IT infrastructure. Contact us today to help your business with Cybersecurity.
<urn:uuid:49a6bb8d-cac4-481e-95ab-cdeecdeb4097>
CC-MAIN-2024-38
https://www.itsasap.com/blog/4-password-management-tips
2024-09-08T00:00:31Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650926.21/warc/CC-MAIN-20240907225010-20240908015010-00684.warc.gz
en
0.953597
771
2.6875
3
Environmental disruption will ultimately impact us all—but already, the effects are unequally dispersed. Women, despite representing half the planet’s population, are disproportionately impacted by environmental challenges, from climate change, to water scarcity, to pollution, to resource degradation and depletion. As we mark this year’s World International Women’s Day, themed, “DigitALL: Innovation and technology for gender equality,” it’s time to recognize that one of the biggest challenges of our time—the environment—is not a gender-neutral one. For the technology industry, it’s crucial to consider the gendered nature of the digital solutions we design and implement, particularly those aimed at sustainability. Tech companies also have an essential role to play in boosting women’s participation in developing environmental solutions, which—in the end—is essential to achieving global sustainability. What’s driving the gender imbalance Through the years, women have made significant contributions to solving environmental challenges. From Silent Spring’s Rachel Carson, to renowned whistleblower Erin Brockovich, to Nobel Laureate Prof. Wangari Maathai, to today’s best known environmental activist Greta Thunberg, women have historically been outspoken environmental advocates. Despite their immense contributions, women’s role in environmentalism continues to be undermined due, in large part, to gender inequalities. The biggest drivers of this imbalance span three key areas: the division of labor, resource access and control, and strategic decision making. - Uneven division of labor. Globally, the largest demographic in the fields of informal work and unpaid care is women; some estimates show women spend 2.6 times more time on unpaid care than men. This means that when natural disasters or climate catastrophes strike or are impending, women are less able to leave the impacted area to seek an alternative livelihood. For instance, in the 1995 Kobe earthquake in Japan, 1.5 times more women lost their lives than men. Similarly, 80% of those left to fend for themselves during Hurricane Katrina in 2005 in New Orleans were women, despite their representing 54% of the city’s population. In rural areas of developing economies, women rely on natural resources to provide for their families, whether walking long distances to collect water, fodder and fuel, or relying on environmentally dependent practices like rainfed farming. This makes women particularly vulnerable when such resources diminish due to environmental change. It’s not unusual for young girls to be pulled out of school to help their mothers, thus interrupting their educational pursuits and damaging their economic prospects. - Access and control over resources, finances and opportunities. Globally, women have lower access to, control over and ownership of resources that are essential to dealing with environmental disasters: natural resources, technology, information, education, finances and assets. When it comes to property and home ownership, less than 15% of all landholders globally are women, which translates to limited decision-making power on issues like productivity improvements or changing to alternative energy sources. Because women also have relatively lower access to technology resources like the internet and smartphones, they are less likely than legal property owners to get vital weather information. According to the International Telecommunication Union, more than 50% of women do not use the internet compared with 31% of men. Lack of property ownership also prevents women from accessing credit sources, which are often pegged to asset collateral; for example, women receive only 2.3% of startup funding and less than 30% of climate funds. The global gender pay gap further limits women’s ability to afford and adopt climate solutions. - Representation in strategic decision-making. Whether in business, government or civic settings, women are often under-represented in decision-making forums. In the United Nations Climate Change Conferences, for example, there was a public outcry about the lack of women in the conference negotiation forums, where they represented just 27% of the delegates. Only 32% of the authors on the Intergovernmental Panel on Climate Change are women, which can result in their insights being excluded from shaping the environmental policies. This under-representation is mirrored in many corporate settings, including in energy and sustainability roles. At the same time, research shows that women have contributed valuable insights on combating environmental change. Bringing women into the climate solution Because the tech industry is a key part of developing solutions to environmental challenges, it’s incumbent on tech companies to close the sustainability gender gap. They can do this by: - Taking a gender-responsive approach to innovation and technology. Tech companies need to ensure their project teams are gender diverse and that they use inclusive design principles that focus on actual end-user needs and contexts when creating digital solutions. For example, companies developing digital agricultural solutions for rural regions should consider women’s potentially lower tech literacy levels in their design. Establishing gender-diverse project teams may be challenging due to the gender gap in STEM fields. Companies should use their internal workforce data to understand why— despite young women outperforming young men in STEM education—women are under-represented in STEM professions. Analysis of data such as gender-disaggregated longitudinal data on recruitment, number of applicants, promotion ratios, retention at the one, three- and five-years mark, and family-friendly benefits can help businesses better understand what pushes women out and how they can retain and advance them. - Supporting organizations that support women in STEM. Technology companies can also direct their corporate social responsibility efforts toward helping to improve girls’ and women’s digital literacy, by providing technical upskilling and education. Through its support of Girlstart and other organizations that support women in STEM, Cognizant Foundation has helped more than 150,000 girls increase their interest in and aptitude for STEM careers through afterschool programs. Cognizant Outreach is also involved in driving digital inclusion across several demographic sectors. - Collaborating with gender-equality advocates. Through collaborative efforts between tech companies and gender-equality advocates, digital solutions can be developed to tackle both environmental challenges and gender inequality. Digital solutions can be developed to model and predict the spread of specific gendered impacts of climate change. By collecting, processing and visualizing data on gender-specific resource utilization like energy use, travel patterns, displacement during environmental disasters and use of digital interventions, tech companies may help answer novel questions in a timely and granular way. Businesses need to see gender equality as not just a reporting requirement or a stand-alone goal but as a crucial component of achieving overall sustainability. The tech industry is in a key position to help the world recognize—and rebalance—the sustainability gender equation.
<urn:uuid:0da30901-59fc-49ba-be41-0fe277d884a4>
CC-MAIN-2024-38
https://www.cognizant.com/us/en/insights/insights-blog/how-to-close-the-sustainability-gender-gap-wf1662544
2024-09-09T04:25:57Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651072.23/warc/CC-MAIN-20240909040201-20240909070201-00584.warc.gz
en
0.943078
1,371
3.78125
4
Combating Internet Worms In recent years, not only has the number of network and computer attacks been on the rise, but also the level of complexity and sophistication with which they strike. The most common and perhaps most damaging of these attacks are called worms. Worms are malicious programs written to exploit vulnerabilities within an operating system or an application environment and to then automatically seek out and find other vulnerable hosts to exploit and infect with the worm code. The worms travel rapidly affecting all neighboring systems of the initially infected host. This exponential propagation induces a large amount of network traffic that overwhelms bandwidth and system resources making applications and network services slow or even unavailable. Some worms also contain payloads including additional code to further exploit the host such as data modification (a web page) or thief of information. Network worms and viruses have existed for well over 20 years. One of the first and famous worm programs to impact the Internet was the Morris Worm in November of 1988. This worm exploited vulnerabilities in the finger and sendmail programs. At that time the Internet consisted of approximately 60,000 hosts. This worm infected approximately 10% of the hosts and caused significant outages and slowdowns of mail servers across the net. In July of 2001 a new worm infection appeared that would significantly raise awareness of the threat posed by these malicious software programs along with the dramatic landscape change of the Internet. An estimated 650 million hosts are today connected to the Internet hence a fundamental shift in the potential number of participants to propagate a worm. CodeRed spread quickly and became the most widespread and damaging worm to hit the Internet since the Morris Worm. An estimated total of 360,000 hosts were infected within a period of 14 hours. Two months after CodeRed another large-scale worm named NIMDA (ADMIN spelled backwards) impacted the Internet. More recently, the Internet saw the appearance of a new type of worm that infected the Internet at such a high rate that it was classified as a flash worm. The fast scanning rate of SQL Slammer in January 2003 was achieved because of its small size (single packet of 376 byte) as well as the fact that the worm was not TCP but UDP based (connectionless). SQL Slammer reached its full scanning rate of 55 million scans/sec within 3 minutes of the start of the infection and infected the majority of vulnerable hosts on the Internet within 10 minutes of the start of the infection with an estimated 250,000 – 300,000 infected hosts overall. Summer 2003 witnessed the infamous Blaster and January 2004 was the turn of MyDoom to impact Internet users. While the underlying exploits used to achieve access to the target hosts varied between these worms the methods and technologies used to mitigate and contain the infection remained the same. In order to protect the network from these threats, the security system must be able to protect and react against both known and unknown attacks. This calls for an integrated security solution that is both flexible and pervasive, providing tighter collaboration between network services, security services, hosts, applications, management and business processes. As worms typically invade an environment in a multi-phased approach, this layered structure is an effective way to protect networks from these threats. There are six steps involved in a worm mitigation methodology, in order: preparation, identification, classification, trace back, reaction, and post-mortem. The reaction phase can broken down into containment, inoculation, quarantine, and treatment. Worm mitigation requires coordination between system administration, network engineering, and security operations personnel. This is critical in responding effectively to a worm incident. The containment phase involves the limiting of the spread of a worm infection to those areas of the network already affected. With the worm infection contained, or at the least, significantly slowed down, the inoculation process further deprives the worm of any available targets. The mobile environment prevalent on networks today poses significant challenges since laptops are routinely taken out of the “secure” environment and connected to potentially “insecure” environments such as home networks. A laptop can be infected with a worm or virus and then bring it back into the “secure” environment where it can infect other systems. The quarantine phase involves tracking down and identifying infected machines within the contained areas and disconnecting, blocking, or removing the infected machines. This isolates these systems appropriately for the final phase. During the treatment phase actively infected systems are disinfected of the worm. This can involve simply terminating the worm process and removing any modified files or system settings that the worm introduced, and patching for the vulnerability the worm used to exploit the system. In other cases a complete re-install of the system may be warranted in order to confidently ensure that the worm and its byproducts are removed.
<urn:uuid:df0bdd23-6b25-4fff-a659-f86a5dafaba4>
CC-MAIN-2024-38
https://www.helpnetsecurity.com/2004/05/10/combating-internet-worms/
2024-09-10T11:33:27Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651241.17/warc/CC-MAIN-20240910093422-20240910123422-00484.warc.gz
en
0.949478
947
3.15625
3
During Cyber Security Awareness Month we hear about how important being #CyberAware is. "Do your part! Be careful!" It's important to take this seriously. But additionally, this field offers job opportunities. You Need a Multi-Part Foundation Cyber security builds on fundamental knowledge and skills. What does it take to harden a server, then a network, and finally an entire ecosystem of connected computers? Much of it comes from being careful and complete. Three Pillars of Cyber Security The field relies on three main technical areas. They are networking, operating systems, and, finally, cryptography. You need to understand those pieces, and then see how they fit together. Let's look at them one at a time. You hear about networking protocols. Those are formal rules for communication across a network. A device must follow these rules to send data. What's more, the rules define how to respond to incoming data. There are many protocols. They accomplish different tasks at different layers. First, Ethernet transmits data frames across a local-area network. One hop, over a short distance. Then, IP gets a packet to a specific host on the other side of the world. Many hops, around the world. Later, UDP delivers a short message to a specific program on a host. Also, TCP connects a two-way data stream to a specific program. Finally, application-layer protocols do what we users think about. "Show me that web page." "Next, store this file." "Now, connect me to the database." The protocols define a language. In human terms, it's like "Here's how to speak French", and also, "When someone speaks French, here how to understand them." However, what if they're not really following the rules? Maybe you've switched the DVD audio to French and Jean Reno is using some tough-guy slang. It's French, sort of. But now you may become confused. On a network, you see, bad guys may send slightly incorrect or inappropriate data. Maybe they're trying to avoid detection. Or, maybe they hope to slip data past firewall rules. Know the network language. Certification exams like CompTIA Security+ and ISC2 CISSP and CCSP will assume you're fluent. How can you get up to speed? Learning Tree's Introduction to Networking Training course is a great overview. If you don't have the time or training budget, try self-study with an older edition of Internetworking With TCP/IP by Doug Comer. Now, what are we connecting? System administration builds on top of user skills. First, become an adept user. Later, you can advance to running servers. Learn PowerShell on Windows. Good system administrators use PowerShell to automate tasks, because it is both a command-line shell and a scripting language. For Linux and other UNIX-family operating systems, learn the fundamentals of system administration. Learning Tree's Linux Administration and Support Training builds on top of user command-line skills learned in the Introduction to Linux Training. Then, make sure you can tie the operating system to the network. Let's say you take the CompTIA Security+ exam. And, let's say you haven't taken Learning Tree's CompTIA Security+® Training course. You may have an unpleasant surprise. Security+ now shows you network command output. Next, it asks you to diagnose the problem causing what you see. This surprises people, because CompTIA's study books don't mention this. Many people taking Security+ are blindsided by the questions about IP configuration. You see, it's those network protocols or rules plus command output specific to both Windows and Linux. Next, things get cryptic. Cryptography is the third major component of cyber security. It contains several parts. Its three basic components are hash functions, symmetric encryption, and asymmetric (or public-key) encryption. Next, combine a hash with asymmetric encryption. Now you have a digital signature. Next, wrap a public key inside a digital signature by a trusted third party. Now you have a digital certificate. Web security requires digital certificates. You must trust these in two ways. First, you must have confidence in the mathematics -- did we use an appropriate hash and cipher? Second, you must trust the issuer, the Certificate Authority -- are they careful and ethical? Things change in cryptography. For a while, everyone used RC4. It's a fast stream cipher. But then cryptographers found weaknesses. Therefore, we quit using RC4. In fact, we had no trustworthy stream cipher for a while. But now, new strong stream ciphers have come along. Everyone has reconfigured their servers again and again over the past few years, because our understanding of what is secure has changed. While cyber security is a challenge, it can be rewarding. What's more, there's job security. Something always needs fixing! Consider the career path. However, don't start down the path unprepared.
<urn:uuid:485f42cc-fdbb-4db6-ab57-dc2cc3d85f3e>
CC-MAIN-2024-38
https://courses.learningtree.com/blog/how-to-enter-the-cybersecurity-field/
2024-09-11T17:34:06Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651390.33/warc/CC-MAIN-20240911152031-20240911182031-00384.warc.gz
en
0.940964
1,037
3.25
3
Definition: Beacon Protocol The Beacon Protocol is a communication framework designed to facilitate the exchange of signals and messages between devices, particularly in wireless networks. This protocol is critical in enabling devices to discover, connect, and maintain communication within a network, ensuring efficient and reliable data transmission. Introduction to Beacon Protocol The Beacon Protocol plays a pivotal role in modern wireless communication, serving as the backbone for various technologies such as Bluetooth Low Energy (BLE), Wi-Fi, and Internet of Things (IoT) devices. By broadcasting periodic signals, beacons allow devices to advertise their presence and provide essential information, enabling other devices to establish and maintain connections. Key Components and Functionality At the core of the Beacon Protocol are the beacon signals, which are periodic transmissions sent by devices to announce their presence and capabilities. These signals typically contain a unique identifier and relevant metadata, such as signal strength and available services. Devices within range can detect these signals, enabling them to initiate communication or take specific actions based on the received information. Discovery and Advertisement One of the primary functions of the Beacon Protocol is to facilitate the discovery of devices within a network. When a device sends out a beacon signal, it effectively advertises itself to other devices. This advertisement can include various details, such as the device’s identity, available services, and connection parameters. Devices scanning for beacons can then respond to these advertisements, initiating the process of establishing a connection. Once a device detects a beacon signal and decides to connect, the Beacon Protocol outlines the steps for establishing a secure and reliable connection. This process often involves exchanging authentication and encryption keys, negotiating communication parameters, and ensuring both devices are ready to transmit data. The protocol ensures that the connection is robust and capable of handling the intended data exchange. Maintenance of Connection After establishing a connection, the Beacon Protocol continues to play a role in maintaining the communication link. It manages tasks such as monitoring signal strength, handling disconnections and reconnections, and ensuring data integrity. By continually broadcasting beacon signals, devices can maintain awareness of each other’s presence and adjust their communication strategies as needed. Benefits of Beacon Protocol The Beacon Protocol significantly enhances connectivity by providing a standardized method for devices to discover and communicate with each other. This leads to more efficient use of network resources and reduces the likelihood of connection failures or interference. Improved Device Interaction By enabling devices to advertise their presence and capabilities, the Beacon Protocol fosters improved interaction between devices. This is particularly important in IoT ecosystems, where seamless communication between sensors, actuators, and control units is essential for optimal performance. Beacon signals are designed to be low-power, making the Beacon Protocol an energy-efficient solution for wireless communication. This is especially beneficial for battery-powered devices, such as IoT sensors and wearables, where conserving energy is crucial for prolonged operation. The Beacon Protocol supports a scalable approach to network communication. As more devices are added to a network, the protocol can efficiently manage increased traffic and maintain reliable connections without significant degradation in performance. Uses of Beacon Protocol One of the most prominent uses of the Beacon Protocol is in indoor navigation systems. Beacons placed throughout a building can provide real-time location data to users’ devices, enabling accurate navigation and location-based services. This technology is widely used in malls, airports, and museums to enhance the visitor experience. In retail environments, the Beacon Protocol enables proximity marketing, where businesses can send targeted advertisements and offers to customers’ smartphones based on their location within the store. This personalized approach increases customer engagement and drives sales. The Beacon Protocol is extensively used in asset tracking applications. By attaching beacons to valuable items, businesses can monitor their location in real-time, reducing the risk of loss or theft. This is particularly useful in logistics, warehousing, and healthcare industries. In smart home environments, the Beacon Protocol facilitates communication between various devices, such as lights, thermostats, and security systems. This enables automated control and enhances the overall efficiency and convenience of the smart home ecosystem. Event organizers use the Beacon Protocol to manage large gatherings effectively. Beacons can provide attendees with real-time information, such as event schedules, location-based notifications, and emergency alerts. This improves the overall event experience and ensures safety and organization. Features of Beacon Protocol Low Energy Consumption One of the defining features of the Beacon Protocol is its low energy consumption. Designed to operate with minimal power, beacon signals extend the battery life of devices, making the protocol suitable for applications requiring long-term deployment. The Beacon Protocol offers high accuracy in location-based services. By leveraging signal strength and triangulation techniques, devices can determine their position with precision, making the protocol ideal for navigation and tracking applications. Security is a critical aspect of the Beacon Protocol. It incorporates robust encryption and authentication mechanisms to ensure that data exchanged between devices is protected from unauthorized access and tampering. The Beacon Protocol is designed to be interoperable with various wireless technologies, such as Bluetooth, Wi-Fi, and Zigbee. This ensures that devices from different manufacturers can communicate seamlessly, fostering a diverse and flexible network environment. Customizable Broadcasting Intervals The protocol allows for customizable broadcasting intervals, enabling devices to adjust the frequency of beacon signals based on specific needs and power constraints. This flexibility enhances the protocol’s adaptability to different use cases. How to Implement Beacon Protocol Selecting the Right Beacons Implementing the Beacon Protocol begins with selecting the appropriate beacons for the intended application. Factors to consider include signal range, battery life, form factor, and compatibility with the target devices and infrastructure. Once the beacons are selected, they need to be configured to broadcast the necessary information. This involves setting up unique identifiers, defining broadcast intervals, and ensuring the beacons are positioned optimally for effective signal transmission. Integrating with Applications To leverage the Beacon Protocol, applications must be developed or adapted to interact with beacon signals. This involves implementing functionality for detecting beacons, processing the received data, and taking appropriate actions based on the information. Testing and Deployment Before full-scale deployment, it is crucial to test the beacon network to ensure reliable performance and coverage. This includes verifying signal strength, connectivity, and the accuracy of location-based services. Once testing is complete, the beacons can be deployed across the desired area. Monitoring and Maintenance Ongoing monitoring and maintenance are essential to ensure the continued effectiveness of the Beacon Protocol. This includes regularly checking beacon battery levels, updating configurations as needed, and addressing any connectivity issues that may arise. Frequently Asked Questions Related to Beacon Protocol What is the Beacon Protocol? The Beacon Protocol is a communication framework that facilitates the exchange of signals and messages between devices in wireless networks. It is essential for enabling device discovery, connection, and communication maintenance, ensuring efficient data transmission. How does the Beacon Protocol enhance connectivity? The Beacon Protocol enhances connectivity by providing a standardized method for devices to discover and communicate with each other, leading to more efficient use of network resources and reducing the likelihood of connection failures or interference. What are the primary uses of the Beacon Protocol? The Beacon Protocol is used in various applications such as indoor navigation, proximity marketing, asset tracking, smart homes, and event management. It enables real-time location data, targeted advertisements, and seamless communication between smart devices. What features make the Beacon Protocol energy-efficient? The Beacon Protocol is energy-efficient due to its low-power beacon signals, which extend the battery life of devices. This makes it suitable for applications requiring long-term deployment, such as IoT sensors and wearables. How can the Beacon Protocol be implemented? Implementing the Beacon Protocol involves selecting the right beacons, configuring them, integrating with applications, testing, and deploying the beacon network. Ongoing monitoring and maintenance are also essential to ensure reliable performance and connectivity.
<urn:uuid:67a82080-d25b-44bb-9d5d-3facfb7f8d60>
CC-MAIN-2024-38
https://www.ituonline.com/tech-definitions/what-is-beacon-protocol/
2024-09-12T22:09:17Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651498.46/warc/CC-MAIN-20240912210501-20240913000501-00284.warc.gz
en
0.914934
1,621
3.953125
4
Let’s start with the important question: Do you understand and trust your AI? As Evert Haasdijk, Senior Manager and member of Deloitte’s AI Centre of Excellence bluntly puts it: “The board room and higher management of a company are often not really aware what developers in the technical and the data analytics departments are working on. They have an idea, but they don’t know exactly. This causes risks for the company.” To make progress, business leaders across all industries must take the necessary steps to understand the inner workings of their AI model. There are several ways in which AI transparency can propel a business forward. With many public examples of AI decision-making going awry, public anxiety regarding the use of machine learning is on the rise. From legal concerns regarding the use of personal data, to ethical concerns on the matter of fairness, the influence of AI models on high-impact decisions is under continuous scrutiny. After all, giants like Apple and Amazon have unknowingly deployed biased AI models to make processes such as credit application and recruitment more efficient. The case of COMPAS, a tool used by US courts to predict the potential risk of recidivism in defendants, is notorious. Machine Learning bias is a complex problem that stands at the crossroads between mathematical analysis and the social implications of determining what is ‘fair’ in decision-making. In these circumstances, the expectation is for AI to produce transparent and understandable results that reflect the ethical standards of society. Transparent AI allows businesses and organizations to offer meaningful and understandable explanations that humans can trust. On top of building trust, insights gained by means of AI transparency can be used to identify weak points and failure cases that can be subsequently used to improve the accuracy and robustness of machine learning models. Machine learning models can pick up spurious correlations and appear to be making the correct decisions, but for the wrong reason. A famous example is a model trained to distinguish between photos of wolves and huskies, which learned to predict “wolf” whenever it saw snow in the background, just because in the training dataset most pictures of wolves had snow, while most pictures of huskies did not. As such, AI transparency promotes continuous optimization and enables a symbiotic relationship to be formed between the human and the machine. As machine learning is becoming a core element in the value proposition of organizations worldwide, the number of threats is also increasing. Malicious attackers often target AI models by means of adversarial attacks. Be it in the form of poisoning, evasion, or model extraction, an adversarial attack can lead to partial or irrevocable damage, from data theft up to complete model decay. By means of AI transparency, business leaders, and data scientists could combat the potential threat of malicious attacks. Big industry names such as Google, Microsoft, and IBM have already started to invest both in developing AI models, but also in securing them against adversarial attacks. Implementing the necessary security measures for AI models should be a crucial step in every business’s defense strategy. Due to the potential biases that AI models may include in their decision-making processes, many countries around the globe have started implementing AI strategies and policies meant to ensure the ethical use of algorithms. The OECD AI Policy Observatory provides a repository of over 700 policies initiated by 60 countries. Perhaps some of the best-known legal frameworks that businesses and organizations employing AI models need to take into consideration are: As Machine Learning technologies are becoming essential to advancing in the digital landscape, AI regulations are here to stay. The question is: Is your business prepared for them? Embracing AI has gained a new nuance of urgency, especially in the context of a post-pandemic economy. A major study of 900 senior executives conducted by HFS Research in conjunction with KPMG shows how AI has become crucial to the future survival of businesses. Fulfilling the positive potential of AI is ultimately only possible by opening the black box and experimenting toward success. It is only through AI transparency and explainability that we can give Machine Learning strategies purpose across all the important dimensions, those of trust, performance, security, and compliance. Lumenova AI can help companies understand the inner workings of their AI models, so they can efficiently mitigate bias and ensure compliance with new and emerging algorithmic regulations. To request a demo, please get in touch with us.
<urn:uuid:69ed6b72-de1a-45a3-a8a9-9f7a7f79c3cb>
CC-MAIN-2024-38
https://www.lumenova.ai/blog/business-leaders-should-care-ai-transparency/
2024-09-14T04:30:04Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651548.18/warc/CC-MAIN-20240914025441-20240914055441-00184.warc.gz
en
0.951899
911
2.796875
3
A DNS Attack is any attack targeting the availability or stability of a network’s DNS service. Attacks that leverage DNS as its mechanism as part of its overall attack strategy, such as cache poisoning, are also considered DNS attacks. In this article, we will get an overview of the common types of DNS attacks out there. Generic Attacks Against DNS Service These attacks focus on attacking the DNS infrastructure itself, either rendering the DNS service itself unavailable or subverting the answers provided by the DNS servers. Keep in mind that DNS is comprised of two (2) separate components, authoritative servers (answer-hosting) and recursive servers (answer-finding), and there are tailored attacks against each component that will be discussed later. Like any other server, DNS servers are prone to all network-based attacks. There are many ways attackers can cause a large amount of network traffic to the DNS servers, such as TCP/UDP/ICMP floods, rendering the service unavailable to other network users by saturating the network link to the DNS servers. Attackers can also leverage a specific vulnerability to the DNS server software or host operating system, to either bypass control measures to create rogue entries in the DNS database, or cause the DNS server to crash. Attacks Against Authoritative Servers Authoritative name servers maintain the DNS zone and records, similar to a database. These are some common attacks against authoritative DNS servers. DNS data by design is supposed to be for public consumption, making this the ideal first step for an attacker trying to learn more about a target environment. This attack does not directly impact service availability or stability, but it is usually part of a long-term strategy of a larger attack. For example, an attacker could deduce that exchange.example.com listed in the MX record is running Microsoft Exchange, and launch specific attacks against it. Authoritative name servers can accept dynamic updates, meaning they can essentially create new DNS records on the fly. However, this feature could be exploited by attackers to sneak unauthorized entries into the DNS zone. This is a type of DoS attack, and its goal is to overwhelm the authoritative name servers to the point that it can no longer respond to legitimate queries. In this attack, the attacker sends a lot of queries for subdomains that probably do not exist, consuming the authoritative server’s resources to the point that it causes disruption to other DNS lookups. For example, instead of querying for example.com, the attacker would query for 111aaa.example.com, 111bbb.example.com, 222aaa.example.com, 333ccc.sub.example.com, etc. Attacks Against Recursive Servers The main job of a recursive server is to build and hold a rich cache of DNS answers. Cache poisoning aims to corrupt the answers stored in the cache, so any subsequent lookup from other clients will get the corrupted answer. This is discussed in Cache Poisoning. NXDOMAIN and Phantom Domain Similar to the subdomain attack against authoritative servers, this attack queries recursive name servers that are known to not exist. This will waste the recursive server’s time in walking the DNS namespace, only to reach the conclusion that the name does not exist, filling up the cache with useless answers. Other DNS-based Attacks and Exploits Amplification + Reflection Attack Previously discussed DDoS attacks are the variety where attackers target DNS servers within an organization. These types of attacks we address now is where attackers take over DNS servers within an organization as part of a DDoS attack to target someone else. It is discussed in more details in DDoS. Domain Hijacking and Redirection This category of attack subverts the users to go to a different destination. A well-known example is the domain name paypaI.com (notice the letter “i” is in uppercase), which looks very much like the real domain name paypal.com, spelled with an L. In a similar category, attackers could infect the target client machine with malware that changes the local DNS settings, so that all DNS requests are sent to the DNS server under the attacker’s control, such as the case for the DNSChanger worm. Speaking of malware, many types of malware today are using DNS as part of their overall function to not only communicate with the command-and-control server but to update and evolve itself. A prime example is the recent WannaCry ransomware, which relies on making an initial successful DNS query before it executes its attack. To learn more about Malware, click here. Data Exfiltration and Tunneling DNS Tunneling is a general technique that encodes messages in DNS queries and answers, mostly to evade detection. While there are legitimate uses of DNS Tunneling, where it gets serious is when someone uses it to exfiltrate sensitive information out of the target environment. This is extremely difficult to detect, due to the ever-changing domain names, and the encoding-decoding schema chosen.
<urn:uuid:d4c20635-8750-4265-837a-b4dd5dd3013a>
CC-MAIN-2024-38
https://www.infoblox.com/dns-security-resource-center/dns-security-faq/what-are-dns-attacks/
2024-09-15T11:05:18Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651622.79/warc/CC-MAIN-20240915084859-20240915114859-00084.warc.gz
en
0.923807
1,032
3.234375
3
Historic technology firsts changed the world and inspired the future. The course of innovation may not always run smooth, but trailblazers never lose their drive: Technology has come a long way from the first computing machines and the days of Alexander Graham Bell, but the spirit that drives the world’s greatest innovators never slows. Take a road trip down history lane: California, 1971 - Pioneering the Personal Computer - Long before the rise of Silicon Valley, computer engineer John Blankenbaker put the California tech industry on the map—and transformed the future of computing—when he developed the first affordable personal computer in 1971. Illinois, 1973 - Can You Hear Me Now? - In 1973, Motorola engineer Martin Cooper—inspired by the always-connected, work-anywhere communication devices on Star Trek—designed a two-and-a-half pound cell phone at his workspace in Chicago. New York, 1979 - Mission to the Moon - When Apollo 11 lifted off from Florida in 1969, it took with it the incredibly complex lunar module that would land Neil Armstrong and Buzz Aldrin on the moon—a marvel of engineering (and modern connectivity) designed in Bethpage, New York. New Hampshire, 1972 - Game On - End your journey in 1972 in New Hampshire, where Magavox released the Odyssey video game system—the first multiplayer, multiprogram console—designed by Ralph Baer. Washington, 1976 – Fast Company – After budget cuts forced computer designer Seymour Cray to halt work on a supercomputer he was designing for his employers, he started his own business to continue working on the project. Cray Research released the Cray-1 supercomputer in 1976, setting a new standard for computing speed. Ohio, 1974 – Setting the Bar – More than two decades after Joe Woodland filed a patent for the barcode—an information storage design inspired by his days learning Morse code as a Boy Scout—a supermarket cashier in Troy, Ohio, scanned the first item with a UPC label: a pack of Wrigley’s Gum. The advancements of the last 50 years have fueled achievements and milestones in every tech sector. Where are you going next? With HPE solutions—including the cost-optimized hybrid cloud services of GreenLake, the guaranteed-available storage of your data, and proven as-a-service virtual desktop infrastructure—your business can leave its own mark on history. See resources that may help steer you in the right direction below. Questions? Chat in lower right or email us. Prepare for your own landmark achievements with HPE solutions that centralize management of on-premises and cloud data; store and protect your essential information; and enable seamless remote work environments. AI-driven, as-a-service storage built for the cloud securely stores, protects, and archives your data—allowing faster insights, optimized operational costs, and guaranteed availability. Pick up the pace of your digital journey with pay-as-you-go GreenLake hybrid cloud services that scale up and down according to your needs, prevent overprovisioning, and centralize apps and data across all your locations. As-a-service virtual desktop infrastructure empowers remote workers to stay connected, secure their critical data and programs, and collaborate smoothly, consistently, and efficiently. Shave essential hours off your IT workload with reliable support services that allow your team to focus on the finish line with a 65 percent shorter time to project deployment. Whether you’re looking to centralize your apps and data, unleash the power of your data, or improve productivity, join HPE on the road to innovation. A major grocery chain was concerned about making it easy for employees to connect at home securely. Our solution addressed their challenge quickly, helped them scale, and was easy to support. When a large university needed high-performance computing (HPC) cluster manageability tools, they leaned on us to provide a solution that would be “easy to use and manage” by their staff. Are you concerned with disaster preparedness and business continuity? See how TCU stood up active-active data centers with zero downtime with HPE 3PAR and VMware in this case study. Are you looking for a way to simplify your IT purchases? Mobius Value Portal (MVP) has the ability to be custom tailored to your company’s needs and to automate the sales and procurement process. get started today Cookie | Duration | Description | cookielawinfo-checkbox-analytics | 11 months | This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics". | cookielawinfo-checkbox-functional | 11 months | The cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional". | cookielawinfo-checkbox-necessary | 11 months | This cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary". | cookielawinfo-checkbox-others | 11 months | This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other. | cookielawinfo-checkbox-performance | 11 months | This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance". | "*" indicates required fields
<urn:uuid:082736d6-4f1e-47da-97c9-4c1ca64bf0e7>
CC-MAIN-2024-38
https://mobiuspartners.com/roadtrip/
2024-09-17T17:57:13Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651829.57/warc/CC-MAIN-20240917172631-20240917202631-00784.warc.gz
en
0.914052
1,146
2.796875
3
Scalar functions receive an input value, and based on this input value the result is delivered. This function can be used with constant values, the column elements of a table (view) as well as with compound value expressions. Scalar functions usually expect a special data type for their arguments. If it is not specified, then an implicit conversion of the data type is attempted, or an error message is displayed. Numeric functions are given a numeric value as input and typically deliver a numeric value as output. Exasol supports the following numeric functions: String functions can either return a string (for example, LPAD) or a numeric value (for example, LENGTH). Exasol supports the following string functions: Date/Time functions manipulate the DATE, TIMESTAMP, TIMESTAMP WITH LOCAL TIME ZONE, and INTERVAL data types. Exasol supports the following date/time functions: To analyze geospatial data, there are many functions available. For more information on geospatial data, refer to the Geospatial Data section. Bitwise functions can compute bit operations on numerical values. Conversion functions can be used to convert values to other data types. Functions for Hierarchical Queries The following functions can be used in combination with CONNECT BY queries. JSON functions extract values from JSON objects stored as strings. Functions that cannot be allocated to one of the above categories are listed here.
<urn:uuid:8ca55384-e44a-47ef-b434-5f01a9c9db3a>
CC-MAIN-2024-38
https://docs.exasol.com/saas/sql_references/functions/scalarfunctions.htm
2024-09-18T23:49:09Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651944.55/warc/CC-MAIN-20240918233405-20240919023405-00684.warc.gz
en
0.796326
305
2.703125
3
When most of the world was on lockdown during COVID-19, Earth seemed to appreciate the time off from commuting. The environmental impact of remote work showed that carbon dioxide production was temporarily lowered by 17 percent with reduced surface traffic. During the same period, there was also a 67 percent decrease in commercial waste production and a 14 percent drop in industrial electrical demand during the same period of time. Unfortunately, these figures have bounced back to pre-pandemic levels. However, with a little conscious effort, hybrid and remote work can have positive green consequences. Commuting is one of the most significant pollutants globally. If every employee who wanted to could work from home half the time, that would be the greenhouse gas equivalent of taking the entire New York State workforce off the road. That’s a big breath of fresh air; there are many other ways to contribute to a green Earth. How can we build more sustainable work practices and trends into our work-from-home habits for Earth Day? First, let’s take a quick look at the history of Earth Day. A Brief History of Earth Day This year marks the 54th anniversary of Earth Day, established on April 22 in 1970. In an era where cars ran on leaded gas, and industry produced unregulated smoke and toxins, Earth Day was a call to save the deteriorating environment — that first Earth Day inspired 20 million Americans — 10 percent of the national population – to demonstrate against industrial development, ecological waste, and harmful pollutants. Within a year, the bipartisan message of Earth Day led to the formation of the Environmental Protection Agency (EPA) and the passing of critical environmental laws that still exist in some form today, including the Clean Air Act (CAA) and the Occupational Safety and Health Act (OSHA). By 1990, Earth Day was officially a global movement, with 200 million people worldwide working together to save the planet. Over one billion people now participate in activities to restore nature yearly, from cleanups to planting trees. The theme for Earth Day 2024 is “Planet vs Plastics,” explicitly calling global citizens to a “60% reduction in the production of ALL plastics by 2040” in a commitment to human and planetary health. Companies that actively develop robust Environmental Social Governance (ESG) standards don’t just have happier employees, better financials, and sustainable business practices — they’re working to make a more viable, profitable future for everyone through green initiatives. Earth Day teaches us that whether you’re a multinational corporation or an individual working from home, everyone’s choices can have a positive environmental impact Here are 10 green workspace tips showing how a small investment in eco-conscious work habits can significantly impact the planet. 10 Tips for Going Greener While Working from Home 1. Unplug your electronics. Did you know that the wasted energy drawn by devices like idle monitors, laptop chargers, television screens, and printers could have powered 11,000,000 homes? So unplug your computer at the end of the day, or ensure everything is plugged into a smart power strip that stops the drain. 2. Swap out your light bulbs. Replacing old incandescent light bulbs with LEDs uses at least 75 percent less energy and can last up to 25 times longer. You can also choose from a range of warmth and brightness options to save money and make sure you look your best on your next video call. 3. Let the fresh air in. Turn off your air conditioning and open the windows if the weather's nice. You can encourage maximum air circulation (and cooling!) by opening windows opposite each other or setting up fans for cross-ventilation. 4. Check your utilities. If you’re curious about how green your energy is, the EPA’s Power Profiler breaks down how your local electricity is made. Some bills even itemize your usage and show times where levels peak. Many providers offer customers the option to select clean energy generated from renewable sources if you want to take the next step in making your home more sustainable. 5. Skip the printer. One of the nice parts about remote work is that most documents are easily shared and accessed in the cloud — no more piles of paper to shuffle through. But if you need to print something, recycle it after reading and adequately dispose of ink cartridges. 6. Upcycle your office furniture. If you’ve been itching for a change of scenery, don’t head straight for the nearest big-box store and dump your old desk on the curb. Resale locations like Goodwill and Habitat for Humanity Restore often have high-quality, gently used furniture — and will take your old pieces off your hands and out of the junkyard. 7. Take stock of your trash. If you’re tossing all your waste directly in the bin, now’s the time to reassess whether something should be destined for the landfill. Every county’s rules for recycling and composting are different, but here are some basic rules for upping your recycling game. 8. Eat (really) local. You're not alone if you’ve been relying on delivery services for lunch. However, even short delivery drives contribute to greenhouse gases, and takeout packaging is often wasteful. It’s never too late to learn how to make simple (but tasty) meals in your kitchen. Bonus points if you’ve been getting your hands dirty in the garden or a planter box. Nothing tastes better — or has a smaller carbon footprint — than something you’ve grown yourself. 9. Meet up online. You may be sick of virtual meetings by now, but they’re here to stay. They’re also the most eco-friendly way to connect with someone over long distances. Even better news is that the right communications software helps make it feel like you’re in the room together. 10. Take a walk. For some, the car has become a mobile office, whether taking calls on the road or simply waiting to pick up their kids after school or practice. Unfortunately, this offsets the benefits of a reduced commute and puts more carbon dioxide into the atmosphere. Weather permitting, wear headphones and take a walking meeting with your smart phone. A change of pace and a breath of fresh air can do wonders for you and the environment. A Sustainable Future with Mitel This Earth Day and every day, we are committed to protecting the environment and reducing pollution. As a leader in unified communications, we provide our partners and customers with the remote collaboration tools they need to work sustainably. Mitel collaboration solutions allow companies to invest in their employees for the planet's wellness.
<urn:uuid:ed109129-1598-4680-9acf-a05d784979df>
CC-MAIN-2024-38
https://www.mitel.com/en-ca/blog/earth-day-tips
2024-09-19T00:17:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651944.55/warc/CC-MAIN-20240918233405-20240919023405-00684.warc.gz
en
0.930988
1,387
3.09375
3
What is DNS Server? How many of you can answer this question i.e. What is DNS Server ? On which port DNS Server runs ? DNS uses TCP or UDP port ? Answer the above mentioned question in the comment section given below before you start reading more of this article. What is DNS Server ? DNS is known as Domain Name System. It is primarily used for name resolution. As computer understands only numbers and not names but it is difficult for human minds to remember numbers. Therefore DNS comes in the picture, it is used to convert the Domain name i.e. FQDN (Fully Qualified Domain Name) to IP Address. Forward Lookup zone is used in DNS for home name to IP address resolution and Reverse DNS Lookup is used to resolve IP Address to Host name. By default Forward Lookup zone gets created automatically and creates all the Host A records when new computers or Servers are added to DNS. Alternatively Reverse DNS Lookup is not created automatically, it needs to be created manually to resolve IP address to Host name. Why do we need DNS Server? Before I explain about DNS, let me ask you about Google. How many of you have visited Google.com, I hope everyone. Next questions is, how often do you visit Google.com. Some of us visit once in day, twice in a day or multiple times in a day. Now the next question, do you know the IP address of Google.com?. No problem if you don’t know the IP address of Google, you can tell me an IP address of Facebook. That’s ok, if you don’t know IP address of Google or Facebook or don’t remember them because human mind is capable of remembering names however remembering IPs or numbers is difficult. Alternatively, computers only understand IPs i.e. numbers and not names. Therefore, Domain Name System comes into the picture. It keeps record of all the domains and map hostname to IP address. When you type Google.com in your browser, DNS converts this name to an IP address and then communicate with Google. Hope your understanding is clear now. Let’s take another example to understand DNS in detail. DNS record is like a contact that we create in our Mobile Phone. When you create a contact in your Mobile, you create a name of person and then map it with their mobile number. When you dial their name, in the background it converts the name to their mobile number and then it dials. Similarly DNS resolves Computer name to IP address and then initiate communication. DNS Port Number On which Port DNS Server runs? DNS Server runs on port number 53. Another very important question is, out of TCP and UDP which protocol is used by DNS and when. DNS server uses both TCP and UDP protocol. TCP is known as connection oriented protocol which requires acknowledgement and UDP is connection less protocol which doesn’t require acknowledgement. DNS uses primarily UDP for name resolution. Whenever client computer connects server for name resolution, it doesn’t required TCP protocol and it uses UDP protocol. Alternatively when secondary DNS server connects primary DNS server for zone transfer then TCP protocol is used and not UDP because it requires acknowledgement. How to Install DNS in Windows Server 2012 R2 In this article, we’ll learn the steps to install DNS Server in Windows Server 2012 R2. As stated earlier it is used to translate Domain Name into IP address and IP address into domain name. AD Integrated DNS would get installed and configured automatically when you are promoting a Domain Controller. However you can still have DNS in your environment even if you don’t have Domain Controller in your environment. Let’s begin with the steps to install DNS in Windows Server 2012 R2 1. Open Server Manager. 2. Click on ‘Add roles and features’. 3. On “Add Roles and Features Wizard” we’ll verify all the prerequisites like administrator account has a password, IP address configured etc. should be completed before we install DNS on this box. To continue, click Next. 4. Select ‘Role based or feature based installation’ and click Next. 5. On “Add Roles and Features Wizard” we have the option to select the destination server i.e. In Server Pool all the remotely managed servers listed on which role can be installed, here DC06. 6. On ‘Add Roles and Features Wizard’ check on DNS Server, then another window opens for adding features that are required. Click on Add Features and then click on next to continue. 7. Here, we can verify that if role is selected to install or not. Click next to continue. 8. All the essential features required with AD DS role are already selected. We don’t need to do any modification in this window. Click Next to continue. 9. Additional information can be seen about DNS. Click next to continue. 10. In “Confirm Installation selections” windows, we can see all the roles and features that we have selected to install. In case of any changes, click on previous button to go back and make the changes. Select the option “Restart the destination server automatically if required”, it will restart the computer if required. Click on Install to continue. 11. In “Installation Progress” windows, the installation is in progress when installation is completed click on close to close the wizard. If you close the wizard while installation is in progress, it will continue the installation in background. 12. To verify if the DNS is installed or not. Click on start button and then click on down arrow. 13. DNS icon confirms that DNS is installed successfully. DNS console can also be opened by typing “DNSMGMT.MSC” in command prompt or in run. Hope you understood all the steps that we have explained in this article. To summarize, we’ll explain that DNS Server is used for name resolution. It is used to resolve Host name to IP Address and IP Address to Host name.
<urn:uuid:6bfbbc41-f79e-490a-a28e-160375d29ff1>
CC-MAIN-2024-38
https://itingredients.com/install-dns-server-in-windows-server-2012-r2/
2024-09-20T08:13:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652138.47/warc/CC-MAIN-20240920054402-20240920084402-00584.warc.gz
en
0.907362
1,270
3.328125
3
Server Virtualization Complete Guide If you want to save on IT costs while maximizing the existing IT infrastructure resources, server virtualization could be worth your consideration. This technique of deploying multiple server applications on one physical system has seen broader adoption in the market, and it's proving quite beneficial for small and large businesses alike. Below, we discussed more on server virtualization, from what it is, its types, pros, cons, and everything in between. What is Server Virtualization? Server virtualization refers to the process of creating servers, infrastructures, services, and multiple computing resources on one virtual platform. Initially, computer software and hardware were designed to support single applications. As a result, servers were forced to process one task at a time, which led to wastage of memory capacity and unused processors. So as additional apps and services were deployed across the organization, the number of servers grew exponentially. Data centers were stretched to their limits by the rising prices and increased demand for space, power, cooling, and maintenance. With the emergence of server virtualization, all of that changed. Here, one physical server is divided into numerous individual and remote virtual spaces, each serving various users. In other words, virtualization adds an extra layer of software to a computer, which controls the computer's virtualized resources, dividing them into logical instances known as virtual machines. The latter is capable of functioning independently. This minimizes the enterprise's IT costs by reducing server counts, easing the burden on data center resources, and improving IT flexibility. How Does Server Virtualization Work? Server virtualization works by extracting /isolating a hardware component from any software in which it is installed. A hypervisor - a specialized software- is used to provide this abstraction. The hypervisor identifies the computer's physical resources, such as CPUs, memory, storage volumes, and network interfaces, and creates logical aliases for them. The true strength of a hypervisor lies in what can be done with those isolated resources, not in its abstraction. In other words, a hypervisor creates logical models of computers or Virtual Machines (VMs) using virtualized resources. Each VM functions as a full-fledged computer. Once a VM is set up, it requires the installation of a full suite of software, including an operating system, drivers, libraries, and, finally, the desired corporate application. This allows an organization to run different operating systems on the same physical computer to handle diverse workloads. Although virtualization allows multiple logical computers to be created from a single physical computer, the number of virtual machines (VMs) that can be created is limited by: The physical resources available on the primary computer. The computing requirements imposed by the corporate apps running in those VMs. Types of Server Virtualization There are several types of server virtualization. We've highlighted the most common types below. The hypervisor, also known as Virtual Machine Monitor or VMM, acts as a layer between the operating system and the hardware. This software package develops and runs VMs and allows a single host machine to support numerous guests' VM by sharing its memory and processing resources. It also handles queues, executes commands, and responds to hardware requests. Since the guest Virtual Machines are independent of the host hardware, they can maximize the system's available resources, allowing for greater IT mobility. The two types of hypervisors available for use are Bare-Metal hypervisors and Hosted hypervisors. Full virtualization is a popular method that combines binary translation and direct execution. The operating system and corresponding hosted software are operated on top of virtual hardware, which separates the computer service demands from the physical hardware that performs them. The challenge with this technique is that it causes continuous traps to the hypervisor and intercepts privileged operations, e.g., the input-output instructions. Para Virtualization (PV) This is a virtualization upgrade in which the guest operating system is altered before it's installed inside a VM, allowing all guest operating systems to share resources. Here, virtual machines become readily available through interfaces tied to the similar underlying hardware. This functionality lowers costs and improves system performance by utilizing VMs that are underutilized in traditional hardware virtualization, hence alleviating full virtualization concerns. Hardware-Assisted and Kernel Level Virtualization Hardware-assisted type is similar to paravirtualization and full virtualization, except that it needs hardware support. Additionally, no modifications to the guest OS are required, and hypervisor overhead is reduced. On the other hand, Kernel Level Virtualization employs a distinct version of the Linux Kernel instead of a hypervisor, allowing numerous VMs to run on a single host. Advantages and Disadvantages of Server Virtualization Server consolidation: Since virtualization permits a single physical server to handle several operations, the total count of servers in an organization can be decreased. Simplified physical infrastructure: The total count of racks and cables in the data center is drastically decreased, thanks to the fewer servers used. Reduced hardware and facilities costs: Consolidating servers minimizes the cost of data center hardware and facilities. This also reduces the overall maintenance costs for enterprises. Increased server adaptability: Because each VM is its own separate instance, it must run its own operating system. On the flip side, the OS can differ between VMs, allowing the organization to run any combination of Windows, Linux, and other operating systems on the same physical hardware. Increased Risk: Using the same physical computer to run numerous workloads poses a risk to the business. Multiple workloads can be affected by a server failure in virtualization, thereby creating further interruption to the organization, its employees, partners, and customers. Virtual server sprawl: Unused or redundant virtual machines continue to consume precious server resources while doing little valuable work; in the meantime, those resources are unavailable to other virtual machines. As VMs multiply, the organization's resources get depleted, compelling it to make unanticipated capacity investments. VM Licensing: Hypervisors and accompanying virtualization-capable management tools add to the organization's costs. Hypervisor licensing must be closely controlled to ensure that the software's licensing agreements' terms and conditions are followed. Requires Professional Servicing: Professional IT personnel must adopt and manage a virtualized environment successfully. Server virtualization serves firms in the IT industry in various ways, including lowering hardware costs and streamlining physical infrastructure. Still, virtualizations come with some drawbacks, so you want to keenly weigh the pros and cons with respect to your organization's IT needs. Before deploying server virtualization, consult an expert IT professional to help you choose a virtualization technique that will guarantee the most benefits to your firm.
<urn:uuid:8858dc1f-2310-42ca-b126-bbc2249b7848>
CC-MAIN-2024-38
https://community.fs.com/article/server-virtualization-complete-guide.html
2024-09-08T02:31:14Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650958.30/warc/CC-MAIN-20240908020844-20240908050844-00784.warc.gz
en
0.924189
1,368
2.84375
3
Did you know that today, an astonishing 3.81 billion people are active on some type of social media? For context, in 2015, there were around 2.07 billion active users on social media. Isn’t it mind-boggling that, in just 5 years, the number almost doubled? Might that have something to do with the fear of missing out or being left out? Surely, the benefits of social media drive many people to join any given platform. And the fact that it’s the most efficient way to keep in touch in an increasingly more digital world is another reason to have an account. But there’s more to those platforms than perks and benefits. As awesome as social media is, it is often just as dangerous. We now know that it causes negative emotions, anxiety, fear, and much more. A decade ago, the fear of missing out (FOMO) was nothing more than a trendy Facebook hashtag. Today, it’s a documented, thoroughly researched, and scientifically verified psychological phenomenon, and we know it’s related to the excessive use of social media. What is FOMO? For anyone who’s been online for the past 5 years, FOMO, or the fear of missing out, might be a familiar term. It refers to the feeling of anxiety caused by a false belief that an exciting event is currently happening and you’re missing out on it. Nearly everyone has experienced it. Remember that time you saw a bunch of your friends at that Drake concert, and they didn’t even bother to let you know? That’s FOMO, and experiencing it can be unimaginably unpleasant. Interestingly, the term FOMO was coined by a consumer behavior expert Dan Herman back in 1996, way before the ubiquitous use of social media. The term took more than a decade to become part of our vocabulary. Now, however, FOMO is omnipresent; we even have a self-assessment scale to measure it, which urges us to raise questions like: Do I get worried when I find out my friends are having fun without me? How does social media create FOMO? It’s true. Social media has made staying in touch and being socially engaged with friends and family a lot easier. But that’s not all. We also get to see what our friends are up to, where they are or have been, what they’re talking about, and even what they are buying. In certain ways, these aspects of social media can be positive, highlighting opportunities and connecting people. However, the unprecedented access to each other’s lives is the perfect hotbed for comparison. Too often, social media has a distorting effect on how we see and interpret other people’s lives as they show them on any given platform. We tend to forget that what we see on social media is a small glimpse of our friends’ lives, mostly the so-called highlights. We never get to see the flip side of the coin, so the mundane bits of our own everyday life can make us feel like we are missing out on something exciting and are the only ones preoccupied with boring stuff. We have to be acutely aware that the way social media platforms are set up can also make us experience FOMO. Just think about it: a feature that shows how many likes a post has received can surely spur fear that you’re not receiving social approval. The infamous “seen” feature is another example. We all know how disgruntling and annoying it can be to be left on “seen”. How to overcome FOMO? So, what can be done about FOMO? How can we overcome and manage it? We have compiled a list of tips to reduce the onset of FOMO: Set limits on smartphone use Create a self-imposed limit on your daily phone usage. If you have trouble doing so, you can get an app that will automatically limit access to your phone. Replace your phone with other activities There’s so much more you can do instead of staring at your phone screen. Read a book, go to the movies, go for a walk, take up a new hobby, hang out with friends or family. Make it a point not to be on your device. Be mindful when spending time on social media Remember: the number of likes doesn’t define you or your value to the world in any significant way. What you see on social media is often a distortion. Don’t forget that everyone experiences the mundane parts of life, but very few of us are willing to share them with the rest of the world. Try not to compare yourself to others. We know it’s hard, but that’s the way to get rid of FOMO. Deactivate your social media accounts Taking a break from social media might be the best thing you can do. Deactivate your accounts for a while and see how you feel without being constantly exposed to other people’s lives. If that helps, consider getting off social media for good. However, if you choose to stay on social media, don’t forget about cybersecurity. Losing a password or falling victim to a data breach can be a stressful and anxiety-inducing experience. It can feel much like FOMO. A recent NordPass study revealed that password management is adding to the stress caused by technology. Almost 30% of respondents noted that losing a password and trying to recover or reset it can be as stressful as becoming unemployed. With NordPass, you can get peace of mind when it comes to password management. You can securely store and access not only passwords but also credit cards, secure notes, and personal information. You will no longer need to type your passwords, as you can use autofill to log in to your favorite online accounts in just a few clicks.
<urn:uuid:3fccedb3-1dbf-4d41-bddb-4ed8184b2493>
CC-MAIN-2024-38
https://nordpass.com/lt/blog/what-is-fomo/
2024-09-08T03:28:33Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700650958.30/warc/CC-MAIN-20240908020844-20240908050844-00784.warc.gz
en
0.956784
1,232
2.578125
3
“What is socialism?” This is an oft asked question, especially by those who want to change their government’s economic policies to something that does not involve the control of the economy by an elite class or government. What exactly is socialism? Basically, socialism is a socialist economic and political system where the ownership of the means of production (also referred to as common or collective ownership) is transferred from the private individuals and/or corporations to the government. These means of production include the tools, machinery, and machines used to create goods that are designed to directly meet the needs of humanity. As a result, the government must distribute resources to provide its citizens with basic services such as health care, education, food and water, electricity, etc. These are all government programs, or socialism. The government’s use of force, i.e. taxation, is another form of socialism. The government uses tax revenue to provide basic services to citizens. While this type of socialism is not necessarily bad, it can be very detrimental if the government decides to spend too much on government programs and doesn’t provide adequate social services for its citizens. Some socialists, however, do not believe in government intervention in the economy. They believe that the government is just a bureaucracy that has no idea how to run a properly functioning society and economy. Thus, they advocate for privately-owned businesses instead. Many socialists also believe that the government is merely an enforcer for a socialist agenda; therefore, they do not mind when the government redistributes wealth from one group of people to another. The reality is that socialist policies have had disastrous effects on the world; not only do these policies harm the economy, but they also hurt individual freedom. For instance, in the United States, many of these socialist programs include welfare, entitlements and social programs such as education, healthcare, and education. In other words, a lot of these benefits go to those who are well-off, but do little for the poor and middle class. In addition to government intervention in the economy, socialism also includes extensive government intervention in the workplace. For instance, businesses must be owned and run by members of the communist party, meaning no one will own the business, no matter what the success or failure Although many consider socialism a failed system, there are also some people who consider it to be a very successful system. For instance, socialism may actually work well in places like Cuba, Venezuela, China, Iran, Cuba, Venezuela, and North Korea, because they are the few countries in the world that have a thriving free market economy and free enterprise. Other nations that have seen a rapid increase in prosperity since they introduced socialist ideas include Russia, Vietnam, Greece, South Africa, and Venezuela. So, as you can see, it is important to understand what is socialism. It may not be the right answer for your country’s economic situation, but it is certainly worth a try. In conclusion, we can say that socialism is different from communism, and in some ways, socialism is worse than communism. It may work in some places, but it doesn’t work in all places. As far as political systems go, we can say that Social Democratic socialism is probably the worst. It is extremely paternalistic and gives the government more control over the private sector than it already has. It also encourages the redistribution of wealth, which is unfair, and has been shown to have disastrous results. On the other hand, free-market capitalism, which is capitalism, is a system in which the government does not interfere with the marketplace and tries to create a business environment in which consumers can buy products and services without the government’s intervention. Although this system doesn’t guarantee economic prosperity, it does promote the economy, which is much more stable and is a very good system overall choice. This is basically the kind of system we live in now. So, the next time you hear someone talking about what is socialism, remember that it’s not a great economic system. It may not be the best choice, but it certainly isn’t the worst. Try to understand socialism and find a better alternative. Editor-in-Chief since 2011.
<urn:uuid:14636e00-c8db-4621-bfa2-2c30638fa7c2>
CC-MAIN-2024-38
https://economystandard.com/what-is-socialism/
2024-09-09T07:32:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651092.31/warc/CC-MAIN-20240909071529-20240909101529-00684.warc.gz
en
0.971998
864
3.15625
3
When NASA launched its James Webb Space Telescope (JWST) a few weeks ago, its mission was as large as the 7.2-ton payload: to reach back in time, to the depths of the ever-expanding universe for clues and insights about the formation of the very first galaxies. It’s no small feat or quest, considering the amount of the visible universe is less than 5% of the total universe. NASA is going well beyond the one billion stars within our Milky Way for this project, and well beyond the approximate several hundred billion neighboring galaxies looking for clarity and answers. Looking for clarity in expanding universes should ring familiar. Back here on Earth we continue to generate, retain and share increasingly larger and larger volumes of data each day. As a result, it’s becoming ever more difficult to capture, analyze, and synthesize that information into useful, strategic decision-making. Consider IDC’s Global DataSphere study from 2021 that estimated that 64.2 zettabytes of data was “created or replicated” in 2020, far outpacing other earlier prognostications (1ZB = about one trillion GB). But that’s not all. The report also predicted that over the next five years the world would generate more than twice the amount of data created since the “advent of digital storage.” A variety of indicators support the projections, from a rise in data hoarding and lax retention programs to machine-based data-generating devices, not to mention cultural phenomena including social media and smart phones. Connected IoT devices, on their own are expected to generate 73.1 zettabytes by 2025. Meanwhile, half of the world’s population is using smart phones, more than one billion are on TikTok, and as of this writing almost two billion sites (and growing by the second) crowd the World Wide Web. Your Guidance System As businesses embark and progress on their digital transformation journeys, corralling and making sense of their own growing data volumes can seem like a drifting dream. Expanding distributed environments, ever-complicated mixes of on-premises datacenters and private and public clouds, scattered edge networks, and entrenched siloed pockets of productivity are just some of the digital challenges they face. What’s needed for the hybrid cloud and data-driven world is guidance, as much as solutions. What’s needed in the crush of product and services material are perspectives from people who have architected massive data migrations, cloud integrations, application modernization, smarter infrastructure mapping, and much more; perspectives that put data and its accessibility, availability, scalability and security, at the core. That’s the charge for Insights by Hitachi Vantara, our new site that’s dedicated to publishing fresh perspectives and points of view on the latest data trends and innovations. Through context and best-practices our goal is to serve as a trusted resource for those architecting or fine-tuning their data-driven strategies; to augment your research, confirm your understanding, or spark new ideas. On Insights by Hitachi Vantara we’ll work to share stories that give you new ways to consider your approach to data, from the datacenter to the edge because we believe in its power to positively impact everything from sustainability to supply chain management, from the enterprise to the environment. The Surround Sound of Silence Consider the nonprofit Rainforest Connection which works with governments around the world to preserve vital ecosystems by preventing illegal logging and fires by listening for anomalies in the forest. With solar-powered IoT acoustic sensors placed across the canopy of a forest, the organization collects acoustic recordings of virtually every living, moving thing from beetles to baboons. It analyzes the mountains of data and develops “bio-acoustic signatures,” base-levels of what a normal day or night sounds like. Once the signatures are created the org uses the sensors to monitor the forest continuously, and through AI, alert officials when deviations in the sounds are detected. Remarkably, what it found was that the sound of the forest changes long before a logger pulls the starter cord on the chainsaw. Animals complain with intrusions, they move, they grow silent. When such anomalies are detected, alerts are automatically sent to officials on the ground, who can deploy to the scene with precision to stop illegal activity before it occurs. That’s a far cry from traditional manual surveillance that includes scheduled physical inspections, which usually results in identifying the remains of illegal activity. The Rainforest Connection’s surround-sound approach to data management and analysis drives home the possibility and opportunity of the data-driven organization. And as organizations around the planet lean in on advanced data management and analysis to optimize their ever-expanding enterprises, they too will reap the benefits of insight and clarity to make better, more strategic decisions. Look for much more in the way of thought-provoking stories and videos with practical applications on the pages of Insights by Hitachi Vantara in the days, weeks and months to come. And welcome. Send us a note with comments, story ideas, reflections and experiences of your own. We’d love to hear from you. And subscribe to get stories emailed to you – as they post. Check out more great stories on Insights. Mike is managing editor of thought leadership, including Hitachi Vantara Insights and corporate Newsroom. Before joining the company, he spent +25 years in journalism and communications, working with edit teams and business leaders to craft stories of import and interest.
<urn:uuid:13854161-100c-4621-bc3e-4670cb98c18f>
CC-MAIN-2024-38
https://www.hitachivantara.com/en-in/insights/letter-from-the-editor-welcome-to-insights
2024-09-09T08:09:33Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651092.31/warc/CC-MAIN-20240909071529-20240909101529-00684.warc.gz
en
0.934801
1,157
2.53125
3
Quantum Tech: Semiconductor ‘Flipped’ to Insulator Above Room Temp (Semi-ConductorDigest) A semiconducting material that performed a quantum “flip” from a conductor to an insulator above room temperature has been developed at the University of Michigan. It potentially brings the world closer to a new generation of quantum devices and ultra-efficient electronics. Observed in two-dimensional layers of tantalum sulfide only a single atom thick, the exotic electronic structure that supported this quantum flip was previously only stable at ultra-cold temperatures of -100 degrees Fahrenheit. The new material remains stable at up to 170 F. “We’ve opened up a new playground for the future of electronic and quantum materials,” said Robert Hovden, U-M assistant professor of materials science and engineering and corresponding author of the study in Nature Communications. “It represents a whole new way to access exotic states.” Hovden explains that exotic quantum properties—like the ability to switch from a conductor to an insulator—could be key to the next generation of computing, providing more ways to store information and faster switching between states. That could lead to far more powerful and more energy-efficient devices. Today’s electronics use tiny electronic switches to store data; “on” is one and “off” is zero, and the data disappears when the power is turned off. Future devices could use other states, like “conductor” or “insulator” to store digital data, requiring only a quick blip of energy to switch between states rather than a steady stream of electricity. In the past, however, such exotic behavior has only been observed in materials at super-cold temperatures. The ultimate goal is to develop materials that can quickly “flip” from one state to another on demand and at room temperature. Hovden says this research could be an important step in that direction.
<urn:uuid:c90cc42f-7bd8-48a5-90dc-d1548bf0dd6b>
CC-MAIN-2024-38
https://www.insidequantumtechnology.com/news-archive/quantum-tech-semiconductor-flipped-to-insulator-above-room-temp/
2024-09-09T07:29:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651092.31/warc/CC-MAIN-20240909071529-20240909101529-00684.warc.gz
en
0.920283
411
3.375
3
25 Most Dangerous Programming Errors Exposed By publicizing these common programming errors, the participating organizations hope to make software code, and by extension the nation's cyberinfrastructure, more secure. Experts from more than 30 U.S. and international cybersecurity organizations plan to disclose the 25 most dangerous programming errors on Monday, at a media event in Washington, D.C. The CWE/SANS Top 25 List was compiled with help from organizations and individuals including Apple, CERT, Microsoft, Oracle, Red Hat, and Symantec, to name a few. It is managed by The SANS Institute and Mitre, and funded by U.S. Department of Homeland Security's National Cyber Security Division and the U.S. National Security Agency, both of which also contributed to the development of the list. CWE stands for Common Weakness Enumeration, a government-sponsored software assurance initiative. By publicizing these common programming errors, the participating organizations hope to make software code, and by extension the nation's cyberinfrastructure, more secure. Just two of these errors led to more than 1.5 million security breaches in 2008, according to the groups. "This activity is an important first step in managing the vulnerability of our networks and technology," said Tony Sager, director of the Vulnerability Analysis Office at the National Security Agency, in a statement. "We need to move away from reacting to thousands of individual vulnerabilities, and focus instead on a relatively small number of software flaws that allow vulnerabilities to occur, each with a general root cause. This allows us to then target improvements in software development practices, tools, and requirements to manage these problems earlier in the life cycle, where we can solve them at large scale and cost-effectively." The hope is that the errors list will serve four major purposes: To make software more secure for buyers by requiring that vendors certify their software is free of these top 25 errors; to incorporate awareness of these errors into software testing tools; to provide information necessary for educators to teach more secure programming techniques; and to provide a guide for employers to determine the abilities of programmers to write code free of these errors. "The first two errors on the Top 25 are improper input validation and improper output encoding, and they earned the top rating for good reason," said project editor Steven Christey of Mitre in a statement. "In 2008, hundreds of thousands of innocent, and generally trusted, Web pages were modified to serve malware by automated programs that burrowed into databases using SQL injection," he said. "The attack worked because countless programmers made the exact same mistake in their software. In a 2005 incident exploiting these same two errors, a teenager used a cross-site scripting attack to create a worm that hit the profiles of over 1 million MySpace users in less than a day, causing a temporary outage for the entire site." The Top 25 List consists of three categories of programming errors: Insecure Interaction Between Components (nine errors), Risky Resource Management (nine errors), and Porous Defenses (seven errors). Examples of errors in the respective categories include: CWE-20: Improper Input Validation; CWE-119: Failure to Constrain Operations within the Bounds of a Memory Buffer; and CWE-285: Improper Access Control. About the Author You May Also Like State of AI in Cybersecurity: Beyond the Hype October 30, 2024[Virtual Event] The Essential Guide to Cloud Management October 17, 2024Black Hat Europe - December 9-12 - Learn More December 10, 2024SecTor - Canada's IT Security Conference Oct 22-24 - Learn More October 22, 2024
<urn:uuid:5e88ddda-3590-4c37-b48a-4e6445e229c0>
CC-MAIN-2024-38
https://www.darkreading.com/cyber-risk/25-most-dangerous-programming-errors-exposed
2024-09-10T15:09:44Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651255.81/warc/CC-MAIN-20240910125411-20240910155411-00584.warc.gz
en
0.926248
753
2.796875
3
Exploring data is an integral part of any data analysis project. Microsoft Fabric is a game changer as it provides various tools for data exploration and preparation. Notebooks, in particular, have been found to be one of the fastest ways for initiating data exploration. Inspiration is drawn from the world of experiment, exploration, and the integration into the Fabric data science. A deep dive is taken into the details of how to read data from Azure Data Lake Storage using shortcuts and thus, capitalize on the Onelake flexibility. Organizing raw data into structured tables serves as a fundamental step for basic data exploration. Data from the diverse and enchanting city of London is used in this exploration, highlighting the accessibility and versatility that Fabric offers for data analysis. For the primary example, an Azure Data Lake Storage (ADLS) account is connected by a shortcut to a workspace in the system, named "Workspace 1". The aim is to replicate that data from ADLS to a new workspace known as "Workspace 2". This is facilitated by the Onelake that sits between the two workspaces. Additional reading can be found here. Data manipulation is crucial to restructure, copy, or modify data to meet specific requirements. Incorporation of several techniques in data manipulation leads to a more effective data exploration process. Fabric, known for its versatility, provides several ways of performing this task. One way is through the mssparkutils tool which is a preferred choice among many users. It allows recursive exploration of subfolders within the main folder, copying the data while still maintaining the same structure at the destination. A block written in Python demonstrates this functionality and the ease with which it is implemented. Once the data is copied and organized into delta tables, the next step is to make use of it. The data is then scrutinized to understand different factors using pivot tables and pie charts, which give a more significant visual representation of the data and insights. For instance, looking specifically at music events in London, statistical items such as standard deviation, the mean, the minimum, and maximum are determined. These serve to measure the variability of data points relative to the average. This enhances our knowledge on variation in music events across different London wards. Ultimately, this blog post does not only illustrate the functionalities Microsoft Fabric provides, but also the adaptability of these tools in various data operations. From data extraction, transformation to data exploration, the Fabric comes packed with interactive and user-friendly resources that can suit individual project demands. Read the full article Fabric Change the Game: Exploring the data The Microsoft Fabric program offers an exciting array of tools for data analysis. Its fast-paced world of exploration and integration makes it a perfect fit for data enthusiasts who want to administrate their data effectively. To better understand Microsoft Fabric, we would take a close look at Microsoft Azure Data Lake Storage. Here, we learn how to read data, use shortcuts, and structure raw data into tables using Onelake's flexibility. This foundational knowledge is crucial in mastering data analysis within Microsoft Fabric. Ever wondered how to unlock the massive potential of a data Lakehouse? Here, we illustrate how an ADLS account connects via a shortcut to a workspace in Fabric. The goal is to copy data from ADLS to a new workspace using Onelake, making data migration seamless within workspaces. If you're curious about the other ways to achieve this, you'll find the Microsoft Spark utilities interesting. It allows you to do a recursive search of subfolders inside the main folder, including copying and creating directories. Once this data copy is complete, we delve into CSV data, reading it and saving it as delta tables inside the Lakehouse using similar folder structures. Afterwards, we explore implements such as Z-order optimization and V-order at the Lakehouse Level to get the best out of our data. Tables created, we proceed to look at the data. For instance, taking a hundred "Wards" (the neighborhood unit that organizes the data), we can explore the variability of data points close to the average. Using measures like minimum values, maximum values, mean, and standard deviation helps us understand the spread of music events across London. By plotting Histograms, calculating for outliers, and looking at frequency distribution, you can grasp an overview of music events across Ward. This helps spot areas with unusual event counts and get an understanding of the variability of music events across "Wards" In summary, Microsoft Fabric makes it simple to copy data from an ADLS shortcut - Workspace1 to Workspace 2, through the Onelake structure. After the data is set in place, we apply the same recursive logic to create delta tables from the CSV files and begin with some basic data exploration. By the end of this journey, you should have a clearer understanding of how you can make the most out of Microsoft Fabric for more effective data analysis. Fabric data exploration, changing game fabric, game changing textiles, explore fabric data, manipulating textile data, fabric innovation game, fabric data game change, exploration in fabric data, game of fabric data, game-changers in fabric data.
<urn:uuid:6946fc7f-afb0-4d49-bc86-71dc6824ab7b>
CC-MAIN-2024-38
https://www.hubsite365.com/en-ww/crm-pages/fabric-change-the-game-exploring-the-data.htm
2024-09-10T13:57:02Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651255.81/warc/CC-MAIN-20240910125411-20240910155411-00584.warc.gz
en
0.904649
1,054
2.578125
3
- What is M-Files? - What is Metadata? - Intro to the M-Files Interface - Accessing M-Files - M-Files Terminology - Saving Documents with Metadata - Introduction to Metadata Cards - How to Save Documents - Finding Information with Metadata - How to Use Quick Search - Organizing with Views - How to Use Views - How to Use the Pinned Tab - The M-Files Way to Collaborate - How to Modify Documents - How to Share Documents - Grouping Information - Creating Views in M-Files - Search Options - How To Create Document Collections - How to Create Multi-File Documents - How To Create Relationships Between Objects Tips and Tricks - How to Create and Complete Assignments - How to Create Notifications in M-Files - How to Use and Create Document Templates - How To Use Offline Mode in M-Files - How to Change the Default Check-In Functionality - How to Use Workflows in M-Files - Permissions in M-Files - How to Convert Documents to PDF Format - How to Avoid Creating Duplicate Content in M-Files Finding Information with Metadata With M-Files you never have to wonder where something is. It’s all in M-Files. All you have to know is something about what it is or what it’s related to, and you’ll find it. In this lesson, we go through how metadata helps you find documents instantly and reliably. Find Documents Instantly We all know the pain of spending minutes, if not hours, looking for where a document might be. In M-Files, you never need to know where a document is. All you have to know is something about the document and you’ll find it immediately. This is because everything in M-Files is tagged with metadata that describes what it is, what it relates to, and how it ought to be managed. So as long as you know something about what you’re looking for, you’ll find it. For example, let’s say that you want to see everything related to a customer – the ESTT corporation. All you have to do is use features like search or views to find all information that’s been tagged with the Customer: ESTT Corporation in the metadata and M-Files will bring the documents to you instantly and reliably. Relationships Through Metadata Most documents are related to some larger context, like a customer, project, or an asset. Whenever two documents are linked to something in common, like a project, then M-Files creates a relationship through that common metadata. What this means is that you can use these relationships to browse and find information. For instance, if you want to know where a project stands, you can easily browse through all the documents that are related to that project. This is another big benefit of metadata – information is no longer hidden and separated by location, it’s all connected and related through metadata. Since everything in M-Files has a metadata card that explains what it is and what it’s related to, to find a document all you need to know is something about it. For example, let’s say that you need to find a proposal that you colleague Tina showed in a meeting, for your customer, XYC corporation. So, what do you know? – it’s a proposal that Tina created, and it’s related to XYC. This is exactly the sort of information that’s captured on metadata cards. All you have to do is ask M-Files, with features like quick search and views, to pull up all information that’s been tagged with that metadata. There are ways you can organize your documents in M-Files in a way that feel like folders. For instance, views look and feel like folders, but they don’t have any of the disadvantages of folders. Creating your own views is a great way to give easy access to the information that matters most to you. Another option is to use the Pinned tab. You can use it to organize information in a way that makes sense to you. Once you get to know these features, you’ll see that M-Files offers you much better ways to organize your information.
<urn:uuid:4955f2d2-e016-48df-8daf-9baef987655f>
CC-MAIN-2024-38
https://help.m-files.com/guides/finding-information-with-metadata/
2024-09-13T00:49:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651506.7/warc/CC-MAIN-20240913002450-20240913032450-00384.warc.gz
en
0.924486
920
2.53125
3
Cybersecurity is a critical concept to take into consideration in the chemical sector given the potential consequences a successful cyber attack would create. This is a sector that has its main core in the production of chemicals then used in a wide range of industries, from healthcare to transportation, to agriculture and other sectors. An attack against any company working in this sector, with the production of these chemical products could have as a consequence the release of hazardous material in the environment, representing also a risk to people and animals. Another consequence of a cyberattack against this sector could also disrupt strategic critical infrastructure and cause economic damage to a wide range of interconnected companies in a certain country. In fact, the chemical sector is definitely considered a critical infrastructure with strategic goals, making it an attractive target for any malicious actor. This is a sector often interconnected with other pivotal sectors, such as energy production and water lines, and it is a critical part of the industry also because it deals with sensitive information linked to the research and development of chemical formulas, the creation of new production processes, and also customer data, making it a perfect target for industrial cyber espionage and intellectual property theft from other countries. In the EU, the chemical sector is regulated by different laws more or less related to cybersecurity. The main framework where any company has to abide by is the General Data Protection Regulation (GDPR), that even if primarily deals with data protection and privacy, it also includes different provisions related to cybersecurity, requiring companies to take measures to ensure the security of personal data, and how it should be stored and processed within their cyber perimeter. There are also other laws that are regulating the European chemical sector. The NIS2 Directive, a EU-wide policy on cybersecurity that came into force in 2023, provides legal measures to boost the overall level of cybersecurity in the European Union, establishing a framework for the management of the eventual cybersecurity risks and incidents, creating requirements for EU member states and digital service providers working in the European territory of EU. The directive also creates a cooperation network among member states and establishes a single point of contact for reporting and sharing information on cybersecurity incidents. Another law affecting the chemical sector is the EU Cybersecurity Act, that came into force on June 27, 2019, introducing a EU-wide cybersecurity certification framework, it created the European Cybersecurity Research and Competence Centre, and the European Cybersecurity Certification Group. The act also enhances the role of the EU Agency for Cybersecurity (ENISA). This regulates the chemical sector because it affects ICT products, services and processes. There’s also another legislation that even if it is not directed to cybersecurity, the Chemicals Regulation (REACH), this governs the registration, evaluation, authorization, and restriction of chemicals in the European Union, and even if it does not specifically address cybersecurity, it does require companies to provide and store information on the hazards and risks associated with the chemicals they produce or use, which may lead to a data leak if the logistics of the data storage or transfer are badly handled. We take some case studies to provide specific examples of attacks against the sector scope of this brief. In 2014, DuPont, an American multinational chemical company, disclosed a data leak scandal of sensitive and private company data, including trade secrets related to the production of titanium dioxide, used to make the pigment for the colour white, that happened within the company. This case was an example of insider threat, in fact as documents and testimony in a 2014 federal trial in San Francisco reveal, a naturalised American citizen stole DuPont’s protocols for producing its superior titanium white from 1997 through 2011, thus being responsible for selling classified documents to companies controlled by the government of the People’s Republic of China (PRC). In 2019, Norsk Hydro, a Norwegian aluminium company, was attacked by ransomware, LockerGoga, forcing them to shut down some of its production facilities. The attack is believed to have originated in Russia, affecting all 35,000 Norsk Hydro employees across 40 countries, locking the files on thousands of servers and PCs. The financial impact would eventually approach $71 million. All of that damage had been set in motion three months earlier when one employee unknowingly opened an infected email from a trusted customer. Another interesting case is the ransomware attack that targeted the German chemical distributor Brenntag, in May 2020, carried on by the hacking group called DarkSide. The company’s IT systems had to be shut down in order to contain the attack, and this affected the company’s operations in several EU regions. To restore its systems, the company had to pay 4.4 million dollars to the criminal group. Regarding the intrusion vector, one of DarkSide’s members claimed to have managed to get access to the network after purchasing stolen credentials. Private data leak page sent to Brenntag after 2020 ransomware attack Again in 2020, the German chemical company Lanxess was targeted by a cyberattack that disrupted its IT systems and caused production delays. The attack is believed to have been carried out by a Chinese hacking group called WinNTI, an Advanced Persistent Threat (APT) likely affiliated to the Chinese Government, and is thought to be part of a larger espionage campaign targeting the chemical industry. In May 2021, Italian chemical company Radici Group was hit by a ransomware attack that disrupted its operations. The attackers, the Russian speaking group BlackBasta, demanded a ransom payment in Bitcoin to restore the company’s systems. According to the Tactics, Techniques and Procedures (TTPs) used by BlackBasta, as per Norsk Hydro incident, also in this case the initial vector was a spear phishing email opened by an employee, containing the entrypoint of the infection chain. BlackBasta infection chain [source: Kroll.com] In August 2021, the Oil and Gas company Shell disclosed that it had suffered a cyberattack that impacted some of the company’s operations in the Netherlands. The hackers managed to target a specific piece of hardware with previously-unreported vulnerabilities. Once they breached the vulnerable servers the hackers began exfiltrating data.The company stated that it had isolated the affected systems and launched an investigation into the incident. Finally, in the late 2022, different companies in the chemical industry based in Europe, Asia, Middle East and South America are targeted by hacking attempts from different malicious actors, including Advanced Persistent Threats (APTs), that are oftenly state-sponsored groups. An espionage campaign led by one APT, the North Korean state-sponsored hacking group, Lazarus, was discovered in November 2022 a campaign targeting the networks of a number of organisations operating in different sectors, including the chemical one. The group was using many attack vectors, from spear phishing mails to vulnerability exploitation in order to install the popular DTrack malware and steal logins, credentials and intellectual property. Timeline of the mentioned attacks During the analysis we conducted a quick research about the possible risks of some of the major European companies operating in the chemical sector, in terms of exposed credentials and vulnerabilities. For the research we chose a sample of 25 players operating in the analysed sector. Regarding the exposed credentials, 79% of the analysed companies have at least one credential belonging to their employees publicly exposed. As known, most of these credentials are stolen from data breaches occurred on third party sites, so it is possible that almost all of them are useless for obtaining an access on a company’s asset. This could be due to the fact that, hopefully, the password used on an internal asset is different from the one used on a third party site. Beyond the direct access to the company’s assets, don’t forget that publicly exposing employees’ emails could represent a risk due to the fact that a malicious threat actor could easily retrieve from the data breach a list of email addresses used as targets for an initial attack against the company. In other cases, the data from the breaches is publicly released after a long period of time and a company could have a password policy which imposes the password to be changed after a certain amount of days, making the exfiltrated credentials useless even if the exfiltrated password was the same of the one used for the company’s assets. However, according to our results, 68% of studied companies have at least one credential coming from data breaches published in 2023, so in the last 3 months. This could increase a bit the possibility of having the same password for internal and external portals, since the password policy usually expects the change to be made every 90 or 180 days. Regarding the external perimeter of the studied companies, the data is encouraging. Just 16% of the sample have at least an exposed vulnerability. The oldest vulnerability found is CVE-2006-20001 affecting Apache server, having a base score of 7.5. The most frequent vulnerability seems to be 2022-31629 affecting the PHP software and having a base score of 6.5. Finally, the most dangerous vulnerabilities found are CVE-2022-31813 and CVE-2022-22720, both affecting the Apache server and having a critical base score (9.8). Most frequent vulnerabilities found during the study From our study, fortunately, we didn’t find any open RDP (3389) port on the analysed companies sample. Exposed RDP server are one of the most exploited way by ransomware groups that, in combination with stolen credentials of a targeted company, can find an easy access to the internal network. These attacks are just significant examples of cyber-attacks on critical infrastructure, highlighting the growing threat of cyber-attacks on critical infrastructure and the need for improved cybersecurity measures to prevent such attacks from occurring in the future. Any company should take proactive measures in order to prevent cyberattacks and to have a robust cybersecurity strategy in place to protect their operations and sensitive information. To mitigate the cyber threats, and avoid replicating the above mentioned examples there are certain steps any company should take. Before anything, for the company it is important to understand and identify all its assets, data and systems that are the most critical to the business and assess any risk that could happen to each one. The company should also implement strong policies for regulating the access to its premises and to its data by employees and visitors. The data should be encrypted and there should also be data backup systems. It is vital to conduct regular security training for all the employees, as it is important to do regular security audits and assessments to all the company resources. And as last resort, having an incident response plan, to limit the damages of a cyber accident of any sort.
<urn:uuid:7b98c381-9cba-49df-8460-71b8f374ced4>
CC-MAIN-2024-38
https://www.duskrise.com/2023/04/12/cybersecurity-risks-and-challenges-in-the-chemical-industry/
2024-09-14T08:00:48Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651559.58/warc/CC-MAIN-20240914061427-20240914091427-00284.warc.gz
en
0.946274
2,173
2.84375
3
How AI Is Shaping an Inclusive and Diverse Future AI's Transformative Impact and Challenges in Developing RegionsArtificial intelligence is steadily emerging as a driver in bridging the digital divide and fostering inclusivity, particularly in developing regions where disparities in health, education and economic opportunities are most acute. The widespread digitization across these regions is not just a technological revolution; it's a data revolution. This abundance of data is pivotal for AI, allowing it to address complex societal challenges in various sectors such as weather forecasting, healthcare and agriculture, thereby unlocking the potential for a more inclusive future. AI in Weather Forecasting One of the most significant examples of AI's impact is seen in India's adoption of AI in weather forecasting. This innovative shift from traditional methods to AI-enhanced models has revolutionized how weather-related information is processed and disseminated. By employing AI for high-resolution data collection at even the village level, India ensures that its diverse geographic and agricultural landscapes are better understood and catered to. AI's application in meteorology extends to public alerts on heat waves and health advisories, such as those for malaria, ensuring that forecasts and warnings are precise, localized and relevant to the varied needs of the vast population. This approach improves the efficacy of weather-related information and democratizes its accessibility, playing a pivotal role in India's inclusive response to the complexities of climate change. AI for Diversity and Inclusion India's forthcoming AI Mission, announced by Prime Minister Narendra Modi, is a proactive step in harnessing AI to enhance diversity and inclusion. This initiative aims to unleash AI's capabilities in crucial sectors such as agriculture, healthcare and education, focusing on extending its benefits to Tier II and Tier III cities. The mission is pivotal in democratizing AI education and skill development through India's extensive network of Industrial Training Institutes. It emphasizes fostering ethical AI practices and developing transparent, unbiased systems to prevent societal disparities. Strategic plans include a substantial increase in computing capacity, creating a Digital Public Infrastructure for startups and facilitating access to anonymized datasets, all aimed at cultivating an inclusive AI ecosystem. These steps highlight India's commitment to using AI for societal inclusivity and bridging technological divides. Challenges for Developing Regions Despite AI's potential to enhance diversity and inclusivity, its implementation in developing regions faces unique challenges. One is ensuring that AI systems are unbiased and representative of the diverse populations they serve, especially given the risk of perpetuating existing societal biases. Addressing the digital literacy gap is critical, as marginalized groups may lack the necessary skills to benefit fully from AI technologies. The risk of exacerbating existing inequalities must be considered, as those with limited access to technology could be further marginalized. Developing inclusive AI solutions requires a concerted effort to understand and integrate these populations' diverse needs and contexts. In Telangana, a faulty algorithm initially intended for criminal profiling mistakenly denied thousands of people essential welfare benefits, including subsidized food. This misuse breaches rights under the Indian food security law and underscores the risks of unregulated AI in critical public welfare decisions. In Uttar Pradesh, AI initiatives are enhancing diversity and inclusion. The state government's implementation of an AI-based common beneficiary database is streamlining the delivery of schemes, ensuring targeted assistance. AI bots in secondary schools are transforming education by monitoring examinations, assessing performance and identifying educational gaps. During the migrant crisis, an AI-powered portal facilitated crucial services such as direct benefit transfers, food distribution and job matching based on migrant skills. And integrating AI education in madrasas brings modern technological skills to these institutions, opening new employment opportunities and mainstreaming education for madrasa students. LLMs to Enhance Communication and Diversity Developing large language models for India's linguistic diversity highlights AI's role in promoting inclusivity. By addressing challenges such as diverse languages, dialects and the predominantly oral nature of many Indian languages, LLMs can transform sectors such as government, healthcare and education by improving services and altering job dynamics. Initiatives such as Bhashini and Bhasha Daan focus on enriching AI models with Indian languages and contexts, which is crucial for breaking down language barriers and fostering growth across India. Tools such as LiveTalk are revolutionizing communication accessibility. This software, equipped with multilingual translation and real-time speech recognition, displays spoken words as text in various languages, aiding deaf individuals and facilitating cross-lingual interactions in international conferences and business meetings. Generative AI advances inclusivity and diversity across multiple dimensions. Enhancing language and communication breaks down linguistic barriers, enabling access to information and services in many languages. In content creation, it reflects diverse cultural perspectives, supporting multicultural understanding. For individuals with disabilities, it offers accessible solutions through real-time translations and descriptions. Generative AI also devises innovative, tailored solutions for unique challenges marginalized communities face, demonstrating its potential as a tool for social good and equality. AI in Public Health and Agriculture In emergency management, AI-driven tools such as disaster maps revolutionize crisis response crisis responses. These tools enable efficient resource allocation and targeted emergency responses by accurately analyzing data, benefiting vulnerable communities. AI has had an impact on public health, especially in developing countries. During the Ebola outbreak in Sierra Leone, a company in Africa developed a natural language processing platform that enabled locals to communicate their experiences, providing insights for more effective public health strategies. Researchers in Rwanda have used anonymized metadata from cellphone networks to develop detailed maps of wealth distribution, guiding aid organizations in resource allocation. In agriculture, AI compensates for resource scarcity. In Africa, AI augments the capacities of agricultural extension workers, boosting productivity. In healthcare, AI-powered advancements in telemedicine and virtual training address the skills deficit in regions that have a shortage of medical professionals. Ethical Deployment and the Future of AI The deployment of AI must be guided by responsible and ethical practices. These include considerations of data protection, transparency and bias. AI systems must be designed to avoid perpetuating societal biases and be accessible to all. Diversity and inclusion in AI development must be balanced. Project teams and stakeholders must understand the criticality of these factors to identify, monitor and mitigate potential risks and challenges. The homogeneity in AI's development community can unintentionally inject biases into AI systems. Addressing this imbalance is crucial, as there is a growing recognition of the need for diversity and inclusion as critical elements in AI development. If steered responsibly, AI offers a pathway to a better world by bridging gaps in healthcare, education and economic opportunities across diverse regions. Incorporating incentives for data owners when their data is used underlines a commitment to inclusivity in AI. This practice is particularly effective in developing regions, encouraging a broader spectrum of individuals and communities to contribute their unique data. Such participation is vital for creating AI systems that reflect and serve diverse populations. Through this approach, we ensure that the benefits of AI are not only widely distributed but also grounded in a wide variety of data sources. In the global arena, AI harbors the potential for both "AI colonization" and the creation of a more equitable world. In AI colonization, dominant tech regions dictate AI norms, infrastructure and applications, potentially widening the digital divide. This risk is particularly pronounced if AI development overlooks diverse needs and contexts, leading to a one-size-fits-all approach that marginalizes less technologically advanced regions. If steered responsibly, AI offers a pathway to a better world by bridging gaps in healthcare, education and economic opportunities across diverse regions. The key lies in global collaboration, ethical AI frameworks and diverse representation in AI development to harness AI's transformative power for inclusive and sustainable global progress. Practical Examples of AI Implementation at Fujitsu Fujitsu is committed to the ethical and responsible deployment of artificial intelligence. We prioritize transparency, ethical development and accountable deployment practices, as demonstrated by our pivotal involvement in AI4People, an initiative to advance AI ethics and establish our AI Ethics and Governance Office. This office symbolizes our dedication to developing secure, safe and ethically aligned AI applications, and it steers our initiatives to meet the highest standards of ethics. We introduced the AI Ethics Impact Assessment tool to embody our commitment to ethical AI use and build trust in its applications. This tool scrutinizes the ethical impacts of AI technologies on individuals and society before their deployment and is a critical instrument for designing and auditing trustworthy AI systems. It is included in the OECD.AI's Trusted AI Tool Catalog. We have also developed technologies such as Wide Learning and Deep Tensor, which aim to enhance AI explainability and transparency and foster trust among our customers. We strictly comply with local AI development and deployment regulations and provide detailed, explainable AI reports when necessary. Our digital twin linkage technology, the Actlyzer, uses AI to enhance workplace safety and efficiency through the 3D analysis of human behavior without retaining personal information. This underscores our firm stance on data privacy and security, with applications spanning various industries - including manufacturing and retail - and showcases our dedication to merging innovation with ethical responsibility. Fujitsu actively promotes the responsible advancement and application of AI through these initiatives and technologies, ensuring that our solutions are beneficial, ethical and congruent with societal values. Our team of AI engineers is committed to projects related to AI ethics, and our AI professionals actively contribute to the discourse on AI ethics. They are sharing their expertise through writings and presentations, both internally and externally, to further the cause of responsible AI. AI presents enormous opportunities for reducing inequalities and promoting inclusivity in developing regions, but its deployment must be guided by ethical practices and a conscious effort to integrate diversity and inclusion at every stage. By addressing these challenges head-on and leveraging AI responsibly, we ensure that the benefits of this revolutionary technology are shared by all, creating a more inclusive and equitable global society.
<urn:uuid:a5d34d33-42b2-442c-942e-66e3232a29c9>
CC-MAIN-2024-38
https://www.inforisktoday.asia/blogs/how-ai-shaping-inclusive-diverse-future-p-3568
2024-09-17T22:02:05Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651835.53/warc/CC-MAIN-20240917204739-20240917234739-00884.warc.gz
en
0.930647
1,999
3.265625
3
There has been much misinformation and misunderstanding around so-called ghost frames. There is no frame type or subtype in the 802.11 standard called a ‘ghost frame’, but rather it’s a byproduct of frame reception that happens by way of 802.11 protocol design around 802.11 Basic Service Sets (BSSs). Before the ill-named concept of a ghost frame can be defined, a concept from the 802.11 standard must be presented. PPDUs Use Multiple Data Rates 802.11 frames are transmitted at 1, 2, 3, or 4 rates (during the course of the transmission of the frame). Let’s first look at PPDU frame structures, from the oldest to the newest, starting with DSSS from 802.11-1997 and again later in 802.11-1999. Fig 1. Direct Sequence Spread Spectrum (DSSS), from 802.11 Prime In the Fig 1 frame format, the PLCP Preamble and the PLCP Header are transmitted at 1Mbps, and the MPDU (which is later called the PSDU) is transmitted at 1 or 2Mbps. This means that DSSS PPDUs are transmitted at 1 or 2 data rates. Fig 2. High Rate Direct Sequence Spread Spectrum (HR/DSSS), Long Preamble, from 802.11b In the Fig 2 frame format, the PLCP Preamble and the PLCP Header are transmitted at 1Mbps (for backwards compatibility with 802.11 Prime), while the PSDU (formerly the MPDU) may be transmitted at 1, 2, 5.5, or 11Mbps. This means that HR/DSSS (Long Preamble) PPDUs are transmitted at 1 or 2 data rates. Fig 3. High Rate Direct Sequence Spread Spectrum (HR/DSSS), Short Preamble, from 802.11b In the Fig 3 frame format, the PLCP Preamble is shortened to 50% of its previous length but remains at 1Mbps. The PLCP Header remains at its previous length but is now transmitted at 2Mbps. This formatting breaks backwards compatibility, so APs have to indicate when Long Preamble Only clients (802.11 Prime and 802.11 HR/DSSS with Long Preamble) are present within the BSS so that all clients (and the AP) will use long preambles. The PSDU is then transmitted at either 2, 5.5, or 11Mbps. This means that HR/DSSS (Short Preamble) PPDUs are transmitted at 2 or 3 data rates. Fig 4. OFDM (802.11a) and ERP-OFDM (802.11g) In the Fig 4 frame format, the PLCP Preamble and SIGNAL field is transmitted at 6Mbps, and the PSDU (and surrounding sub-fields) are transmitted at the data rate specified in the RATE field, which is 6-54Mbps. This means that OFDM and ERP-OFDM PPDUs are transmitted at either 1 or 2 data rates. Fig 5. DSSS-OFDM (802.11g) with Long Preambles In the Fig 5 frame format, the PLCP Preamble and PLCP Header mimic that of DSSS and HR/DSSS (Long Preamble), using 1Mbps data rates (for backwards compatibility). The OFDM Sync (preamble) and the OFDM Signal Field are both transmitted at 6Mbps as with standard OFDM. The OFDM Data Symbols (the payload) is transmitted from 6-54Mbps. This means that DSSS-OFDM (Long Preamble) PPDUs are transmitted at either 2 or 3 data rates. Fig. 6. DSSS-OFDM (802.11g) with Short Preambles In the Fig 6 frame format, the PLCP Preamble and PLCP Header mimic that of HR/DSSS (Short Preamble) with 1Mbps and 2Mbps data rates. The OFDM Sync (preamble) and the OFDM Signal Field are both transmitted at 6Mbps as with standard OFDM. The OFDM Data Symbols (the payload) is transmitted from 6-54Mbps. This means that DSSS-OFDM (Short Preamble) PPDUs are transmitted at either 3 or 4 data rates. Since 802.11n, 802.11ac, and 802.11ax are all based on OFDM, the same operational rules apply to them, so there is no need to go any further than 802.11a for the purposes of this discussion. The point of this section was to show that PPDUs are not transmitted at a single data rate. What Is A “Ghost Frame?” Ben Miller’s concept of a ghost frame is defined distinctly as: Any frame transmission where only the PHY header can be decoded by a receiver. Just for clarification, you could say it in a slightly different (and longer) way: Any transmitted frame whose PHY header data rate differs from its MAC and/or PAYLOAD data rate so that at a given distance from the transmitter, a receiver could only receive the PHY header. That definition applies to all frames (and frame types) transmitted by any 802.11 transmitter, that are using a MAC/PAYLOAD data rate >1M or >6M (depending on band), whereby only the PHY header is decodable by a receiver. If the expression ‘ghost frame’ was initially meant to be an avoidable exception, it certainly isn’t. With his own illustration, Ben cements his definition. Fig 7. Ben Miller’s Own Illustration of His Definition of ‘Ghost Frames’ This concept goes much further than how Ben tries to position it, as I will explain below. Consider, for the time-being, all of the variables involved in whether a receiver can receive the PHY header, but not receive the MAC/PAYLOAD: - Receiver sensitivity - Receiver position/distance - FSPL and RF loss from objects in the environment - Output power of transmitter - Data rate (MCS rate) of MAC/PAYLOAD transmission - You could have a transmitter sending an OFDM PPDU at 6M/54M to a receiver that has very low sensitivity, and the range at which the 6M PHY header would cause backoff would be much larger than the decodable range of 54M. - You could have a transmitter sending an OFDM PPDU at 6M/9M to a receiver that has very high sensitivity, and the range at which the 6M PHY header would cause backoff would be negligibly larger than the decodable range of 9M. My point is that it’s continually and inexplicably variable, across all frame types, and manifestly unavoidable. Do Ghost Frames Matter? Ben asserts that the problems caused by PPDU multi-rate are: - An 802.11 receiver (or capture application) cannot decode the frames that are using a MAC/PAYLOAD data rate above 6Mbps when positioned outside the data rate’s decode boundary for that particular receiver. The data rate decode boundary is dependent on a receiver’s receive sensitivity, and every receiver is different. This means that every receiver (AP or client) has a different set of data rate decode boundaries: 1) from every transmitter (AP or client), 2) for every transmission, and 3) of every frame type. This isn’t a problem, but rather simply how WiFi is supposed to work. - The PHY header’s LENGTH (or RATE and LENGTH with OFDM) cause receivers to back off First, the LENGTH (or LENGTH and RATE for OFDM) determine PHY layer backoff for the frame currently being transmitted. DURATION in the MAC header is different in that it causes receivers to back off for the duration of the entire frame exchange unless the MAC frame is destined to their own MAC address. The LENGTH (or LENGTH/RATE) (Layer 1) work in unison with the DURATION field (Layer 2) so that all frames are protected, including initial frames in a frame exchange sequence. The area past the physical MAC/PAYLOAD decodable boundary, while still allowing the PHY header to be decoded, for any given transmission, for any given receiver, becomes continually variable based on continual data rate changes, receiver sensitivity, and possibly a moving receiver. While it does in fact cause contention, it is not a concept that we can plan our networks around. Ben also asserts that we need to understand the decodable range of each BSA (basic service area), all the way to the 1M (DSSS) or 6M (OFDM) rates (depending on the band in question) due to the possibility of receiving only PHY headers. I refer you to Fig 8 below, where in Ekahau ESS I have configured a Cisco 2702i for 1mW (0dBm) Intentional Radiator (IR) power, positioned at one end of 5 American football fields (120 yards each), scaled properly (for a total of 600 yards or 549 meters). The one-way normalized (FSPL only) voice-grade coverage area for -65dBm on 2.4GHz is ~30 yards, while the interference area (down to ~4dB SNR @6Mbps) is ~600 yards. If you extrapolate for best-practice power output, two-way normalization (includes receiver sensitivity in the math) for a 2×2:2 receiver, and a 20MHz 5GHz channel, you quickly come to the conclusion that not only are these 6Mbps boundaries a really long way from a transmitter, but so are even higher data rates such as 9, 12, and 18Mbps. Fig 8. Minimum Decodable Data Rate Range, One-way Normalized At 2.4GHz, we often use 6-10dBm IR power (on the AP) and usually have 3-4dBi antennas. You can quickly see that the contention range, if the signal is not blocked by many walls (and other loss objects) is extraordinarily large. At 5GHz, we often use 10-14dBm IR power (on the AP) and usually have 4-5dBi antennas. 5GHz experiences an additional 7dB of loss in the first meter with omni antennas, and since the Inverse Square Law, when in open space, defines loss at -6dB for every 2 times the distance (2D), then we can roughly estimate that 5GHz has half of the usable range and contention range of 2.4GHz. In this case, “roughly half” would still be enormous. With a sensitive receiver (e.g. a 3×3:3 laptop), the contention range for either band is so large, without numerous, lossy walls to block the signal, that consideration of the PHY vs. MAC/PAYLOAD data rate differential is inconsequential. Ben’s assertion (below) that we need to enable 6Mbps data rates for surveys is unnecessary, as I will clearly explain in the next section. Three Causes of Contention (CCI) Because APs aren’t the only transmitters in an ESS, there are three causes of contention (not counting ACI or non-WiFi interference), that I call Intra-BSS, Inter-BSS, and Client-to-Client (or Client-to-Opposing-AP). These three types are illustrated in Fig 9 below. There is far more contention on your network than you may think, especially if you didn’t take the time to design and validate it. Fig 9. Three Causes of Contention The reason that mentioning the three causes of contention is to support my assertion that clients cause a much larger contention area than APs do, for a given BSS. The contention radius of a BSS is primarily determined by: - Client group position around the AP, which includes distances from the AP - Output power of the AP - Output power of the client group - Receiver sensitivity of the associated clients - Receiver sensitivity and location of any nearby receiver, who is not in the BSS - Preamble & PHY header modulation type (BPSK) - Any loss objects between a transmitter and receiver There are also some modest causes of contention, such as: - AP’s Minimum Basic Rate (MBR), as this will positively impact clients that do not roam well, serving as a hard boundary (per client) for forced roaming. - The primary reason for raising the MBR is to decrease airtime consumption of management frames such as beacons and probe responses. Given a client transmitter between the AP (to which it is associated) and a nearby receiver (that is outside the BSS), client transmissions will be far louder to the receiver (and decodable at a much longer range than AP transmissions) due strictly to the effects of FSPL on AP transmissions, as shown in Fig 10 below. Fig 10. 5GHz FSPL Math Example AP transmissions are the primary cause of contention (from a BSS), as shown in Fig 11 below, when: - There are no associated clients - There are no associated clients in a given direction, whereby the AP is the closest transmitter to a receiver Fig 11. When AP Transmissions Are the Primary Cause of Contention (from a BSS) Ben’s counter-point to my assertion that clients are the primary causers of contention from a BSS (within an ESS), is that APs transmit more data (and therefore more often) than clients do, and if his assertion alludes to any single client, then he is generally correct. If his assertion alludes to all clients collectively, then that is only partially correct because it depends entirely on the situation at any given moment in any given network. My argument is that it doesn’t matter whether Ben is fully or partially correct about the percentage of airtime consumption by individual client devices or all client devices collectively within a specific BSS. My rationale is two-fold: 1) As shown in Fig 12 below, that client devices meaningfully, continually, and variably extend the contention range of a BSS in such a manner that where InterBSS contention may not be present in an ESS, Client-to-Client and/or Client-to-Opposing-AP contention very likely is present – the amount of which depends on the situation, and 2) all downlink (DL) data has to be acknowledged in the uplink direction. Whatever the percentage of airtime that is consumed by client devices within the BSS, that is the amount of airtime lost in other same-channel BSSs within the ESS. You cannot put a single BSS in a box and say what applies to it applies generally to the entire ESS because ESS design is rarely perfect for a variety of reasons. My friend Chuck Lucaszewski says (paraphrasing), “You have a channel reuse factor of 1 unless you can prove you don’t.” That means that you have contention (aka CCI) unless you can prove you don’t. Client-induced contention is the primary cause. An additional consideration is that enterprises are very rarely confined to a BSS, but rather have many BSSs within an ESS. Each BSS within the ESS has transmitting clients at given ranges from their associated AP, and with the limited number of 5GHz channels available, it is very difficult (especially in multi-floor environments) to get rid of 5GHz contention. It’s nearly impossible to get rid of 2.4GHz contention in such environments. The further client devices move from an AP, the more contention they cause to neighboring same-channel BSSs within reception range. If we had no clients, it would be difficult to rid ourselves of 5GHz InterBSS contention, but with clients present and mobile, it is nearly impossible to get rid of contention. The amount of contention caused by PHY data rates differing from MAC/PAYLOAD data rates pales in comparison and is in fact altogether ‘in the noise.’ Fig 12. Client-induced Contention Extends the BSA Contention Area The more active clients that are in a BSS, the more of the airtime is consumed with uplink transmissions, at a greater distance than the AP’s contention boundary, causing contention for any neighboring same-channel APs and clients. In Fig 12 above, the reference to “~4dB SNR” for the contention boundary comes from a detailed discussion here. Using RF Design and Survey Software Ben has asserted on many occasions that we need to know the contention range of each BSS for design purposes, but he doesn’t advocate using RF modeling or RF survey software. Instead, he advocates that all surveys should be done only with client devices, whereby heat maps are not available. The industry at large disagrees with this approach because we can, in fact, offset our RF designs and RF surveys for the “least-capable, most important” device (or said another way, the least-capable device that is important to the customer). You can find out more about that here. While today’s RF modeling tools help us reduce InterBSS contention, and design best practices help us reduce IntraBSS contention, we do not currently have tools that will indicate Client-to-Client contention. Manual surveys, using actual client devices (as Ben suggests) will allow us to assess Client-to-Opposing-AP contention, but that’s still not the entirety of the contention domain. I have done such manual surveys in the past, and they were helpful, but of limited use because they give the customer no verifiable data. This is why we use RF Site Survey software. Ben has also asserted that the antennas used in RF Site Survey kits (USB NICs, Sidekicks™, etc.) aren’t the same as those of client devices, and therefore survey data is wrong. I disagree because we use measurements from the customer’s actual client devices to offset our RF Models and RF Site Surveys, so that our heat maps look as any given client device would experience the RF environment. Corrections to Ben’s Blog on So-called Ghost Frames Below, is an excerpt from Ben’s blog on so-called ghost frames. The LENGTH field in an OFDM PPDU denotes the number of octets (bytes) the payload (PSDU and surrounding subfields) will be, and the RATE notes the MCS to be used for the payload. Together, they detail how long (in time) that it is estimated by the transmitter for THIS FRAME to make it across the air. The RATE/LENGTH (or just LENGTH in DSSS and HR/DSSS, which is interpreted as time in microseconds) is used to protect THIS frame transmission, and to force any station who can hear it to be quiet until such a time as they receive the rest of the frame. Once the MAC header is processed, they know whether or not to: 1) respond (because the transmission is for them), or 2) to be quiet the remainder of the entire frame exchange (which is denoted by the DURATION value in the MAC header). All of this is by design, and it is not a shortcoming of any kind. The airtime loss that Ben describes below is an inconsequential side effect of backwards compatibility and caused when a receiver is outside of the decodable range of the MAC/PAYLOAD of a frame. Fig 13. An Excerpt from Ben’s Blog On Ghost Frames While there is no 802.11 frame type called “ghost,” the concept of a PHY header being received when its MAC header and payload cannot does exist. No frame is transmitted as a ghost frame, rather the term ‘ghost frame’ as given by Ben Miller alludes to a receiver affect (described above) that may cause an inconsequential contention effect, for any frame type, that the 802.11 protocol designers purposefully intended to protect frame transmissions. In my opinion, the concept is misleadingly named (for marketing purposes) and irrelevant to Wi-Fi design, deployment, and validation.
<urn:uuid:97c0bdb6-0025-4292-8d6e-862edad39f9d>
CC-MAIN-2024-38
https://divdyn.com/so-called-ghost-frames-not-exist/
2024-09-19T04:03:55Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651981.99/warc/CC-MAIN-20240919025412-20240919055412-00784.warc.gz
en
0.937181
4,321
2.765625
3
Data governance is a practice that controls and handles all the data in an organization. The basic aim of the entire concept is to maintain quality data and ensure that accurate and correct information is provided at the right places. Not abiding by this can have serious consequences for your business and a severe impact on profitability. Indeed in many circumstances, you would already have people in your organization who are checking the quality of your data. As an example, your accountants ensure that invoices are sent to the right receipts on which the correct payments are received. Thus, most of your data is already a part of a management process, but you are mostly interested in the value and quantity. Those areas which are not inspected so often are the master or reference data, which you would be using in most of your processes. So with data governance, you are actually implementing a system that analyzes the quality of this reference data. Data governance lets you follow a proactive approach instead of a reactive one. In most cases, you realize that your data is not of a good quality only when one of your processes fails or when your IT system suffers a lag. However, if you have data governance in place, there would be someone whose sole responsibility is to manage this data, which will reduce the chances of disasters. If you want to implement data governance, you have to realize that this is more than just a tool or a method. Data governance is more of an organizational procedure in which processes are defined and responsibilities are assigned. True there are related tools, but they are only used for assistance. By no means do they completely define data governance. If you look at the scenario from a high level, there are just two main activities which a data governance organization performs. However, these two activities can be of an extremely complex nature and can only be achieved when a complete resource network is utilized. Once a set of standards has been defined and all data is in accordance with them, then all modifications to these standards must be controlled. As an example, if all dates are supposed to be stored in the day/month/year format, then it would not be quite a hassle to change all of the records to the month/day/year format. If there is a data governance team, it would analyze the effects of this modification and will conduct a cost and benefit analysis. If the results conclude that change is appropriate, the team would also monitor this change as it takes place in all business areas. If there are any rules, there are always policing requirements, which are the duties of a data governance team. The team must evaluate how the organization complies to the required standards and it must also take measures to improve the compliance level.
<urn:uuid:8c4af869-1247-4e27-9791-23e6bac13987>
CC-MAIN-2024-38
http://www.best-practice.com/governance/governance-11092013/
2024-09-20T11:45:37Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652246.93/warc/CC-MAIN-20240920090502-20240920120502-00684.warc.gz
en
0.965779
546
2.921875
3
Sometimes it is hard to understand how so many professionals in an industry latch on to one statistic on which they base all their theories, predictions, and forecasts without ever revisiting the basis of the original statistic or even tracking their performance to see if they are improving or not. A simple statistic published back in 2006 about how much electrical energy was consumed by data centers launched the industry on an epic effort to improve energy efficiency in these facilities. As an industry, we developed and vociferously debated multiple metrics, created a draft LEED standard specifically for data centers, modified ASHRAE Standards, and produced an untold number of solutions for energy efficiency in data centers. We attended conference after conference to discuss our corporate social responsibilities (CSR) for energy consumption and carbon footprints. Government at every level has actively engaged the industry and challenged us to develop more and better further performance standards and to promote discussion about the right solutions. Yet for all these positive efforts, article after article, and speaker after speaker continue to refer to Jonathan G. Koomey, Ph.D.,’s 2007 paper, Estimating Total Power Consumption by Servers in the U.S. and the World. In this paper, Koomey, a staff scientist at Lawrence Berkeley National Laboratory and consulting professor, Stanford University, reported that data centers constituted 1.2 percent of the U.S. electricity sales. As one might expect, as this figure continues to be repeated it has grown and now is referenced as high as 2 percent. The data in this report were compiled in 2005, and so in 2010 we have no new updated information. Industry analysts, vendors hawking their merchandise, and politicians looking to eliminate global warming all refer back to this old estimate. What have we accomplished in five years? We deployed virtualization, cloud computing, air-side economizers, high-efficiency UPSs, in-row cooling, and much more. Where are these successes quantified? Which of the five projected curves did we hitch a ride on? (See figure 2.) With the explosion in computing power, do data center power purchases now constitute 5, 8, or 10 percent of total electrical sales, or have we maintained or actually reduced the total electrical use of data centers? What would be the percent had we not taken these extraordinary measures? We live in a world where instantaneous and accurate answers are expected. Do a search for more current data on the web and what do you get? Listings for Cisco, Cybercon, data center designers, The Green Data Project, data collection, and more. There is no shortage of white papers and vendors putting their best foot forward. Perhaps I just do not know the right “key search words” or perhaps I need a specialized search engine, but I did not find any updated estimates of where the industry is going with respect to energy consumption. Further, if these data did exist, one would expect to it be fully exploited in the press and the vendor community alike. At least part of the answer lies in the industry’s continued unwillingness to share information. Take the U.S. Environmental Protection Agency’s (EPA) data collection efforts, for example. Despite its extraordinary efforts over the past four years, the EPA could get information from only 100 data centers on which to base its programs. That’s 100 data centers out of tens of thousands nationwide. The data center industry should be embarrassed by this woefully inadequate participation. A web search for colocation and hosting firms will turn up over 1,000 companies with multiple facilities. Why aren’t all these commercial entities, each of which measure every microwatt consumed in their operations contributing to the EPA’s database or some similar effort? What about the thousands of enterprise data centers? What about their CSR when it comes to helping the nation accurately trend its energy consumption and its carbon footprint? What about the health-care industry, where virtually everything is now regulated and reported? Everything, that is, except their data center power consumption. OK, so I have asked a lot of questions and have no answers, which means we need a call to action. Koomey’s 2007 report, based on 2005 data, was a wake up call, but most of us seems to have reached over and hit the snooze button. So now I call upon all of you-users, operators, vendors, consultants, and the like-to organize and start sharing/trending consumption data. Why can’t we create a real-time index like the NASDAQ or the Dow Jones Industrial Average? Who among us is willing to contribute to an industry index? Yes folks, I am talking about real-time reporting of data center power consumption. Can we as an industry step up to the plate or will it take an act of Congress to force us to do what we should be doing anyway? Let’s hear from you? In December 28, 2009, New York City Mayor Michael Bloomberg signed a series of energy bills into law that require NYC owners to conduct energy audits to assess efficiency and annually document energy consumption. The benchmarking bill requires owners to track and disclose energy consumption annually so they can see how efficiently their buildings function and enable prospective buyers and tenants to better assess the value of the building. Further, a sub-metering component of the bill recognizes that tenants, not building owners, are the consumers of electricity.We will no doubt be seeing more of these regulations that include data centers located within buildings.
<urn:uuid:71a62bbe-6e7e-4534-9dc8-e89d3967ddfc>
CC-MAIN-2024-38
https://www.missioncriticalmagazine.com/articles/82494-still-relying-on-old-data
2024-09-11T23:16:08Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651405.61/warc/CC-MAIN-20240911215612-20240912005612-00584.warc.gz
en
0.950038
1,129
2.5625
3
I’m pretty sure you will all agree with me if I describe humans as social beings. We are all part of a society with members who are interdependent. This dependency extends from the food we eat, the clothes we wear and all the other services we rely on. Even though we often like believing that we are independent, we realize that it is just an illusion, when we fall into situations, we can’t do anything about! In such periods we often look up to people who can handle these situations better than us – the first responders. Getting to know our heroes- the first responders Can you, for a moment, imagine the usual route to your office? Walking past the usual signboard, your eyes fall on that random object, and on looking closer, you realize that it’s a person! What will you do? Obviously, you will call an ambulance or 911, isn’t it? That’s how confident we are in our heroes – the first responders. We are pretty sure that they will be quick in reaching the emergency spot and will save the life of the person in need. According to U.S. Homeland Security Presidential Directive, The term ‘‘first responder’’ refers to those individuals who in the early stages of an incident are responsible for the protection and preservation of life, property, evidence, and the environment, including emergency response providers What are the tools that first responders need? First responders often need to work in situations that can be life-threatening. So, it is indeed our responsibility to ensure that these selfless warriors of ours are taken care of. With technology on the rise, it’s possible to add many such features like location tracking, GPS, etc. Into devices like smartphones or tablets. But there is also another challenge here, the harsh conditions they work in are not always favorable for smartphones and tablets. They need devices like smartphones, tablets, routers, modems, wearables, etc. which are also capable of resisting impact, unexpected drop or water penetration. What makes rugged devices the best choice? Rugged devices are designed to work in harsh conditions. Being part of the emergency services, they often find themselves in a wide range of situations. It can range from firefighting, emergency medical services to rescue operations in coastguards or even mountains, often involving unexpected events and high-stress levels. These conditions demand devices with greater capabilities over commonly used commercial smartphones. A closer analysis First responders play a vital role in ensuring the safety of our citizens. It would be really hard to imagine a society without their aura of protection around us. Let’s have a closer look at their functioning. Firefighters work in risky environments. We often hear about instances of firefighters being injured during rescue operations. This is primarily due to the lack of protective equipment. We can ensure that they work normally even in such complex environments by adopting smart devices and other equipment. These devices can give them a clearer picture of the challenges, helping them prepare for them well in advance. There is no doubt regarding what device type must be used in such situations. Obviously, rugged devices are the best choice here. These devices designed for challenging environments, capable of withstanding drops, submersion in water and even extreme temperatures, are preferred over consumer-grade devices even if we add heavy-duty cases adding shock resistance to them.Emergency medical services (EMS) Emergency Medical Services (EMS) refers to the system providing emergency medical care to patients. It involves a chain of processes activated by incidents causing severe illness or injury. Their prime focus is the emergency medical care of patients. These chains of events involve everything from the emergency medical care of patients with illness and injuries out of the hospital to transporting them to definitive care. Considering the critical nature of the situation, the ambulance team should have the ability to act fast, have dedicated equipment and professional skills and should also be able to handle any unexpected situation that they encounter with ease. Rugged devices become the preferred choice here as they are engineered to tolerate the harshest of situations. During emergency care, we need such devices as downtime is not at all among the list of emergencies we would want to face. It comprises the agencies and the employees responsible for providing the first response to emergencies and other threats to public safety. They play a pivotal role in maintaining public order and public safety. They are involved in the investigation, apprehension and detention of individuals suspected of criminal activities. Being responsible for a lot of activities, from helping victims calling out for help to making complaints and enforcing the law and preventing their violation, they have a lot of issues to attend to. So, by using the latest technology, we can ensure that they work efficiently and are well-equipped to face all the threats they may encounter. Rugged devices have made things easier for law enforcement. The ability to deliver high-level data coupled with communication functions even under challenging environments makes them suitable for their demanding job profile. Emergencies can come in different forms. The most common emergencies are usually weather-related or the results of geological events with impacts that may or may not be localized. Well, emergencies are not just restricted to natural disasters like thunderstorms, flood or earthquake; it extends to man-made hazards, including hazardous material releases, acts of terrorism, or nuclear accidents. So, in order to face this diverse set of challenges, the first responders should have all the necessary tools at their disposal. Rugged devices that ensure accurate tracking of victims can help reduce the number of causalities even in the toughest of situations. The other aspect where rugged devices are finding applications is in the post-disaster period. It helps in keeping track of the people, resources and other equipment. Why are devices alone not sufficient? Normalizing the use of rugged devices by first responders is definitely a great move in helping them respond better to emergencies. But, like any smart device, there are challenges associated with this implementation, which, if taken lightly, can make things even worse. You need to ensure that all your devices are functioning in a way that is efficient as well as secure. For this you should have efficient mechanisms that cover the entire process; from device onboarding, security, data management to real-time location tracking, to ensure that the whole process is running smoothly. With Hexnode’s UEM you can manage all your device aspects from a single console. Let’s have a brief overview: Device enrollment is the first step in the process that changes our endpoints from mere devices to managed devices. The easier the process, the better, isn’t it? With Hexnode’s easy enrollment methods you can enroll all your devices without any trouble. Some of the enrollment methods include: Security is one of the key aspects in any enterprise. In this period of uninterrupted internet connectivity, a proper security framework is vital in ensuring that your enterprise data is inaccessible to any potential intruder. Some of these features include: Hexnode’s UEM is an all-in-one platform that helps you manage most of the device aspects from a single console. It includes: Sometimes we face situations where we need to remotely manage certain device aspects. By pushing commands remotely, you can ensure that these immediate requirements are implemented almost instantaneously. Some of them include: Rugged devices are indeed the best fit for first responders considering the criticality of the situations and the harsh climatic conditions they are often exposed to. However, making these devices a part of your enterprise alone is not sufficient; you need efficient solutions to ensure that your devices and your data are protected even if it falls into unsafe hands. Hexnode’s UEM makes the whole process of rugged device management easier, thus ensuring that the first responders are actually responding to emergencies and not wasting their valuable time worrying about the safety of their rugged devices.
<urn:uuid:4cba9aa5-5064-4918-a432-f5a1ced3c1c9>
CC-MAIN-2024-38
https://www.hexnode.com/blogs/manage-first-responder-devices-with-uem/?utm_source=hexnode_blog_devices_for_first_responders&utm_medium=referral&utm_campaign=internal_link
2024-09-13T04:13:17Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651507.67/warc/CC-MAIN-20240913034233-20240913064233-00484.warc.gz
en
0.953559
1,618
2.546875
3
- Sora, a diffusion model, uses generative machine learning to create images or videos. It refines random noise into structured patterns based on learned data distributions. - Sora has the ability to create intricate scenes featuring multiple characters, precise motion types, and detailed subject and background elements. OpenAI has recently introduced Sora, a novel text-to-video model capable of producing videos with a duration of up to one minute. It ensures both visual excellence and alignment with the user’s provided prompt. OpenAI’s Sora Text-to-video technology is arguably the next significant advancement in artificial intelligence, and OpenAI is not the pioneer in this domain. Meta Platforms Inc., Google LLC, Runway AI Inc., and other entities also provide comparable services. The overarching challenge across these services has been achieving high quality. While some existing services produce remarkably impressive videos, the ultimate goal is to create realistic videos, a feat not all of them have mastered. OpenAI’s Sora operates as a diffusion model, falling under the category of generative machine learning models. It crafts data, including images or videos, by iteratively refining random noise into organized patterns, relying on learned data distributions. Sora possesses the capability to produce intricate sequences comprising numerous characters, distinct forms of motion, and precise particulars of the subject and backdrop. Furthermore, the model possesses knowledge not only of the information requested in the query but also of the physical manifestations of that which was requested. OpenAI asserts that the model possesses a profound comprehension of language, facilitating precise interpretation of prompts and the generation of “compelling characters that express vibrant emotions.” Additionally, the service can produce multiple scenes within a single generated video, adeptly capturing the essence of characters and visual style. Creditably, OpenAI has been transparent about the limitations of the Sora text-to-video AI model. In its current testing phase, Sora exhibits weaknesses, such as challenges in accurately simulating the physics of intricate scenes and potential difficulties in understanding specific cause-and-effect instances. Spatial details in the prompt, like distinguishing between left and right, might be confused by the model. Moreover, Sora might face difficulties in delivering accurate descriptions of events unfolding over time, such as tracking a specific camera trajectory. While the model has its imperfections, it is in its early stages, and some of the initial demonstrations are remarkably impressive. While OpenAI Sora appears impressive, ChatGPT users will need to exercise patience before gaining access. Presently, Sora is exclusively accessible to designated “red teamers” for evaluating potential risks and areas of concern. OpenAI is also extending access to visual artists, designers, and filmmakers, seeking feedback to enhance the model and tailor it to be most beneficial for creative professionals. OpenAI said, “We’re sharing our research progress early to start working with and getting feedback from people outside of OpenAI and to give the public a sense of what AI capabilities are on the horizon.”
<urn:uuid:ab91edfb-5107-4d9c-bf74-3544f8b04b16>
CC-MAIN-2024-38
https://www.ai-demand.com/news/tech-news/artificial-intelligence-news/openais-sora-text-to-video-ai-enters-the-content-creation-race/
2024-09-14T11:00:07Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651579.22/warc/CC-MAIN-20240914093425-20240914123425-00384.warc.gz
en
0.922508
612
2.953125
3
20 years ago, IT systems were huge mainframes and information was stored on tapes. These systems were considered to be in a closed environment – no firewall, no anti-virus protection, and you would definitely be noticed if you ever tried to share information between systems. Today, the way we use information is drastically different. With every new technological advance comes a new set of security considerations with various levels of vulnerabilities, threats and risk levels. That’s why the Privacy Shield framework includes Security as a key pillar – it makes organizations responsible for taking reasonable and appropriate measures to protect data from loss, misuse, and destruction. Information security managers are constantly required to meet the fundamental principles of security: confidentiality, integrity, availability and traceability, what I refer to collectively as CIAT: Confidentiality is the necessary level of secrecy which should be designated every time information or documents are created, updated, or transmitted. Think of confidentiality as having measures designed to prevent sensitive information from ending up in the wrong hands. Integrity focuses on maintaining trustworthiness and ensures the accuracy and reliability of the information. For example, how do you prevent unauthorized users from altering data? It is common to assign permissions only to a handful of people, enable version control, and store data backups to prevent accidental deletion to our information. Availability has more accent on systems but is also significant for the reliability and timely access to data too. Organizations have to put reasonable safeguards against data loss in unpredictable events such as system failure, natural disasters, or, as we have seen in recent cases, ransomware attacks. Having a backup copy of your data ensures your business has the ability to continue operating in such occurrences. Traceability, also known as audit trail, is often a prerequisite for accountability. Traceability is ensured by providing a detailed log of the actions done by a user who can be held responsible in some occasions such as: - Suspicious activities from employees after business hours or on their last working day: Ex-employees tend to leave with company data on their last days of work. Having a solution that monitors and audits such suspicious activities could help prevent possible data leakage before it’s too late. - Loss of data: Employee activities like deleting important documents un/intentionally could be commonly seen in eDiscovery or legal investigations. You may trust your employees, but it’s always better to have more control where possible. Best Practices for Data Security As systems become more and more integrated into business and personal activities, unexpected interruptions (i.e. data breaches) have much more potential to severely disrupt our lives. Nowadays, information is worth as much as gold – or even more depending on the consequences you would face if the information were exposed. Data security is more than just having a password, anti-virus software, a firewall, or a shiny router. It establishes best practices that focus on protecting information throughout the entire lifecycle. Privacy by Design One best practice is to require adoption of these security measures as early as possible within company projects. Having Security and Privacy by Design is a key factor before implementing any of the CIAT criteria. Privacy Impact Assessments Understand how your employees are working with sensitive data on an everyday basis by using a Privacy Impact Assessment. Even the most common process like sharing a document internally or outside of the organization, requires companies to know: - Is the document classified as sensitive or confidential? - Is the recipient supposed to have access or receive the document? - For how long does the recipient need to have access to the document? - Should the document be encrypted or read-only if it is opened from another country? - Do we have any monitoring or tracing mechanisms to assess who did what and when? An employee most likely wouldn’t have all the answers to the above questions, nor the time to go through such a repetitive process for every document. However, these requirements are usually driven by company policies that mandate a layered data protection solution that would accommodate CIAT requirements. These layers should include: - Data Discovery and Data Analysis: Understand where your sensitive data lives to identify potential risk and protect confidential information. - Data Classification: Classify data based on content sensitivity, criticality or confidentiality. Develop a security awareness that protects organizational assets via accountability, classification, and inventory. - Data Loss Prevention: Apply security controls and integrity layers on data based on classification. Control access to information based on business requirements or need to know basis. - Monitoring and reporting on any unusual user activities: Track compliance whether it is aligned with the policies. Organizations need to determine at early stages which types of documents/information are critical, how should they be protected or handled, and that everyone has a firm understanding of the CIAT criteria. More importantly, they need to understand how CIAT requirements directly relate to the company needs. A Security or Privacy Impact Assessment should be the first step in implementing a successful information lifecycle management practice in conjunction with an automated and measurable controls. To learn more about the Privacy Shield fundamental principles, sign up for our EU-U.S. Privacy Shield Guide! During his tenure as a Senior Compliance Technical Specialist at AvePoint, Esad was responsible for research, technical and analytical support on current as well as upcoming industry trends, technology, standards, best practices, concepts and solutions for information security, risk analysis and compliance. View all posts by Esad I.
<urn:uuid:6b7301bb-2372-4e4e-9ec3-dc01aeb70f3b>
CC-MAIN-2024-38
https://www.avepoint.com/blog/protect/data-security-protection-lifecycle
2024-09-14T11:07:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651579.22/warc/CC-MAIN-20240914093425-20240914123425-00384.warc.gz
en
0.954485
1,111
2.515625
3
On March 1st 2018 several media reports came out stating that GitHub was hit by a missive DDoS attack on February 28th. The attack was executed using an amplification attack as shown by the figure to the right. In an amplification attack an adversary (A) sends packets to servers (B and C) using spoofed IP addresses belonging to the victim (V). The servers (B and C) receiving the spoofed packets will send the responses to the victim (V). If the responses are larger than the packets sent by the adversary, an amplification is created. The amplification allows an adversary with little resources to generate high amounts of traffic towards the victim. This high amount of traffic saturates the link of the victim rending the services hosted by the victim unreachable. In the case of the attack on GitHub the adversaries used the memcached service to perform an amplification attack. The memcached service was introduced to speed up the fetching of resources. Retrieving data from disk is usually slow, therefore the memcached service uses system memory on servers to store frequently used data. This speeds up the request time as the data can now be retrieved from fast RAM instead of disk. The protocol lends itself well for being used in amplification attacks for three reasons: first, there is no authentication required, and second, memcached servers may be accessed by TCP and UDP. As UDP does not require the completion of a handshake, it is trivially possible to forge the IP address of a UDP request, which means tha data from the memcached service can be requested by any entity and delivered to an arbitrary destination. Finally, as the data requests are small and would result in very large answers, the potential amplification factor of the memcached protocol is high, it is possible to trigger downloads tens of thousands of times the original request size. Leveraging servers running the memcached service allowed the attackers to generate an unprecedented traffic surge of 1.35 Terabit per second towards GitHub. As stated in the title of this article, the DDoS on github.com is interesting because of its new exploitation vector and massive volume, but also because the media attention of the attack created a surge in activity itself towards the memcached service. In order for adversaries to abuse servers running the memcached service in an attack, they first need to know where these servers are. The obvious way to accomplish this is by scanning the Internet, that is probing each IP on the Internet and see if the servers behind the IP is responds to a memcached protocol request. As adversaries would most likely scan the entire Internet looking for vulnerable instances and use their own IP address to collect the responses, it is possible based on this probing activity to track who is looking for the service and get a fingerprint of by their exploitation attempt. In this study, we leveraged a large unused IP space and collected all traffic destined for potential memcached servers towards addresses in said space. The figure on top shows the probing activity measured in number of unique IP addresses probing for servers running the memcached service between August and April. By studying the number of unique IP addresses probing for servers running the memcached protocol an estimate can be made for adversaries probing for potential servers to abuse. The red line on the plot indicates the time of the attack on GitHub, 28th of February, one day before media reports. The graph shows that prior to the media coverage, actually little scanning activity can be seen. However, after the media has covered the attack, the activity suddenly increases from a few unique IPs to between 25 and 50 IPs scanning per hour for vulnerable instances. One exception can be noted at the end of January when around 125 unique IP addresses scan for memcached servers. From the figure it seems that the media reports have triggered an increase in activity. Likely adversaries trying to uncover vulnerable server which can be used in future attacks. The number of unique IP addresses probing for memcached servers is a good indication of activity. However, it only tells a part of the story, it does not show the intensity of the activity. We can measure intensity in terms of probes received per hour, after all the faster the attacker scans the Internet, the sooner a sizeable list of exploitable services will be available. Given the size of our monitoring IP ranges, we can obtain a solid estimate of the scanning speed of each actor. The figure on the left shows the intensity of probing activity, measured in packets received per hour. It is similar to the figure portraying the unique IPs per hour, showing the same increase, in terms of packets per hour, around the time of the attack on GitHub. However, contrary to the previous figure, there are four distinct intensity spikes prior to the attack. This indicates that prior to the attack, some reconnaissance missions are performed to find servers which can be used in an attack. The spike in intensity of late January coincides with the spike in number of unique IP addresses, the three spikes prior do not. The discrepancy between activity and intensity give an insight into another aspect of adversaries and probing techniques. The three first spikes are created by one IP address, the fourth one is generated by 125 different IP addresses from the same /24 subnet working together. The use of multiple IP addresses shows that there is a difference in adversary skills or resources. Using multiple IP addresses is more costly and complicated than using a single IP for probing. Based on the scan progression, we can see several reconnaissance campaigns prior to the attack. Additionally, it reveals that some adversaries are more skilled or possess more resources than others. When we look at the payload of the probing packets, we can obtain further information about the adversaries, for example in how they structure their packets, whether they make mistakes in doing so, or how they optimize their techniques to increase the impact. Indeed, analyzing the different payload of the probes reveals that there is a plethora of different probing packets, indicating different knowledge and skill levels. Some adversaries manage to immediately craft correct probing packets whereas other take several attempts to make valid probes. The difference in payload and how the packets are formed also allows some fingerprinting of the tools used for the probing. Some probe types are specific to single IP addresses whereas othes gain fast some wide-spread adoption. As shown on the right, the majority of senders finally converge to the "correct" solution, which will trigger a sufficiently large response from any unsecured memcached server without a mistake in the request packet format. This development seems greatly helped with the publication of two proof-of-concept programs on pastebin, that demonstrate how to send a valid packet in the required format. Shortly after each PoC publication, we see an influx in people participating in memcached probing and a spike in more and more actors adopting this technique. The progression towards the right solution also shows us that some adversaries are more skilled than others, and shows clear evidence for the collaboration between IP addresses. For a more in depth read please refer to the original publication: How Media Reports Trigger Copycats: An Analysis of the Brewing of the Largest Packet Storm to Date. In addition to the increase in activity related to media reports and the difference in skill set the paper analyses the convergence of commands used for probing, the research also analyzes the response of memcached server operators on the events following the media reports.
<urn:uuid:8a4f3458-7563-44dc-8bde-ec103a1e64bb>
CC-MAIN-2024-38
https://www.cyber-threat-intelligence.com/research/memcached-copycats/
2024-09-14T10:20:31Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651579.22/warc/CC-MAIN-20240914093425-20240914123425-00384.warc.gz
en
0.952551
1,527
2.90625
3
The left and right halves of our brains develop differently, as each hemisphere ‘specialises’ in certain functions. For example, for most people the left hemisphere–controlling the right hand–is dominant for language. But brain asymmetry is sometimes affected in people with developmental or psychiatric disorders like autism, which is characterised by impaired social cognition, repetitive behaviour and restricted interests. PhD student Merel Postema explains: “Previous studies have suggested that people with autism spectrum disorder are less likely to have the typical asymmetries for language dominance or hand preference. However, it has not been clear whether asymmetry of the brain’s anatomy is affected in autism, because different studies have reported different findings.” Do people with autism have a more symmetrical brain? To settle this question, scientists from the international ENIGMA consortium of brain researchers decided to do a large-scale study, based on brain scans that were collected in different countries over more than 20 years. This is by far the largest ever study of this question, using brain scan data from 1,774 people with autism and 1,809 healthy controls. The team found that the left and right cerebral hemispheres of the brain are indeed more similar in people with autism. In other words, people with ASD had less brain asymmetry. The reduced asymmetry was mostly found for cortical thickness, at various locations across the brain’s surface. In the healthy brain, the thickness of the cerebral cortex (the thin layer of gray matter that covers the brain) differs between the left and right hemispheres. Importantly, the anatomical differences did not depend on age, sex, IQ, the severity of symptoms, or medication use. Altered asymmetry of cerebral cortex thickness in autism spectrum disorder. The stars show the affected brain regions. The image is credited to Clyde Francks, MPI Nijmegen. “The very small average differences in brain asymmetry between affected people and controls mean that changes of brain asymmetry will not be useful in terms of clinical prediction”, says study leader Clyde Francks. “But the findings might inform our understanding of the neurobiology of autism spectrum disorder”. As the bulk of the data were from children, the findings suggest that altered development of the brain’s left-right axis is involved in autism, affecting widespread brain regions with diverse functions. For example, many of the affected brain regions overlap with the default mode network, which is a network of inter-connected brain regions that is particularly active during passive rest and mind-wandering, rather than when doing a specific task. Why this might relate to ASD may be a topic of future research. Social deficits are a defining feature of autism spectrum disorder (ASD), but also frequently affect children with attention-deficit/ hyperactivity disorder (ADHD)1, and can occur in obsessive-compulsive disorder (OCD) as well2–5. Increasingly, recognition of the overlapping and related nature of both the symptoms and the biology of different neurodevelopmental disorders6–8 has led to a call for research that spans diagnostic boundaries, and focuses instead on dimensions of psychopathology9. Decades of neuroimaging research have begun to delineate the neural substrates of sociality. Different theoretical models of the social brain have been put forward from meta-analyses10,11 and reviews of the literature12–14 (summarized in Supplementary Table 1). Across models, three functional/structural groups emerge. Brain regions hypothesized to be involved in mentalization and empathy (group 1), cluster along the midline and lateral aspects of the brain, including the temporal–parietal junction (TPJ), superior temporal gyrus/sulcus (STS/STG), dorsal medial prefrontal cortex, temporal poles, and the posterior cingulate. Anterior and prefrontal regions (group 2), including the anterior and dorsal cingulate, orbital frontal cortex, and the dorsal and ventral lateral prefrontal cortices, may contribute to executive function and cognitive control over affective and social processes. Deeper cortical and subcortical structures (group 3), including the insula, amygdala, hippocampus, and the dorsal and ventral striatum, are more central to affective responding, memory, and social reward processing (Supplementary Table 1). Structural neuroanatomical differences in many of these regions have been detected in individuals with OCD, ASD, and ADHD compared to controls8,15–21. For example, recent meta-analyses suggest increased frontal lobe thickness in ASD20, thinner temporal/parietal thickness in ASD and OCD20,22, and smaller subcortical volumes in ASD and ADHD20,23. The extent to which social deficits may differentially localize to specific brain regions/networks in different disorders is unclear, however24. Cortical gray matter volume is a product of cortical thickness and surface area. Recent work has provided evidence that cortical thickness and cortical surface area measurements are under distinct genetic influences34, and follow unique developmental timelines35,36, necessitating that they be studied independently. Cortical thickness measurements on magnetic resonance imaging (MRI) are thought to reflect the underlying cortical microstructure, involving the number and organization of cortical neurons, neuronal dendritic arborization, the number and size of glial cells, and to some extent the maturation of the adjacent white matter37. Cortical thickness, including the timing and rate of cortical thinning, has been a major area of study across neurodevelopmental disorders, particularly in ASD24. The association between brain (i.e., cortical thickness/subcortical volume) and behavior (i.e., social deficits) was the focus of the following study, in efforts to identify and compare the neural substrates of sociality across disorders. Given that the specific behavioral dimensions that contribute social impairments may vary across neurodevelopmental disorders (e.g., impaired mentalization in ASD38, executive function in ADHD39, and reward processing in OCD40), one hypothesis is that the neutral substrates of sociality will also differ by diagnosis (e.g., primarily lateral mentalization regions in ASD, frontal cognitive regions in ADHD, and subcortical regions OCD). An alternative hypothesis is that the brain regions associated with social deficits will span diagnostic boundaries. In children with ASD, ADHD, and OCD, for example, white matter fractional anisotropy correlated with adaptive functioning abilities, irrespective of diagnosis8. Neuroimaging analyses comparing the structural neuroanatomical correlates of social deficits across children with ASD, ADHD, or OCD have not yet been performed. To address this knowledge gap, we first compared cortical thickness/subcortical volume measurements in social brain regions across a group of children with ASD, ADHD, OCD, or controls. Next, we examined how cortical thickness/subcortical volume corresponded with social deficits across disorders. We hypothesized that social deficits would correlate with structural anatomy, irrespective of diagnosis. Max Planck Institute Marjolein Scherphuis – Max Planck Institute The image is credited to Clyde Francks, MPI Nijmegen. Original Research: Open access “Altered structural brain asymmetry in autism spectrum disorder in a study of 54 datasets”. Postema, M.C., Van Rooij, D., Anagnostou, E., Arango, C., Auzias, G., Behrmann, M., … Francks, C.. Nature Communications doi:10.1038/s41467-019-13005-8.
<urn:uuid:3ccfa8ee-a7e4-499e-aa43-0b7d88106cd7>
CC-MAIN-2024-38
https://debuglies.com/2019/11/01/people-with-autism-have-left-and-right-brain-areas-more-symmetrical/
2024-09-15T17:34:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651632.84/warc/CC-MAIN-20240915152239-20240915182239-00284.warc.gz
en
0.911687
1,585
3.75
4
Common Types of Cyber Attacks Table of Contents - By Bree Ann Russ - Published: Jul 13, 2022 - Last Updated: Jul 18, 2022 The internet has grown significantly over the past 20 years and continues to grow at an accelerated rate. The cyber threat landscape continues to grow as more people and businesses invest in digital services. There are many types of cyber attacks that cyber attackers use to try to gain access to your data or devices or even try to take control of them. There are also several red flags that you should watch out for if you suspect that a cyber attack is targeting you. Keeping informed about the latest threats, staying updated on security best practices, and regularly updating software can all help protect against different types of cyber security attacks. While there are many different ways a cyber attack can occur, some common types of attacks occur most often. Distributed Denial of Service (DDoS) Attacks A denial of service attack, or DDoS attack for short, is when someone sends a large amount of traffic to a computer network or website that exceeds the expected capacity for that network. The influx of traffic leads to the inability to access services for many users or can cause the website or network to crash. It is a common cyber attack. DDoS attacks can happen to any type of computer network, including websites, banks, electrical grids, and government services. There are two main types of DDoS attacks: The first type is when the cyber attacker has an extensive network of computers under their control (a botnet), and they use those computers to send a large amount of traffic to a single website or network. DDoS attacks can also happen when a cyber attacker uses a command and control (C&C) server to send a large amount of illegitimate traffic from many computers they don't control. DDoS attacks often have malicious intent, such as to extort money from organizations or companies, gain publicity, or try to silence organizations or individuals from speaking out. Some cyber attackers will attempt to gain access to a network and then demand a ransom payment in exchange for the stolen data. Ransomware is a type of malicious software (or malware) that can infect a computer or network and block or prevent users from accessing their data. Ransomware attacks have risen over the past few years and have become increasingly sophisticated as time goes on. Ransomware typically gets onto your computer through a malicious email attachment or link or by you clicking on a malicious advertisement online. When you click on the attachment or link, the ransomware gets downloaded onto your computer and attempts to lock up your data and make it inaccessible. Ransomware can be challenging to remove, and if it's not removed within a certain amount of time, it can permanently lock you out of your data. A data breach occurs when someone steals, hacks, or improperly accesses sensitive data. Data breaches often occur when organizations don’t have proper security measures in place to protect their data. The most common types of data breaches occur when hackers gain access to computer systems and networks through malware, backdoors, and vulnerabilities in software. In some cases, there is malicious intent behind the data breach. However, in other cases, the breach is accidental. Data breaches often lead to identity theft and financial fraud, so it's essential to be aware of potential signs of a breach. These can include: - Locked accounts - Changing file locations or names - Unusual activity - New files or software - Significantly slowed networks - Extraordinary amounts of traffic (in or out) - Devices that have damage or noticeable signs of tampering Hacking is the act of controlling a computer without authorization. Hackers can gain access to computer systems by exploiting vulnerabilities, installing malicious software, tricking people into giving them their login credentials, and more. Many people fall victim to hacking by clicking on a seemingly familiar link that does not go where they believe it will go. Hackers may use the computers they gain access to for malicious purposes, like stealing data, sending spam and phishing emails, distributing viruses, and other unsavory acts. Hackers commonly target large organizations, especially ones with lots of sensitive data, like banks, hospitals, and government agencies. To avoid being hacked, make sure you regularly update your software and install patches as they become available, keep an eye on your system for suspicious activity, and be extra cautious when clicking on links or attachments from emails. Phishing is a type of social engineering where cyber attackers attempt to trick people into providing sensitive information, like usernames, passwords, credit card numbers, and other personal data. They do this by pretending to represent a trusted organization. Phishing happens when a cyber attacker sends an email (or texts you) that looks like it's from a company like your bank, your internet service provider, or your phone company. Phishing emails are often compelling, so it can be difficult to tell that they're not actually from the organization they say they're from. These emails often contain links to websites that look like they’re from a trusted company but lead users to another site that the cyber attacker controls. Phishing emails are usually easy to spot if you're aware of the latest phishing scams and regularly update your software. Some of the more common phishing examples include: - Emails saying you must log into your account from a specific link to prove you are who you claim to be - Any email asking you to respond with your username and password to a website - Texts or posts claiming you won money, a prize, or have someone trying to reach you to give you something Proactivity and Education Can Help Keep You Safe Cyber attacks are on the rise, and everyone needs to be aware of the latest threats and how to protect against them. While there are many different types of cyber attacks, there are also ways to help reduce your risk of being targeted, such as using identity protection services. It's essential to keep informed about new threats, stay updated on security best practices, and regularly update your software to protect against different types of cyber attacks. The more you know, the safer you can remain.
<urn:uuid:3221a276-d11c-4548-88c3-c8af3bf179bb>
CC-MAIN-2024-38
https://www.idstrong.com/sentinel/common-types-of-cyber-attacks/
2024-09-16T22:27:20Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651714.51/warc/CC-MAIN-20240916212424-20240917002424-00184.warc.gz
en
0.952555
1,261
3.34375
3
A Good Clinical Practice (GCP) audit is a critical component of clinical trials, ensuring adherence to standards and safeguarding participants’ rights, safety, and data accuracy. GCP audits are essential to minimize errors in data analysis, maintain regulatory compliance, and uphold the integrity of clinical trial results. The clinical trial regulatory environment is governed by stringent regulations from government authorities and health agencies. GCP audits play a crucial role in ensuring compliance with these regulations, with the International Council for Harmonization (ICH) guidelines providing a global foundation for audit consistency. Several types of GCP audits, including pre-study, site initiation, interim monitoring, and close-out audits, ensure compliance with regulatory requirements and protect participant safety and data integrity. Preparing for a GCP audit involves understanding audit types, regulatory landscapes, and compiling necessary documents, such as protocols and consent forms. Anticipating needs, reviewing organizational policies, and organizing records are crucial steps for successful preparation. During the audit, adherence to proper procedures, objective assessments, thorough documentation review, interviews with relevant individuals, and identification of risk factors are essential for a successful GCP audit. Deficiencies often encountered during GCP audits include inadequate infrastructure, insufficient staff training, missing Standard Operating Procedures (SOPs), or irregular data documentation. Immediate action to rectify these deficiencies is necessary, with auditors providing solutions and recommendations. Effective record-keeping during GCP audits involves documenting inspections, collecting supporting evidence, using templates or checklists, regularly updating records, and secure data storage. Managing GCP documents requires secure storage, retention of originals, setting up an organized file management system, regular document reviews, and ensuring quick access to documents during audits. Reporting GCP audit results involves providing a briefing to the sponsor, followed by a detailed written report outlining audit findings, observations, nonconformances, and recommended corrective actions. This report is submitted to regulatory bodies. In cases of severe deficiencies, consequences can include fines or clinical trial suspensions. Developing a remediation plan with timelines and additional training is necessary to address issues and gain approval from governing bodies. In conclusion, GCP audits are integral to clinical trials, ensuring compliance, data integrity, and participant safety. Adequate preparation, proper conduct, and effective management of records and documents are key to a successful audit process. GCP audits maintain the credibility and trustworthiness of clinical trial data, contributing to the overall success of these trials. Introduction to GCP Audits A Good Clinical Practice (GCP) audit is an important part of clinical trials that ensures that the methods used and data collected are meeting certain standards. Through such an audit, sponsors can observe and evaluate the trial protocols, data, staff, and processes related to the clinical trial. The audit process is an essential part of the clinical trial process since its purpose is to protect the rights, safety, and well-being of study participants, while at the same time ensuring the data is accurate and reliable. GCP audits are also important because they help minimize the risk of errors in the analysis of trial data, which could lead to erroneous findings and conclusions. Additionally, GCP audits can help ensure that the data submitted to regulatory authorities for review is complete, accurate, and up to date. In this guide, we will provide an overview of GCP audits, including what they are, why they are important, types of audits, preparation and conduct, potential findigns, and reporting the results. Clinical Trial Regulatory Environment Clinical trials are subject to stringent regulations from government authorities and health agencies. Good Clinical Practice (GCP) audits are an important part of ensuring that clinical trial protocols and documentation are compliant with prevailing regulatory guidelines. The purpose of a GCP audit is to assess the quality and accuracy of all processes and procedures related to a clinical trial. They also provide assurance that clinical trial data and results are reliable and meaningful. The framework of current pharmaceutical development and clinical trial regulations is established by the International Council for Harmonization (ICH). This body is composed of leading regulatory authorities from the United States, the European Union, Japan, and other countries. The ICH guidelines provide the foundation for GCP audits and help to ensure uniformity and consistency across global clinical trials. Types of GCP Audits Good Clinical Practice (GCP) Audits are essential for upholding the safety and reliability of clinical trials. They protect the rights, safety and well-being of trial participants, as well as to ensure the quality and integrity of the data collected. The aim of these audits is to evaluate the procedures and documents associated with the trial and identify any discrepancies that may exist. There are several different types of GCP audits that can take place during a clinical trial. These include: - Pre-Study Audit – conducted prior to the launch of the trial - Site Initiation Visit – conducted after the pre-study audit, but before participants are enrolled in the trial - Interim Monitoring Visit – conducted during the trial to assess ongoing trial procedures - Close-Out Visit – conducted at the close of the trial to confirm data accuracy and compliance with regulations These audits help to ensure that all trial activities are carried out in accordance with regulatory requirements and good clinical practice standards. This helps to protect the safety of study participants and the accuracy of the data generated by the trial. Preparation for a GCP Audit When preparing for an upcoming GCP audit, it is important to take the necessary steps to ensure that you and your clinical trial team are ready. Proper preparation involves understanding the different types of audits that might occur during the clinical trial process and familiarizing yourself with the regulatory landscape. This means doing research about the regulations that cover the conduct of a trial and reading up on any guidance documents related to GCP. It’s also helpful to compile any list of documents that are necessary for the audit and to make sure that everything is organized and accessible. This often includes protocols, consent forms, investigational product related documents, and any records of adverse events. Be sure to place these in a secure, accessible location, such as a cloud-based platform, so that they can be easily accessed during the audit. You should also anticipate needs that you may have during the audit, such as additional staff or support personnel, so that you can be better prepared in advance. Lastly, review any policies and procedures related to GCP that your organization might have established. This will help you understand the expectations that may be set during the audit. Conducting the Audit When conducting a Good Clinical Practice (GCP) audit, it is important to ensure that all steps in the process are properly followed. The audit should involve the use of an objective assessment, while also making sure that the right questions are asked and that all necessary information is collected. Here are some tips for running a successful GCP audit: - Establish Clear Expectations: Make sure that everyone involved in the audit knows the objectives, so that the process can be completed in an efficient and comprehensive manner. - Verify Documentation: Ensure that all documentation is reviewed and verified to make sure it is accurate and complete. - Perform Interviews: It is important to speak to relevant individuals during the audit to gain a better understanding of the practices being examined and their potential effects. - Determine Risk Factors: Make a list of any potentially problematic areas during the audit, as they can help to provide insights into how to improve practices in the future. - Provide Feedback: Provide feedback on the results of the audit, so that any issues can be addressed more quickly and efficiently in the future. By following these tips, you can ensure that a GCP audit is conducted professionally and accurately, while also helping to protect the integrity of a clinical trial. When conducting a Good Clinical Practice (GCP) audit, it is important to be prepared for the possibility that certain deficiencies may be found. Common deficiencies encountered during GCP audits include inadequate infrastructure and facilities, insufficient training for staff, missing Standard Operating Procedures (SOPs), or irregularities in the documentation of data. These deficiencies can have a negative impact on the quality of the clinical trial and the accuracy of its results. If any major deficiencies are found during the audit process, the GCP auditors must take immediate action to minimize the risk of further errors. This should include providing solutions to rectify the identified problems and recommendations to help prevent similar issues occurring in the future. For example, if inadequate training for staff is discovered then the auditor may suggest additional employee training or the establishment of an internal educational program. When performing a GCP Audit, it is important to keep accurate and detailed records of the process. Keeping organized records can help ensure that the audit is successful and all findings are considered in their proper context. Here are some helpful tips for keeping clear records during the audit process: - Make sure to document any inspections or observations. - Collect sufficient evidence to support the audit findings. - Create templates or checklists to stay on track and consistently record data. - Update records regularly to make sure everything is up-to-date. - Use secure systems to store all files and documents in a safe place. By following these best practices for record keeping during the audit process, one can guarantee a smooth and successful GCP audit. Management of GCP Documents When it comes to a GCP audit, records of all relevant documents are essential. Adequately managing and keeping track of these documents can make a big difference in the success and smooth running of the audit process. Here are some tips for effectively managing GCP documents: - Make sure documents are stored securely and safely. Physical records should be kept in a safe place and digital records must be password-protected and stored on a secure server. - Retain originals of all documents and keep copies if needed. Maintain backups of all records for added security. - Set up a file management system that is easy to navigate and understand. Include an index of documents complete with titles, dates, and other important information. - Ensure documents are kept up-to-date and tracked regularly. Review records regularly and delete obsolete records. - Make sure the system is designed to accommodate auditing requests. Ensure that documents can be quickly found and provided when requested. Reporting the Results Once the audit is complete, it is important to record and report the findings. The process for doing this varies depending on the scope of the audit and the regulations that must be followed. Generally, the auditors will provide a briefing to the sponsor of the clinical trial to explain the audit results. The purpose of the briefing is to properly inform the sponsor and discuss any corrective action that must be taken. After the briefing, the auditors will provide a written report that outlines the audit findings and recommendations. The report should include detailed information such as background information about the clinical trial, the audit findings, observations, any nonconformances, and suggested corrective action. It is also important to provide details about the resources dedicated to the audit, as well as any follow-up actions that are necessary. The report should be reviewed by the sponsor and then provided to the applicable regulatory body. Impacts & Remediation When a GCP audit does not go well and serious deficiencies are identified, the consequences can be severe. Depending on the nature of the finding, there may be fines or even clinical trial suspensions that could hinder the progress of your research. It’s important to be aware of any potential impacts of a failed audit and know how to address the situation if it does occur. One of the remedies for corrective action is to develop a remediation plan. This typically includes outlining the steps you plan to take to rectify the situation, along with proposed timelines for implementation of those changes. You may also need to make additional training available to personnel and make sure adequate documentation is maintained. The remediation plan should be submitted to the relevant governing body for approval. If your remediation plan is accepted, you may be able to continue on with the clinical trial. However, if the findings are deemed to be too significant, the governing agency may require additional research to be conducted before you can proceed. It’s important to follow all the protocols for completing a GCP audit, so if a problem does arise, you are already aware of the corrective steps to take in order to get back on track. Good Clinical Practice (GCP) Audits are an essential part of any clinical trial. By ensuring that all protocols, processes, and documents are in compliance, these audits help to protect the accuracy and integrity of a clinical trial’s results. They also act as a way to assess the trustworthiness of any trial data. As such, conducting regular GCP audits is vital to the success and credibility of any clinical trial. Knowing what to expect from GCP audits and preparing ahead of time can help ensure that any audit goes smoothly. By understanding the regulatory requirements of clinical trials, keeping accurate records, and staying organized, a clinical trial can meet the demands of any GCP audit. GCP audits play an important role in keeping clinical trials safe, reliable, and trustworthy. Knowing what to expect beforehand and being adequately prepared is the key to successfully passing a GCP audit. GCP Audits: Are You Ready? Good Clinical Practice (GCP) Audits are an important part of any clinical trial. They are designed to ensure the safety and integrity of the procedures and processes involved in a clinical trial. It is essential to be adequately prepared before a GCP audit takes place. In this guide, we’ll provide an overview of the regulatory environment of clinical trials, look at the different types of GCP audits, give suggestions on how to prepare for one, look at common deficiencies that may be found during the audit process, and discuss the potential impacts and corrective measures that must be taken if any are found. We will also review the importance of keeping accurate records throughout the audit process, explain the process for reporting the results, and provide helpful resources for learning more about GCP audits. By becoming familiar with the topics outlined here, you’ll be able to confidently begin an audit and ensure a successful outcome. Frequently Asked Questions about Clinical Trial Good Clinical Practice (GCP) Audits 1. What are Good Clinical Practice (GCP) Audits? GCP Audits are designed to evaluate the validity and integrity of clinical trials in order to ensure that they are conducted in accordance with international standards. The audits assess the systems, processes, and documentations related to clinical research. 2. What is the purpose of Good Clinical Practice (GCP) Audits? GCP Audits are essential for ensuring the safety and well-being of human participants in clinical trials. They help ensure that the best possible data is being collected and that ethics and legal requirements are being met. 3. What are the different types of GCP Audits? There are different types of GCP Audits, such as pre-study audits, mid-study audits, and post-study audits. Pre-study audits are typically focused on an organization’s ability to conduct the trial; mid-study audits review trial progress; and post-study audits review the data collected. 4. How should I prepare for a GCP Audit? It’s important to prepare thoroughly in order to have a successful audit. Start by taking stock of all relevant documentation and procedures so that everything is easily accessible during the audit. Familiarizing yourself with current GCP standards and regulations is also key. 5. What common deficiencies may be found during a GCP Audit? Common deficiencies typically relate to deviations from GCP protocols, inadequate training or record-keeping, lack of monitoring procedures, poor management, and insufficient resources. 6. What are the impacts of failing a GCP Audit? Failing a GCP Audit can have serious consequences, such as suspension of a study, increased costs due to a delay in the timeline, or decreased trust in a company’s ability to carry out a clinical trial reliably. 7. How do I learn more about Good Clinical Practice (GCP) Audits? Useful information can be found by visiting the websites of regulatory bodies such as the US FDA and EMA, as well as the websites of professional organizations such as the Association for Clinical Research Professionals (ACRP) and the Society for Clinical Research Sites (SCRS).
<urn:uuid:df0a253f-69fb-49b0-9373-5f0a9ef85926>
CC-MAIN-2024-38
https://msbdocs.com/security-compliance/what-are-gcp-audits/
2024-09-18T05:56:22Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651836.76/warc/CC-MAIN-20240918032902-20240918062902-00084.warc.gz
en
0.943035
3,378
2.625
3
XG-GPON and XGS-PON: Understanding the Principles and Applications The Evolution of GPON Technology GPON(Gigabit-capable Passive Optical Network) is the access technology of a passive optical network (PON) based on the ITU-T G.984.x standard. Its development has undergone continuous evolution and improvement, while also driving the development and popularization of fiber optic access networks. With the increasing demand from users for higher bandwidth, new generations of PON technology such as XG-PON(10 Gigabit-capable Passive Optical Network) and XGS-PON(10 Gigabit Symmetrical Passive Optical Network) have emerged. XG-PON and XGS-PON are collectively referred to as 10G GPON. XG-PON is an upgraded version of GPON technology, capable of providing up to 10 Gbps downstream bandwidth and 2.5 Gbps upstream bandwidth. XGS-PON, on the other hand, is a further development of GPON technology aimed at providing symmetrical upstream and downstream bandwidth, with both upstream and downstream data transmission rates reaching 10 Gbps. These technologies offer higher rates and bandwidth than traditional GPON, catering to the growing demand for broadband and emerging applications such as 4K/8K videos, virtual reality (VR), and augmented reality (AR). For more details about the difference between GPON, XG-PON, and XGS-PON, please check: GPON vs. XG PON vs. XGS PON: Which PON Technology Is Right for Your Network? Unlock Network Architecture: XG-PON and XGS-PON In fact, XG-PON and XGS-PON have the same network structure as GPON, consisting of OLT, ONU, and ODN, following the standard P2MP architecture of PON networks, and can be widely used in FTTx scenarios. OLT: A convergence device placed at the central office that terminates the PON protocol. ONU: A user-side unit or terminal located on the client side, providing various interfaces to the user. ODN: Consisting of optical fibers and one or more Passive Optical Splitters (POS), it connects the OLT and the ONU. For more information about these components, please check: Decoding OLT, ONU, ONT, and ODN in PON Network Exploring the Working Principles of XG-PON and XGS-PON 10G GPON uses Wavelength Division Multiplexing (WDM) technology to achieve bidirectional data transmission on a single fiber within the same ODN network by using different wavelengths for upstream and downstream, as shown in the figure: Downstream Data Transmission The downstream link refers to the process of data transmission from the OLT to the ONU. In this direction, data is sent via broadcasting. The OLT broadcasts data to all ONUs, and each ONU selects and receives the data intended for itself, discarding other data. The diagram below illustrates the downstream service forwarding process. On the OLT, data flows are encapsulated into GEM ports within the service processing unit. The OLT broadcasts the data encapsulated into GEM ports to all ONUs. Upon receiving the data sent by the OLT, each ONU determines whether to process or discard the data flow based on the XGEM port ID. Upstream Data Transmission The upstream link refers to the process of data transmission from the ONU back to the OLT. To avoid signal conflicts and interference, the upstream link is managed using a Time Division Multiple Access (TDMA) mechanism. The OLT allocates time slots for each ONU, and each ONU sends data within its designated time slot to prevent data collisions. The optical signals aggregated at the OLT are converted back into electrical signals for processing, and the data is forwarded to higher-level networks. The upstream service forwarding process is illustrated in the diagram below. On the ONU, data flows are encapsulated into the corresponding GEM ports and then mapped to the corresponding T-CONTs. The OLT schedules data flows based on T-CONTs and sends them to the OLT in sequence according to time slots. The OLT then decapsulates the data flows from the GEM ports and forwards them to the corresponding service units. How to Deploy 10G GPON Network 10G GPON Network 10G GPON network refers to a network infrastructure entirely based on 10 Gigabit Passive Optical Network (GPON) technology. The 10G GPON OLT connects to 10G GPON ONUs, with signal transmission exclusively over the ODN network. This network configuration features a uniform device type (solely 10G GPON equipment), making it easy to deploy and maintain. It is suitable for new deployments in 10G GPON FTTx network scenarios. 10G GPON and GPON Hybrid Networking This networking method involves having both 10G GPON and GPON equipment within the same PON network, sharing a single ODN. In a hybrid network, 10G GPON and GPON access boards are inserted into the same OLT to handle access services. Due to the different wavelengths used by 10G GPON and GPON (10G GPON typically uses 1577nm for downstream and 1270nm for upstream, while GPON uses 1490nm for downstream and 1310nm for upstream), passive xWDM/xPON Multiplexers is needed for wavelength allocation. It is important to note that on the OLT, a 10G GPON access board can only connect to 10G GPON ONUs, and a GPON access board can only connect to GPON ONUs. This networking approach achieves a smooth transition from GPON to 10G GPON while maximizing investment protection. However, the coexistence of 10G GPON and GPON devices in the network also presents challenges in terms of network complexity and maintenance. 10G GPON & GPON Shared Board Hybrid Networking The key difference between this hybrid networking method and the previous one is that it uses shared service boards that support simultaneous processing of 10G GPON and GPON signals. These types of service boards can handle two different wavelengths simultaneously. This method uses built-in xWDM/xPON Multiplexers, commonly referred to as Combo-PON equipment. This type of networking provides a high degree of flexibility, allowing 10G GPON and GPON services to be deployed flexibly on the same OLT according to demand. By using shared service boards, the cost of equipment procurement can also be reduced, resulting in cost efficiency. As the next-generation PON technology, 10G GPON meets the increasing demands of high-bandwidth network applications while enabling high-bandwidth access at a lower cost. FS supports the customization of XG(S)-PON devices and the FS solutions team will continue to provide you with reliable and professional services. For more information on PON technology and equipment, please visit FS PON Networks.
<urn:uuid:9a5f1d9e-d5d3-4373-b223-990400f837f8>
CC-MAIN-2024-38
https://community.fs.com/article/xggpon-and-xgspon-understanding-the-principles-and-applications.html
2024-09-20T12:42:46Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652278.82/warc/CC-MAIN-20240920122604-20240920152604-00784.warc.gz
en
0.903866
1,478
2.8125
3
Robot Theorizes, Proves Own Scientific Discoveries "Adam" can conduct experiments and form hypotheses to explain the results. April 3, 2009 Scientists at the United Kingdom's Cambridge and Aberystwyth universities have created a "robot scientist" that they believe is the first automaton to make its own scientific discoveries. The robotic computer system, named Adam, automates the scientific process, carrying out each stage on its own. The robot has discovered simple but new scientific knowledge about the genomics of baker's yeast, or Saccharomyces cerevisiae, according to the scientists. Baker's yeast can serve as a model for more complex life forms. The scientists said their manual experiments confirmed the robot's hypotheses and its findings. Adam used artificial intelligence to hypothesize that some genes in baker's yeast code for specific enzymes cause biochemical reactions in yeast. The robot came up with experiments to test the hypothesis, ran the experiments using laboratory robotics, interpreted the results, and repeated the cycle. Ross King, a computer scientist at Aberystwyth in Wales, worked with systems biologists at the University of Cambridge. The team plans to build a second robot, Eve, to help scientists find new drugs to combat diseases like malaria and schistosomiasis, a tropical parasitic infection. King said scientists could one day have teams of robots and people working together in laboratories. He said the detailed records of biological experiments that are sometimes "irksome for human scientists" are easy for robots. Stephen Oliver, co-author of the paper and professor of systems biology and biochemistry at England's University of Cambridge said Adam's uniqueness comes from the machine's ability to reason and come up with theories. "As we start to consider living systems in a holistic manner, the complexity of such systems means that it will become increasingly difficult for scientists to formulate hypotheses unaided," Oliver said in a prepared statement. "Thus it will be necessary for human and robot scientists to work together to achieve the goals of biological research." The Biotechnology and Biological Sciences Research Council funded the work, which was published in Friday's issue of the journal Science. Each year, InformationWeek honors the nation's 500 most innovative users of business technology. Companies with $250 million or more in revenue are invited to apply for the 2009 InformationWeek 500 before May 1. About the Author You May Also Like
<urn:uuid:26b13ba2-c607-4767-8951-ba9b11fd5d02>
CC-MAIN-2024-38
https://www.networkcomputing.com/network-infrastructure/robot-theorizes-proves-own-scientific-discoveries
2024-09-20T14:29:44Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700652278.82/warc/CC-MAIN-20240920122604-20240920152604-00784.warc.gz
en
0.941375
489
3.125
3
The device can monitor heart rate, blood pressure, blood oxygenation levels, the patient’s weight, and stroke volume The Rochester Institute of Technology and the University of Rochester Medical Center (URMC) are planning a clinical trial of AI-enabled toilet seats to determine if vital signs can be monitored to reduce re-hospitalization of heart patients. The venture is being funded by a $3 million grant from the National Institutes of Health (NIH). The “Fully-Integrated Toilet Seat” (FIT) will use AI and sensors to provide physicians with updated patient data, and an early-alert system to help identify possible deterioration sooner. “The cost of readmission usually exceeds the cost of the original hospitalization when individuals have a diagnosis of heart failure,” said David Borkholder, former Bausch and Lomb professor in RIT’s Kate Gleason College of Engineering. “We are using machine learning and AI to develop an early alert system so that we can identify when a patient’s health begins to subtly deteriorate and hopefully see it much sooner than the patient would detect it themselves.” Borkholder is leaving RIT to join Heart Health Intelligence, the startup that will be producing the connected toilet seats. The sensors in the toilet seats measure the electrical and mechanical activity of the heart, and can monitor heart rate, blood pressure, blood oxygenation levels, the patient’s weight, and even stroke volume – the amount of blood pumped out of the heart with every beat. Development of the smart seat started at RIT in 2014, and now AI is being incorporated to analyze the output in real time. The first phase of the clinical trials will start with about 250 patients, an RIT spokesperson told AI Business. “There are a number of factors that can be evaluated in these patients,” Wojciech Zareba, professor of Cardiology at the University of Rochester Medical Center, said. “It is like having a patient on bedside monitoring in an intensive care unit. “At home, people don’t usually have these monitoring tools. This seat is serving as a good monitoring tool and even if it is not continuous, it will be used by patients several times per day, and each time, it will record data and send it to be processed.” The systems in the seat can be trained to recognize patterns unique to each person using it and are able to distinguish between more than one household user. “Individuals have unique features, from distinct eye and hair color to biological characteristics, such as an electrocardiogram (ECG),” Zareba said. “Looking at ECGs from two people, physicians can see differentiations, so the data on a given patient will not be mixed up. “Medicine today is oriented toward prevention. If you have good tools which will allow you to catch certain signs and symptoms that could be worsening, it will be easier to do it. This is the future of medicine.” About the Author You May Also Like
<urn:uuid:773073f1-7722-41e7-8865-a6a87b4d80a1>
CC-MAIN-2024-38
https://aibusiness.com/verticals/researchers-use-ai-and-connected-toilet-seats-to-monitor-heart-patients
2024-09-09T14:21:42Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651103.13/warc/CC-MAIN-20240909134831-20240909164831-00884.warc.gz
en
0.938547
646
2.59375
3
You can make a database available to users in different locations, on different networks, or in different time zones, by creating replicas of the database. All replicas share a replica ID which is assigned when the database is first created. The file names of two replicas can be different, and each replica can contain different documents or have a different database design; however, if their replica IDs are identical, replication can occur between them. As users add, edit, and delete documents in different replicas of a database, the content in the replicas is no longer identical. To ensure that the content in all replicas remains synchronized, you use Connection documents to schedule replication between the servers that store the replicas. Then multiple sites, teams, and users can make changes to a database and share those changes with everyone else who has access to that database. In addition, using replicas and scheduling replication reduces network traffic. Users never need to connect to a single central server that stores the only replica of a particular database. Instead, they can access a replica of that database on one or more local servers. These distributed replicas can also be Web sites that are hosted on different Domino® servers. Then users aren't dependent on one server when they attempt to access critical applications over the Internet. If one server is unavailable, users can access another replica of the database on another server. You can also use replicas to help manage ongoing Web site design. On one server, you can set up a Web staging area where you design and test new pages. When the design changes are tested and ready to be released, you can replicate this server with the server storing the replica of the Web site that is available to users. By using replicas and replication this way, you prevent Web users from seeing your "work-in-progress." A replica of a database isn't the same as a copy of a database that you make by choosing . Although a copy of a database may look the same as the original database, a copy doesn't share a replica ID with the original database and so it can't replicate with it.Deciding when to create a replica Plan your replica strategy carefully, and create replicas on servers only when necessary. The more replicas, the greater the demand on server and network resources and the greater the need for additional maintenance. To prevent unnecessary proliferation of replicas, assign Create Replica server access to only a few administrators. Then tell users and application developers to send their requests for new replicas to these administrators. Create a replica of a database to: - Improve performance of a heavily used database. - Distribute network traffic. - Keep a database that you're redesigning separate from a production version of the database. - Keep a database available even if one server goes down. - Make a database available to users in remote locations. - Provide a replica containing only a subset of information that is relevant to a particular workgroup. - Set up Domino® system administration -- for example, you must create replicas of the Domino® Directory, the Administration Requests database, and other critical system databases. - Place a replica of a master template on each server that stores a database that inherits from the master template. - Create a backup database from which you can restore information if data becomes corrupted; since corrupted data often replicates, use this only as a secondary backup method. Keep in mind that two replicas will contain slightly different content between replications. If users need access to the most up-to-date information in a database, you can create replicas on clustered servers and then set up replication in clusters. In a cluster, all replicas are always identical because each change immediately replicates to other servers in the cluster.
<urn:uuid:fd163f90-8525-425c-9196-7be47d0ccb66>
CC-MAIN-2024-38
https://help.hcl-software.com/domino/14.0.0/admin/conf_replicas_c.html
2024-09-09T14:50:41Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651103.13/warc/CC-MAIN-20240909134831-20240909164831-00884.warc.gz
en
0.901062
771
2.71875
3
(CosmosMagazine) Two new studies suggest researchers may be getting closer to achieving “quantum advantage”, the point at which quantum computing can solve a problem that no classical computing can in a feasible amount of time – that is, solve a problem that is otherwise computationally impossible. In Europe, Danish and German researchers have built a chip that is a promising advancement in the race to build the world’s first photonic quantum computer, while a team from China used an unusual experimental set-up to demonstrate quantum advantage for the second time ever. The first study describes the development of a nanochip less than a tenth the width of the human hair yet capable of producing hundreds of stable photons, the fundamental particle of light, that can be used to store huge amount of data – and therefore form the hardware of quantum computers, like transistors in today’s computers. Practically speaking, reaching this milestone involves simultaneously controlling at least 50 quantum bits. Most big players in the quantum race today – like Google, IBM and Microsoft – have instead been focusing on using superconducting architectures, such as Google’s 53-qubit Sycamore chip. “One key difference between the two architectures thus becomes how information is detected, transmitted and manipulated, and the second is the speed. Optical frequencies are significantly higher and have the potential for an overall faster operation – similar to the difference between the classical copper telecommunication network and the fibre network for the NBN.” Photonic computing is again near the top of the candidate list for quantum computation, thanks to several advances over the past few years – including more efficient photon detectors, improvements in single photon sources, and breakthroughs in novel ways to prevent errors when encoding qubits. In fact, a team led by Jian-Wei Pan at the University of Science and Technology of China in Shanghai has just made the first definitive demonstration of quantum advantage using photons. They used a distinctly untraditional computer – a novel tabletop set-up of lasers, mirrors, prisms and photon detectors – to perform a technique called Gaussian boson sampling. The calculation took just 200 seconds – but would have been impossible on classical computers. heir results are published in Science and far surpass Google’s demonstration; Pan’s team observed 76 photons interacting.
<urn:uuid:9e2879a3-532b-4033-bdf8-1ab3559af576>
CC-MAIN-2024-38
https://www.insidequantumtechnology.com/news-archive/the-search-for-quantum-advantage-advances-in-europe-china/amp/
2024-09-12T03:38:14Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651420.25/warc/CC-MAIN-20240912011254-20240912041254-00684.warc.gz
en
0.931998
475
3.84375
4
Symmetric cryptography encrypts data in blocks of sixty-four bits, due to computer technology of the time of implementation. The key to encrypt and decrypt the data is a stream of ones and zeros, of varying size from sixty-four to four hundred and forty-eight bits. The same key is used to encrypt and decrypt the data. Symmetric cryptography is faster to decrypt and encrypt and the key is shared between both users, providing confidentiality and access controls. Symmetric being faster to compute makes it less secure. Symmetric is better to encrypt larger amounts of data. Asymmetric cryptography encrypts data using two keys, one public key for encrypting and the private key for decrypting. Key sizes can be much larger the symmetric and takes significantly more time to encrypt and decrypt, so it is best used for small amount of data due to time. The key exchange is more secure as it is handled by the PKI, a third-party system to maintain integrity. Asymmetric encryption offers better security and tamper proof.
<urn:uuid:3ef1aae0-4ad0-4396-8a38-545d74f47c42>
CC-MAIN-2024-38
https://mile2.com/forums/reply/91905/
2024-09-15T19:50:21Z
s3://commoncrawl/crawl-data/CC-MAIN-2024-38/segments/1725700651647.78/warc/CC-MAIN-20240915184230-20240915214230-00384.warc.gz
en
0.950804
214
4.03125
4